A CPU-GPU Hybrid System of Environment Perception and 3D Terrain Reconstruction for Unmanned Ground Vehicle
- Authors
- Song, Wei; Zou, Shuanghui; Tian, Yifei; Sun, Su; Fong, Simon; Cho, Kyungeun; Qiu, Lvyang
- Issue Date
- Dec-2018
- Publisher
- 한국정보처리학회
- Keywords
- Driving Awareness; Environment Perception; Unmanned Ground Vehicle; 3D Reconstruction
- Citation
- JIPS(Journal of Information Processing Systems), v.14, no.6, pp 1445 - 1456
- Pages
- 12
- Indexed
- SCOPUS
ESCI
KCI
- Journal Title
- JIPS(Journal of Information Processing Systems)
- Volume
- 14
- Number
- 6
- Start Page
- 1445
- End Page
- 1456
- URI
- https://scholarworks.dongguk.edu/handle/sw.dongguk/8853
- DOI
- 10.3745/JIPS.02.0099
- ISSN
- 1976-913X
2092-805X
- Abstract
- Environment perception and three-dimensional (3D) reconstruction tasks are used to provide unmanned ground vehicle (UGV) with driving awareness interfaces. The speed of obstacle segmentation and surrounding terrain reconstruction crucially influences decision making in UGVs. To increase the processing speed of environment information analysis, we develop a CPU-GPU hybrid system of automatic environment perception and 3D terrain reconstruction based on the integration of multiple sensors. The system consists of three functional modules, namely, multi-sensor data collection and pre-processing, environment perception, and 3D reconstruction. To integrate individual datasets collected from different sensors, the pre-processing function registers the sensed LiDAR (light detection and ranging) point clouds, video sequences, and motion information into a global terrain model after filtering redundant and noise data according to the redundancy removal principle. In the environment perception module, the registered discrete points are clustered into ground surface and individual objects by using a ground segmentation method and a connected component labeling algorithm. The estimated ground surface and non-ground objects indicate the terrain to be traversed and obstacles in the environment, thus creating driving awareness. The 3D reconstruction module calibrates the projection matrix between the mounted LiDAR and cameras to map the local point clouds onto the captured video images. Texture meshes and color particle models are used to reconstruct the ground surface and objects of the 3D terrain model, respectively. To accelerate the proposed system, we apply the GPU parallel computation method to implement the applied computer graphics and image processing algorithms in parallel.
- Files in This Item
- There are no files associated with this item.
- Appears in
Collections - College of Advanced Convergence Engineering > Department of Computer Science and Artificial Intelligence > 1. Journal Articles

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.