Cargando…

Towards a Meaningful 3D Map Using a 3D Lidar and a Camera

Semantic 3D maps are required for various applications including robot navigation and surveying, and their importance has significantly increased. Generally, existing studies on semantic mapping were camera-based approaches that could not be operated in large-scale environments owing to their comput...

Descripción completa

Detalles Bibliográficos
Autores principales: Jeong, Jongmin, Yoon, Tae Sung, Park, Jin Bae
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2018
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6111277/
https://www.ncbi.nlm.nih.gov/pubmed/30082618
http://dx.doi.org/10.3390/s18082571
_version_ 1783350623335350272
author Jeong, Jongmin
Yoon, Tae Sung
Park, Jin Bae
author_facet Jeong, Jongmin
Yoon, Tae Sung
Park, Jin Bae
author_sort Jeong, Jongmin
collection PubMed
description Semantic 3D maps are required for various applications including robot navigation and surveying, and their importance has significantly increased. Generally, existing studies on semantic mapping were camera-based approaches that could not be operated in large-scale environments owing to their computational burden. Recently, a method of combining a 3D Lidar with a camera was introduced to address this problem, and a 3D Lidar and a camera were also utilized for semantic 3D mapping. In this study, our algorithm consists of semantic mapping and map refinement. In the semantic mapping, a GPS and an IMU are integrated to estimate the odometry of the system, and subsequently, the point clouds measured from a 3D Lidar are registered by using this information. Furthermore, we use the latest CNN-based semantic segmentation to obtain semantic information on the surrounding environment. To integrate the point cloud with semantic information, we developed incremental semantic labeling including coordinate alignment, error minimization, and semantic information fusion. Additionally, to improve the quality of the generated semantic map, the map refinement is processed in a batch. It enhances the spatial distribution of labels and removes traces produced by moving vehicles effectively. We conduct experiments on challenging sequences to demonstrate that our algorithm outperforms state-of-the-art methods in terms of accuracy and intersection over union.
format Online
Article
Text
id pubmed-6111277
institution National Center for Biotechnology Information
language English
publishDate 2018
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-61112772018-08-30 Towards a Meaningful 3D Map Using a 3D Lidar and a Camera Jeong, Jongmin Yoon, Tae Sung Park, Jin Bae Sensors (Basel) Article Semantic 3D maps are required for various applications including robot navigation and surveying, and their importance has significantly increased. Generally, existing studies on semantic mapping were camera-based approaches that could not be operated in large-scale environments owing to their computational burden. Recently, a method of combining a 3D Lidar with a camera was introduced to address this problem, and a 3D Lidar and a camera were also utilized for semantic 3D mapping. In this study, our algorithm consists of semantic mapping and map refinement. In the semantic mapping, a GPS and an IMU are integrated to estimate the odometry of the system, and subsequently, the point clouds measured from a 3D Lidar are registered by using this information. Furthermore, we use the latest CNN-based semantic segmentation to obtain semantic information on the surrounding environment. To integrate the point cloud with semantic information, we developed incremental semantic labeling including coordinate alignment, error minimization, and semantic information fusion. Additionally, to improve the quality of the generated semantic map, the map refinement is processed in a batch. It enhances the spatial distribution of labels and removes traces produced by moving vehicles effectively. We conduct experiments on challenging sequences to demonstrate that our algorithm outperforms state-of-the-art methods in terms of accuracy and intersection over union. MDPI 2018-08-06 /pmc/articles/PMC6111277/ /pubmed/30082618 http://dx.doi.org/10.3390/s18082571 Text en © 2018 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Jeong, Jongmin
Yoon, Tae Sung
Park, Jin Bae
Towards a Meaningful 3D Map Using a 3D Lidar and a Camera
title Towards a Meaningful 3D Map Using a 3D Lidar and a Camera
title_full Towards a Meaningful 3D Map Using a 3D Lidar and a Camera
title_fullStr Towards a Meaningful 3D Map Using a 3D Lidar and a Camera
title_full_unstemmed Towards a Meaningful 3D Map Using a 3D Lidar and a Camera
title_short Towards a Meaningful 3D Map Using a 3D Lidar and a Camera
title_sort towards a meaningful 3d map using a 3d lidar and a camera
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6111277/
https://www.ncbi.nlm.nih.gov/pubmed/30082618
http://dx.doi.org/10.3390/s18082571
work_keys_str_mv AT jeongjongmin towardsameaningful3dmapusinga3dlidarandacamera
AT yoontaesung towardsameaningful3dmapusinga3dlidarandacamera
AT parkjinbae towardsameaningful3dmapusinga3dlidarandacamera