Cargando…
Towards a Meaningful 3D Map Using a 3D Lidar and a Camera
Semantic 3D maps are required for various applications including robot navigation and surveying, and their importance has significantly increased. Generally, existing studies on semantic mapping were camera-based approaches that could not be operated in large-scale environments owing to their comput...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2018
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6111277/ https://www.ncbi.nlm.nih.gov/pubmed/30082618 http://dx.doi.org/10.3390/s18082571 |
_version_ | 1783350623335350272 |
---|---|
author | Jeong, Jongmin Yoon, Tae Sung Park, Jin Bae |
author_facet | Jeong, Jongmin Yoon, Tae Sung Park, Jin Bae |
author_sort | Jeong, Jongmin |
collection | PubMed |
description | Semantic 3D maps are required for various applications including robot navigation and surveying, and their importance has significantly increased. Generally, existing studies on semantic mapping were camera-based approaches that could not be operated in large-scale environments owing to their computational burden. Recently, a method of combining a 3D Lidar with a camera was introduced to address this problem, and a 3D Lidar and a camera were also utilized for semantic 3D mapping. In this study, our algorithm consists of semantic mapping and map refinement. In the semantic mapping, a GPS and an IMU are integrated to estimate the odometry of the system, and subsequently, the point clouds measured from a 3D Lidar are registered by using this information. Furthermore, we use the latest CNN-based semantic segmentation to obtain semantic information on the surrounding environment. To integrate the point cloud with semantic information, we developed incremental semantic labeling including coordinate alignment, error minimization, and semantic information fusion. Additionally, to improve the quality of the generated semantic map, the map refinement is processed in a batch. It enhances the spatial distribution of labels and removes traces produced by moving vehicles effectively. We conduct experiments on challenging sequences to demonstrate that our algorithm outperforms state-of-the-art methods in terms of accuracy and intersection over union. |
format | Online Article Text |
id | pubmed-6111277 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2018 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-61112772018-08-30 Towards a Meaningful 3D Map Using a 3D Lidar and a Camera Jeong, Jongmin Yoon, Tae Sung Park, Jin Bae Sensors (Basel) Article Semantic 3D maps are required for various applications including robot navigation and surveying, and their importance has significantly increased. Generally, existing studies on semantic mapping were camera-based approaches that could not be operated in large-scale environments owing to their computational burden. Recently, a method of combining a 3D Lidar with a camera was introduced to address this problem, and a 3D Lidar and a camera were also utilized for semantic 3D mapping. In this study, our algorithm consists of semantic mapping and map refinement. In the semantic mapping, a GPS and an IMU are integrated to estimate the odometry of the system, and subsequently, the point clouds measured from a 3D Lidar are registered by using this information. Furthermore, we use the latest CNN-based semantic segmentation to obtain semantic information on the surrounding environment. To integrate the point cloud with semantic information, we developed incremental semantic labeling including coordinate alignment, error minimization, and semantic information fusion. Additionally, to improve the quality of the generated semantic map, the map refinement is processed in a batch. It enhances the spatial distribution of labels and removes traces produced by moving vehicles effectively. We conduct experiments on challenging sequences to demonstrate that our algorithm outperforms state-of-the-art methods in terms of accuracy and intersection over union. MDPI 2018-08-06 /pmc/articles/PMC6111277/ /pubmed/30082618 http://dx.doi.org/10.3390/s18082571 Text en © 2018 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Jeong, Jongmin Yoon, Tae Sung Park, Jin Bae Towards a Meaningful 3D Map Using a 3D Lidar and a Camera |
title | Towards a Meaningful 3D Map Using a 3D Lidar and a Camera |
title_full | Towards a Meaningful 3D Map Using a 3D Lidar and a Camera |
title_fullStr | Towards a Meaningful 3D Map Using a 3D Lidar and a Camera |
title_full_unstemmed | Towards a Meaningful 3D Map Using a 3D Lidar and a Camera |
title_short | Towards a Meaningful 3D Map Using a 3D Lidar and a Camera |
title_sort | towards a meaningful 3d map using a 3d lidar and a camera |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6111277/ https://www.ncbi.nlm.nih.gov/pubmed/30082618 http://dx.doi.org/10.3390/s18082571 |
work_keys_str_mv | AT jeongjongmin towardsameaningful3dmapusinga3dlidarandacamera AT yoontaesung towardsameaningful3dmapusinga3dlidarandacamera AT parkjinbae towardsameaningful3dmapusinga3dlidarandacamera |