Cargando…

Enhanced Perception for Autonomous Driving Using Semantic and Geometric Data Fusion

Environment perception remains one of the key tasks in autonomous driving for which solutions have yet to reach maturity. Multi-modal approaches benefit from the complementary physical properties specific to each sensor technology used, boosting overall performance. The added complexity brought on b...

Descripción completa

Detalles Bibliográficos
Autores principales: Florea, Horatiu, Petrovai, Andra, Giosan, Ion, Oniga, Florin, Varga, Robert, Nedevschi, Sergiu
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9269728/
https://www.ncbi.nlm.nih.gov/pubmed/35808555
http://dx.doi.org/10.3390/s22135061
_version_ 1784744291883548672
author Florea, Horatiu
Petrovai, Andra
Giosan, Ion
Oniga, Florin
Varga, Robert
Nedevschi, Sergiu
author_facet Florea, Horatiu
Petrovai, Andra
Giosan, Ion
Oniga, Florin
Varga, Robert
Nedevschi, Sergiu
author_sort Florea, Horatiu
collection PubMed
description Environment perception remains one of the key tasks in autonomous driving for which solutions have yet to reach maturity. Multi-modal approaches benefit from the complementary physical properties specific to each sensor technology used, boosting overall performance. The added complexity brought on by data fusion processes is not trivial to solve, with design decisions heavily influencing the balance between quality and latency of the results. In this paper we present our novel real-time, 360 [Formula: see text] enhanced perception component based on low-level fusion between geometry provided by the LiDAR-based 3D point clouds and semantic scene information obtained from multiple RGB cameras, of multiple types. This multi-modal, multi-sensor scheme enables better range coverage, improved detection and classification quality with increased robustness. Semantic, instance and panoptic segmentations of 2D data are computed using efficient deep-learning-based algorithms, while 3D point clouds are segmented using a fast, traditional voxel-based solution. Finally, the fusion obtained through point-to-image projection yields a semantically enhanced 3D point cloud that allows enhanced perception through 3D detection refinement and 3D object classification. The planning and control systems of the vehicle receives the individual sensors’ perception together with the enhanced one, as well as the semantically enhanced 3D points. The developed perception solutions are successfully integrated onto an autonomous vehicle software stack, as part of the UP-Drive project.
format Online
Article
Text
id pubmed-9269728
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-92697282022-07-09 Enhanced Perception for Autonomous Driving Using Semantic and Geometric Data Fusion Florea, Horatiu Petrovai, Andra Giosan, Ion Oniga, Florin Varga, Robert Nedevschi, Sergiu Sensors (Basel) Article Environment perception remains one of the key tasks in autonomous driving for which solutions have yet to reach maturity. Multi-modal approaches benefit from the complementary physical properties specific to each sensor technology used, boosting overall performance. The added complexity brought on by data fusion processes is not trivial to solve, with design decisions heavily influencing the balance between quality and latency of the results. In this paper we present our novel real-time, 360 [Formula: see text] enhanced perception component based on low-level fusion between geometry provided by the LiDAR-based 3D point clouds and semantic scene information obtained from multiple RGB cameras, of multiple types. This multi-modal, multi-sensor scheme enables better range coverage, improved detection and classification quality with increased robustness. Semantic, instance and panoptic segmentations of 2D data are computed using efficient deep-learning-based algorithms, while 3D point clouds are segmented using a fast, traditional voxel-based solution. Finally, the fusion obtained through point-to-image projection yields a semantically enhanced 3D point cloud that allows enhanced perception through 3D detection refinement and 3D object classification. The planning and control systems of the vehicle receives the individual sensors’ perception together with the enhanced one, as well as the semantically enhanced 3D points. The developed perception solutions are successfully integrated onto an autonomous vehicle software stack, as part of the UP-Drive project. MDPI 2022-07-05 /pmc/articles/PMC9269728/ /pubmed/35808555 http://dx.doi.org/10.3390/s22135061 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Florea, Horatiu
Petrovai, Andra
Giosan, Ion
Oniga, Florin
Varga, Robert
Nedevschi, Sergiu
Enhanced Perception for Autonomous Driving Using Semantic and Geometric Data Fusion
title Enhanced Perception for Autonomous Driving Using Semantic and Geometric Data Fusion
title_full Enhanced Perception for Autonomous Driving Using Semantic and Geometric Data Fusion
title_fullStr Enhanced Perception for Autonomous Driving Using Semantic and Geometric Data Fusion
title_full_unstemmed Enhanced Perception for Autonomous Driving Using Semantic and Geometric Data Fusion
title_short Enhanced Perception for Autonomous Driving Using Semantic and Geometric Data Fusion
title_sort enhanced perception for autonomous driving using semantic and geometric data fusion
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9269728/
https://www.ncbi.nlm.nih.gov/pubmed/35808555
http://dx.doi.org/10.3390/s22135061
work_keys_str_mv AT floreahoratiu enhancedperceptionforautonomousdrivingusingsemanticandgeometricdatafusion
AT petrovaiandra enhancedperceptionforautonomousdrivingusingsemanticandgeometricdatafusion
AT giosanion enhancedperceptionforautonomousdrivingusingsemanticandgeometricdatafusion
AT onigaflorin enhancedperceptionforautonomousdrivingusingsemanticandgeometricdatafusion
AT vargarobert enhancedperceptionforautonomousdrivingusingsemanticandgeometricdatafusion
AT nedevschisergiu enhancedperceptionforautonomousdrivingusingsemanticandgeometricdatafusion