Cargando…
Enhanced Perception for Autonomous Driving Using Semantic and Geometric Data Fusion
Environment perception remains one of the key tasks in autonomous driving for which solutions have yet to reach maturity. Multi-modal approaches benefit from the complementary physical properties specific to each sensor technology used, boosting overall performance. The added complexity brought on b...
Autores principales: | , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9269728/ https://www.ncbi.nlm.nih.gov/pubmed/35808555 http://dx.doi.org/10.3390/s22135061 |
_version_ | 1784744291883548672 |
---|---|
author | Florea, Horatiu Petrovai, Andra Giosan, Ion Oniga, Florin Varga, Robert Nedevschi, Sergiu |
author_facet | Florea, Horatiu Petrovai, Andra Giosan, Ion Oniga, Florin Varga, Robert Nedevschi, Sergiu |
author_sort | Florea, Horatiu |
collection | PubMed |
description | Environment perception remains one of the key tasks in autonomous driving for which solutions have yet to reach maturity. Multi-modal approaches benefit from the complementary physical properties specific to each sensor technology used, boosting overall performance. The added complexity brought on by data fusion processes is not trivial to solve, with design decisions heavily influencing the balance between quality and latency of the results. In this paper we present our novel real-time, 360 [Formula: see text] enhanced perception component based on low-level fusion between geometry provided by the LiDAR-based 3D point clouds and semantic scene information obtained from multiple RGB cameras, of multiple types. This multi-modal, multi-sensor scheme enables better range coverage, improved detection and classification quality with increased robustness. Semantic, instance and panoptic segmentations of 2D data are computed using efficient deep-learning-based algorithms, while 3D point clouds are segmented using a fast, traditional voxel-based solution. Finally, the fusion obtained through point-to-image projection yields a semantically enhanced 3D point cloud that allows enhanced perception through 3D detection refinement and 3D object classification. The planning and control systems of the vehicle receives the individual sensors’ perception together with the enhanced one, as well as the semantically enhanced 3D points. The developed perception solutions are successfully integrated onto an autonomous vehicle software stack, as part of the UP-Drive project. |
format | Online Article Text |
id | pubmed-9269728 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-92697282022-07-09 Enhanced Perception for Autonomous Driving Using Semantic and Geometric Data Fusion Florea, Horatiu Petrovai, Andra Giosan, Ion Oniga, Florin Varga, Robert Nedevschi, Sergiu Sensors (Basel) Article Environment perception remains one of the key tasks in autonomous driving for which solutions have yet to reach maturity. Multi-modal approaches benefit from the complementary physical properties specific to each sensor technology used, boosting overall performance. The added complexity brought on by data fusion processes is not trivial to solve, with design decisions heavily influencing the balance between quality and latency of the results. In this paper we present our novel real-time, 360 [Formula: see text] enhanced perception component based on low-level fusion between geometry provided by the LiDAR-based 3D point clouds and semantic scene information obtained from multiple RGB cameras, of multiple types. This multi-modal, multi-sensor scheme enables better range coverage, improved detection and classification quality with increased robustness. Semantic, instance and panoptic segmentations of 2D data are computed using efficient deep-learning-based algorithms, while 3D point clouds are segmented using a fast, traditional voxel-based solution. Finally, the fusion obtained through point-to-image projection yields a semantically enhanced 3D point cloud that allows enhanced perception through 3D detection refinement and 3D object classification. The planning and control systems of the vehicle receives the individual sensors’ perception together with the enhanced one, as well as the semantically enhanced 3D points. The developed perception solutions are successfully integrated onto an autonomous vehicle software stack, as part of the UP-Drive project. MDPI 2022-07-05 /pmc/articles/PMC9269728/ /pubmed/35808555 http://dx.doi.org/10.3390/s22135061 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Florea, Horatiu Petrovai, Andra Giosan, Ion Oniga, Florin Varga, Robert Nedevschi, Sergiu Enhanced Perception for Autonomous Driving Using Semantic and Geometric Data Fusion |
title | Enhanced Perception for Autonomous Driving Using Semantic and Geometric Data Fusion |
title_full | Enhanced Perception for Autonomous Driving Using Semantic and Geometric Data Fusion |
title_fullStr | Enhanced Perception for Autonomous Driving Using Semantic and Geometric Data Fusion |
title_full_unstemmed | Enhanced Perception for Autonomous Driving Using Semantic and Geometric Data Fusion |
title_short | Enhanced Perception for Autonomous Driving Using Semantic and Geometric Data Fusion |
title_sort | enhanced perception for autonomous driving using semantic and geometric data fusion |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9269728/ https://www.ncbi.nlm.nih.gov/pubmed/35808555 http://dx.doi.org/10.3390/s22135061 |
work_keys_str_mv | AT floreahoratiu enhancedperceptionforautonomousdrivingusingsemanticandgeometricdatafusion AT petrovaiandra enhancedperceptionforautonomousdrivingusingsemanticandgeometricdatafusion AT giosanion enhancedperceptionforautonomousdrivingusingsemanticandgeometricdatafusion AT onigaflorin enhancedperceptionforautonomousdrivingusingsemanticandgeometricdatafusion AT vargarobert enhancedperceptionforautonomousdrivingusingsemanticandgeometricdatafusion AT nedevschisergiu enhancedperceptionforautonomousdrivingusingsemanticandgeometricdatafusion |