Cargando…
Guided Depth Completion with Instance Segmentation Fusion in Autonomous Driving Applications
Pixel-level depth information is crucial to many applications, such as autonomous driving, robotics navigation, 3D scene reconstruction, and augmented reality. However, depth information, which is usually acquired by sensors such as LiDAR, is sparse. Depth completion is a process that predicts missi...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9781309/ https://www.ncbi.nlm.nih.gov/pubmed/36559946 http://dx.doi.org/10.3390/s22249578 |
_version_ | 1784857042507268096 |
---|---|
author | El-Yabroudi, Mohammad Z. Abdel-Qader, Ikhlas Bazuin, Bradley J. Abudayyeh, Osama Chabaan, Rakan C. |
author_facet | El-Yabroudi, Mohammad Z. Abdel-Qader, Ikhlas Bazuin, Bradley J. Abudayyeh, Osama Chabaan, Rakan C. |
author_sort | El-Yabroudi, Mohammad Z. |
collection | PubMed |
description | Pixel-level depth information is crucial to many applications, such as autonomous driving, robotics navigation, 3D scene reconstruction, and augmented reality. However, depth information, which is usually acquired by sensors such as LiDAR, is sparse. Depth completion is a process that predicts missing pixels’ depth information from a set of sparse depth measurements. Most of the ongoing research applies deep neural networks on the entire sparse depth map and camera scene without utilizing any information about the available objects, which results in more complex and resource-demanding networks. In this work, we propose to use image instance segmentation to detect objects of interest with pixel-level locations, along with sparse depth data, to support depth completion. The framework utilizes a two-branch encoder–decoder deep neural network. It fuses information about scene available objects, such as objects’ type and pixel-level location, LiDAR, and RGB camera, to predict dense accurate depth maps. Experimental results on the KITTI dataset showed faster training and improved prediction accuracy. The proposed method reaches a convergence state faster and surpasses the baseline model in all evaluation metrics. |
format | Online Article Text |
id | pubmed-9781309 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-97813092022-12-24 Guided Depth Completion with Instance Segmentation Fusion in Autonomous Driving Applications El-Yabroudi, Mohammad Z. Abdel-Qader, Ikhlas Bazuin, Bradley J. Abudayyeh, Osama Chabaan, Rakan C. Sensors (Basel) Article Pixel-level depth information is crucial to many applications, such as autonomous driving, robotics navigation, 3D scene reconstruction, and augmented reality. However, depth information, which is usually acquired by sensors such as LiDAR, is sparse. Depth completion is a process that predicts missing pixels’ depth information from a set of sparse depth measurements. Most of the ongoing research applies deep neural networks on the entire sparse depth map and camera scene without utilizing any information about the available objects, which results in more complex and resource-demanding networks. In this work, we propose to use image instance segmentation to detect objects of interest with pixel-level locations, along with sparse depth data, to support depth completion. The framework utilizes a two-branch encoder–decoder deep neural network. It fuses information about scene available objects, such as objects’ type and pixel-level location, LiDAR, and RGB camera, to predict dense accurate depth maps. Experimental results on the KITTI dataset showed faster training and improved prediction accuracy. The proposed method reaches a convergence state faster and surpasses the baseline model in all evaluation metrics. MDPI 2022-12-07 /pmc/articles/PMC9781309/ /pubmed/36559946 http://dx.doi.org/10.3390/s22249578 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article El-Yabroudi, Mohammad Z. Abdel-Qader, Ikhlas Bazuin, Bradley J. Abudayyeh, Osama Chabaan, Rakan C. Guided Depth Completion with Instance Segmentation Fusion in Autonomous Driving Applications |
title | Guided Depth Completion with Instance Segmentation Fusion in Autonomous Driving Applications |
title_full | Guided Depth Completion with Instance Segmentation Fusion in Autonomous Driving Applications |
title_fullStr | Guided Depth Completion with Instance Segmentation Fusion in Autonomous Driving Applications |
title_full_unstemmed | Guided Depth Completion with Instance Segmentation Fusion in Autonomous Driving Applications |
title_short | Guided Depth Completion with Instance Segmentation Fusion in Autonomous Driving Applications |
title_sort | guided depth completion with instance segmentation fusion in autonomous driving applications |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9781309/ https://www.ncbi.nlm.nih.gov/pubmed/36559946 http://dx.doi.org/10.3390/s22249578 |
work_keys_str_mv | AT elyabroudimohammadz guideddepthcompletionwithinstancesegmentationfusioninautonomousdrivingapplications AT abdelqaderikhlas guideddepthcompletionwithinstancesegmentationfusioninautonomousdrivingapplications AT bazuinbradleyj guideddepthcompletionwithinstancesegmentationfusioninautonomousdrivingapplications AT abudayyehosama guideddepthcompletionwithinstancesegmentationfusioninautonomousdrivingapplications AT chabaanrakanc guideddepthcompletionwithinstancesegmentationfusioninautonomousdrivingapplications |