Cargando…

Guided Depth Completion with Instance Segmentation Fusion in Autonomous Driving Applications

Pixel-level depth information is crucial to many applications, such as autonomous driving, robotics navigation, 3D scene reconstruction, and augmented reality. However, depth information, which is usually acquired by sensors such as LiDAR, is sparse. Depth completion is a process that predicts missi...

Descripción completa

Detalles Bibliográficos
Autores principales: El-Yabroudi, Mohammad Z., Abdel-Qader, Ikhlas, Bazuin, Bradley J., Abudayyeh, Osama, Chabaan, Rakan C.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9781309/
https://www.ncbi.nlm.nih.gov/pubmed/36559946
http://dx.doi.org/10.3390/s22249578
_version_ 1784857042507268096
author El-Yabroudi, Mohammad Z.
Abdel-Qader, Ikhlas
Bazuin, Bradley J.
Abudayyeh, Osama
Chabaan, Rakan C.
author_facet El-Yabroudi, Mohammad Z.
Abdel-Qader, Ikhlas
Bazuin, Bradley J.
Abudayyeh, Osama
Chabaan, Rakan C.
author_sort El-Yabroudi, Mohammad Z.
collection PubMed
description Pixel-level depth information is crucial to many applications, such as autonomous driving, robotics navigation, 3D scene reconstruction, and augmented reality. However, depth information, which is usually acquired by sensors such as LiDAR, is sparse. Depth completion is a process that predicts missing pixels’ depth information from a set of sparse depth measurements. Most of the ongoing research applies deep neural networks on the entire sparse depth map and camera scene without utilizing any information about the available objects, which results in more complex and resource-demanding networks. In this work, we propose to use image instance segmentation to detect objects of interest with pixel-level locations, along with sparse depth data, to support depth completion. The framework utilizes a two-branch encoder–decoder deep neural network. It fuses information about scene available objects, such as objects’ type and pixel-level location, LiDAR, and RGB camera, to predict dense accurate depth maps. Experimental results on the KITTI dataset showed faster training and improved prediction accuracy. The proposed method reaches a convergence state faster and surpasses the baseline model in all evaluation metrics.
format Online
Article
Text
id pubmed-9781309
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-97813092022-12-24 Guided Depth Completion with Instance Segmentation Fusion in Autonomous Driving Applications El-Yabroudi, Mohammad Z. Abdel-Qader, Ikhlas Bazuin, Bradley J. Abudayyeh, Osama Chabaan, Rakan C. Sensors (Basel) Article Pixel-level depth information is crucial to many applications, such as autonomous driving, robotics navigation, 3D scene reconstruction, and augmented reality. However, depth information, which is usually acquired by sensors such as LiDAR, is sparse. Depth completion is a process that predicts missing pixels’ depth information from a set of sparse depth measurements. Most of the ongoing research applies deep neural networks on the entire sparse depth map and camera scene without utilizing any information about the available objects, which results in more complex and resource-demanding networks. In this work, we propose to use image instance segmentation to detect objects of interest with pixel-level locations, along with sparse depth data, to support depth completion. The framework utilizes a two-branch encoder–decoder deep neural network. It fuses information about scene available objects, such as objects’ type and pixel-level location, LiDAR, and RGB camera, to predict dense accurate depth maps. Experimental results on the KITTI dataset showed faster training and improved prediction accuracy. The proposed method reaches a convergence state faster and surpasses the baseline model in all evaluation metrics. MDPI 2022-12-07 /pmc/articles/PMC9781309/ /pubmed/36559946 http://dx.doi.org/10.3390/s22249578 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
El-Yabroudi, Mohammad Z.
Abdel-Qader, Ikhlas
Bazuin, Bradley J.
Abudayyeh, Osama
Chabaan, Rakan C.
Guided Depth Completion with Instance Segmentation Fusion in Autonomous Driving Applications
title Guided Depth Completion with Instance Segmentation Fusion in Autonomous Driving Applications
title_full Guided Depth Completion with Instance Segmentation Fusion in Autonomous Driving Applications
title_fullStr Guided Depth Completion with Instance Segmentation Fusion in Autonomous Driving Applications
title_full_unstemmed Guided Depth Completion with Instance Segmentation Fusion in Autonomous Driving Applications
title_short Guided Depth Completion with Instance Segmentation Fusion in Autonomous Driving Applications
title_sort guided depth completion with instance segmentation fusion in autonomous driving applications
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9781309/
https://www.ncbi.nlm.nih.gov/pubmed/36559946
http://dx.doi.org/10.3390/s22249578
work_keys_str_mv AT elyabroudimohammadz guideddepthcompletionwithinstancesegmentationfusioninautonomousdrivingapplications
AT abdelqaderikhlas guideddepthcompletionwithinstancesegmentationfusioninautonomousdrivingapplications
AT bazuinbradleyj guideddepthcompletionwithinstancesegmentationfusioninautonomousdrivingapplications
AT abudayyehosama guideddepthcompletionwithinstancesegmentationfusioninautonomousdrivingapplications
AT chabaanrakanc guideddepthcompletionwithinstancesegmentationfusioninautonomousdrivingapplications