Cargando…

Guided Direct Time-of-Flight Lidar Using Stereo Cameras for Enhanced Laser Power Efficiency

Self-driving vehicles demand efficient and reliable depth-sensing technologies. Lidar, with its capability for long-distance, high-precision measurement, is a crucial component in this pursuit. However, conventional mechanical scanning implementations suffer from reliability, cost, and frame rate li...

Descripción completa

Detalles Bibliográficos
Autores principales: Taneski, Filip, Gyongy, Istvan, Al Abbas, Tarek, Henderson, Robert K.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10650695/
https://www.ncbi.nlm.nih.gov/pubmed/37960642
http://dx.doi.org/10.3390/s23218943
_version_ 1785135839619055616
author Taneski, Filip
Gyongy, Istvan
Al Abbas, Tarek
Henderson, Robert K.
author_facet Taneski, Filip
Gyongy, Istvan
Al Abbas, Tarek
Henderson, Robert K.
author_sort Taneski, Filip
collection PubMed
description Self-driving vehicles demand efficient and reliable depth-sensing technologies. Lidar, with its capability for long-distance, high-precision measurement, is a crucial component in this pursuit. However, conventional mechanical scanning implementations suffer from reliability, cost, and frame rate limitations. Solid-state lidar solutions have emerged as a promising alternative, but the vast amount of photon data processed and stored using conventional direct time-of-flight (dToF) prevents long-distance sensing unless power-intensive partial histogram approaches are used. In this paper, we introduce a groundbreaking ‘guided’ dToF approach, harnessing external guidance from other onboard sensors to narrow down the depth search space for a power and data-efficient solution. This approach centers around a dToF sensor in which the exposed time window of independent pixels can be dynamically adjusted. We utilize a 64-by-32 macropixel dToF sensor and a pair of vision cameras to provide the guiding depth estimates. Our demonstrator captures a dynamic outdoor scene at 3 fps with distances up to 75 m. Compared to a conventional full histogram approach, on-chip data is reduced by over twenty times, while the total laser cycles in each frame are reduced by at least six times compared to any partial histogram approach. The capability of guided dToF to mitigate multipath reflections is also demonstrated. For self-driving vehicles where a wealth of sensor data is already available, guided dToF opens new possibilities for efficient solid-state lidar.
format Online
Article
Text
id pubmed-10650695
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-106506952023-11-03 Guided Direct Time-of-Flight Lidar Using Stereo Cameras for Enhanced Laser Power Efficiency Taneski, Filip Gyongy, Istvan Al Abbas, Tarek Henderson, Robert K. Sensors (Basel) Article Self-driving vehicles demand efficient and reliable depth-sensing technologies. Lidar, with its capability for long-distance, high-precision measurement, is a crucial component in this pursuit. However, conventional mechanical scanning implementations suffer from reliability, cost, and frame rate limitations. Solid-state lidar solutions have emerged as a promising alternative, but the vast amount of photon data processed and stored using conventional direct time-of-flight (dToF) prevents long-distance sensing unless power-intensive partial histogram approaches are used. In this paper, we introduce a groundbreaking ‘guided’ dToF approach, harnessing external guidance from other onboard sensors to narrow down the depth search space for a power and data-efficient solution. This approach centers around a dToF sensor in which the exposed time window of independent pixels can be dynamically adjusted. We utilize a 64-by-32 macropixel dToF sensor and a pair of vision cameras to provide the guiding depth estimates. Our demonstrator captures a dynamic outdoor scene at 3 fps with distances up to 75 m. Compared to a conventional full histogram approach, on-chip data is reduced by over twenty times, while the total laser cycles in each frame are reduced by at least six times compared to any partial histogram approach. The capability of guided dToF to mitigate multipath reflections is also demonstrated. For self-driving vehicles where a wealth of sensor data is already available, guided dToF opens new possibilities for efficient solid-state lidar. MDPI 2023-11-03 /pmc/articles/PMC10650695/ /pubmed/37960642 http://dx.doi.org/10.3390/s23218943 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Taneski, Filip
Gyongy, Istvan
Al Abbas, Tarek
Henderson, Robert K.
Guided Direct Time-of-Flight Lidar Using Stereo Cameras for Enhanced Laser Power Efficiency
title Guided Direct Time-of-Flight Lidar Using Stereo Cameras for Enhanced Laser Power Efficiency
title_full Guided Direct Time-of-Flight Lidar Using Stereo Cameras for Enhanced Laser Power Efficiency
title_fullStr Guided Direct Time-of-Flight Lidar Using Stereo Cameras for Enhanced Laser Power Efficiency
title_full_unstemmed Guided Direct Time-of-Flight Lidar Using Stereo Cameras for Enhanced Laser Power Efficiency
title_short Guided Direct Time-of-Flight Lidar Using Stereo Cameras for Enhanced Laser Power Efficiency
title_sort guided direct time-of-flight lidar using stereo cameras for enhanced laser power efficiency
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10650695/
https://www.ncbi.nlm.nih.gov/pubmed/37960642
http://dx.doi.org/10.3390/s23218943
work_keys_str_mv AT taneskifilip guideddirecttimeofflightlidarusingstereocamerasforenhancedlaserpowerefficiency
AT gyongyistvan guideddirecttimeofflightlidarusingstereocamerasforenhancedlaserpowerefficiency
AT alabbastarek guideddirecttimeofflightlidarusingstereocamerasforenhancedlaserpowerefficiency
AT hendersonrobertk guideddirecttimeofflightlidarusingstereocamerasforenhancedlaserpowerefficiency