Cargando…
Depth from a Motion Algorithm and a Hardware Architecture for Smart Cameras
Applications such as autonomous navigation, robot vision, and autonomous flying require depth map information of a scene. Depth can be estimated by using a single moving camera (depth from motion). However, the traditional depth from motion algorithms have low processing speeds and high hardware req...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2018
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6338951/ https://www.ncbi.nlm.nih.gov/pubmed/30583606 http://dx.doi.org/10.3390/s19010053 |
_version_ | 1783388523859017728 |
---|---|
author | Aguilar-González, Abiel Arias-Estrada, Miguel Berry, François |
author_facet | Aguilar-González, Abiel Arias-Estrada, Miguel Berry, François |
author_sort | Aguilar-González, Abiel |
collection | PubMed |
description | Applications such as autonomous navigation, robot vision, and autonomous flying require depth map information of a scene. Depth can be estimated by using a single moving camera (depth from motion). However, the traditional depth from motion algorithms have low processing speeds and high hardware requirements that limit the embedded capabilities. In this work, we propose a hardware architecture for depth from motion that consists of a flow/depth transformation and a new optical flow algorithm. Our optical flow formulation consists in an extension of the stereo matching problem. A pixel-parallel/window-parallel approach where a correlation function based on the sum of absolute difference (SAD) computes the optical flow is proposed. Further, in order to improve the SAD, the curl of the intensity gradient as a preprocessing step is proposed. Experimental results demonstrated that it is possible to reach higher accuracy (90% of accuracy) compared with previous Field Programmable Gate Array (FPGA)-based optical flow algorithms. For the depth estimation, our algorithm delivers dense maps with motion and depth information on all image pixels, with a processing speed up to 128 times faster than that of previous work, making it possible to achieve high performance in the context of embedded applications. |
format | Online Article Text |
id | pubmed-6338951 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2018 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-63389512019-01-23 Depth from a Motion Algorithm and a Hardware Architecture for Smart Cameras Aguilar-González, Abiel Arias-Estrada, Miguel Berry, François Sensors (Basel) Article Applications such as autonomous navigation, robot vision, and autonomous flying require depth map information of a scene. Depth can be estimated by using a single moving camera (depth from motion). However, the traditional depth from motion algorithms have low processing speeds and high hardware requirements that limit the embedded capabilities. In this work, we propose a hardware architecture for depth from motion that consists of a flow/depth transformation and a new optical flow algorithm. Our optical flow formulation consists in an extension of the stereo matching problem. A pixel-parallel/window-parallel approach where a correlation function based on the sum of absolute difference (SAD) computes the optical flow is proposed. Further, in order to improve the SAD, the curl of the intensity gradient as a preprocessing step is proposed. Experimental results demonstrated that it is possible to reach higher accuracy (90% of accuracy) compared with previous Field Programmable Gate Array (FPGA)-based optical flow algorithms. For the depth estimation, our algorithm delivers dense maps with motion and depth information on all image pixels, with a processing speed up to 128 times faster than that of previous work, making it possible to achieve high performance in the context of embedded applications. MDPI 2018-12-23 /pmc/articles/PMC6338951/ /pubmed/30583606 http://dx.doi.org/10.3390/s19010053 Text en © 2018 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Aguilar-González, Abiel Arias-Estrada, Miguel Berry, François Depth from a Motion Algorithm and a Hardware Architecture for Smart Cameras |
title | Depth from a Motion Algorithm and a Hardware Architecture for Smart Cameras |
title_full | Depth from a Motion Algorithm and a Hardware Architecture for Smart Cameras |
title_fullStr | Depth from a Motion Algorithm and a Hardware Architecture for Smart Cameras |
title_full_unstemmed | Depth from a Motion Algorithm and a Hardware Architecture for Smart Cameras |
title_short | Depth from a Motion Algorithm and a Hardware Architecture for Smart Cameras |
title_sort | depth from a motion algorithm and a hardware architecture for smart cameras |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6338951/ https://www.ncbi.nlm.nih.gov/pubmed/30583606 http://dx.doi.org/10.3390/s19010053 |
work_keys_str_mv | AT aguilargonzalezabiel depthfromamotionalgorithmandahardwarearchitectureforsmartcameras AT ariasestradamiguel depthfromamotionalgorithmandahardwarearchitectureforsmartcameras AT berryfrancois depthfromamotionalgorithmandahardwarearchitectureforsmartcameras |