Cargando…

Road User Position and Speed Estimation via Deep Learning from Calibrated Fisheye Videos

In this paper, we present a deep learning processing flow aimed at Advanced Driving Assistance Systems (ADASs) for urban road users. We use a fine analysis of the optical setup of a fisheye camera and present a detailed procedure to obtain Global Navigation Satellite System (GNSS) coordinates along...

Descripción completa

Detalles Bibliográficos
Autores principales: Berviller, Yves, Ansarnia, Masoomeh Shireen, Tisserand, Etienne, Schweitzer, Patrick, Tremeau, Alain
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10007371/
https://www.ncbi.nlm.nih.gov/pubmed/36904841
http://dx.doi.org/10.3390/s23052637
_version_ 1784905504506511360
author Berviller, Yves
Ansarnia, Masoomeh Shireen
Tisserand, Etienne
Schweitzer, Patrick
Tremeau, Alain
author_facet Berviller, Yves
Ansarnia, Masoomeh Shireen
Tisserand, Etienne
Schweitzer, Patrick
Tremeau, Alain
author_sort Berviller, Yves
collection PubMed
description In this paper, we present a deep learning processing flow aimed at Advanced Driving Assistance Systems (ADASs) for urban road users. We use a fine analysis of the optical setup of a fisheye camera and present a detailed procedure to obtain Global Navigation Satellite System (GNSS) coordinates along with the speed of the moving objects. The camera to world transform incorporates the lens distortion function. YOLOv4, re-trained with ortho-photographic fisheye images, provides road user detection. All the information extracted from the image by our system represents a small payload and can easily be broadcast to the road users. The results show that our system is able to properly classify and localize the detected objects in real time, even in low-light-illumination conditions. For an effective observation area of 20 m × 50 m, the error of the localization is in the order of one meter. Although an estimation of the velocities of the detected objects is carried out by offline processing with the FlowNet2 algorithm, the accuracy is quite good, with an error below one meter per second for urban speed range (0 to 15 m/s). Moreover, the almost ortho-photographic configuration of the imaging system ensures that the anonymity of all street users is guaranteed.
format Online
Article
Text
id pubmed-10007371
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-100073712023-03-12 Road User Position and Speed Estimation via Deep Learning from Calibrated Fisheye Videos Berviller, Yves Ansarnia, Masoomeh Shireen Tisserand, Etienne Schweitzer, Patrick Tremeau, Alain Sensors (Basel) Article In this paper, we present a deep learning processing flow aimed at Advanced Driving Assistance Systems (ADASs) for urban road users. We use a fine analysis of the optical setup of a fisheye camera and present a detailed procedure to obtain Global Navigation Satellite System (GNSS) coordinates along with the speed of the moving objects. The camera to world transform incorporates the lens distortion function. YOLOv4, re-trained with ortho-photographic fisheye images, provides road user detection. All the information extracted from the image by our system represents a small payload and can easily be broadcast to the road users. The results show that our system is able to properly classify and localize the detected objects in real time, even in low-light-illumination conditions. For an effective observation area of 20 m × 50 m, the error of the localization is in the order of one meter. Although an estimation of the velocities of the detected objects is carried out by offline processing with the FlowNet2 algorithm, the accuracy is quite good, with an error below one meter per second for urban speed range (0 to 15 m/s). Moreover, the almost ortho-photographic configuration of the imaging system ensures that the anonymity of all street users is guaranteed. MDPI 2023-02-27 /pmc/articles/PMC10007371/ /pubmed/36904841 http://dx.doi.org/10.3390/s23052637 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Berviller, Yves
Ansarnia, Masoomeh Shireen
Tisserand, Etienne
Schweitzer, Patrick
Tremeau, Alain
Road User Position and Speed Estimation via Deep Learning from Calibrated Fisheye Videos
title Road User Position and Speed Estimation via Deep Learning from Calibrated Fisheye Videos
title_full Road User Position and Speed Estimation via Deep Learning from Calibrated Fisheye Videos
title_fullStr Road User Position and Speed Estimation via Deep Learning from Calibrated Fisheye Videos
title_full_unstemmed Road User Position and Speed Estimation via Deep Learning from Calibrated Fisheye Videos
title_short Road User Position and Speed Estimation via Deep Learning from Calibrated Fisheye Videos
title_sort road user position and speed estimation via deep learning from calibrated fisheye videos
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10007371/
https://www.ncbi.nlm.nih.gov/pubmed/36904841
http://dx.doi.org/10.3390/s23052637
work_keys_str_mv AT bervilleryves roaduserpositionandspeedestimationviadeeplearningfromcalibratedfisheyevideos
AT ansarniamasoomehshireen roaduserpositionandspeedestimationviadeeplearningfromcalibratedfisheyevideos
AT tisserandetienne roaduserpositionandspeedestimationviadeeplearningfromcalibratedfisheyevideos
AT schweitzerpatrick roaduserpositionandspeedestimationviadeeplearningfromcalibratedfisheyevideos
AT tremeaualain roaduserpositionandspeedestimationviadeeplearningfromcalibratedfisheyevideos