Cargando…
A Deep Learning Framework for Accurate Vehicle Yaw Angle Estimation from a Monocular Camera Based on Part Arrangement
An accurate object pose is essential to assess its state and predict its movements. In recent years, scholars have often predicted object poses by matching an image with a virtual 3D model or by regressing the six-degree-of-freedom pose of the target directly from the pixel data via deep learning me...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9607309/ https://www.ncbi.nlm.nih.gov/pubmed/36298375 http://dx.doi.org/10.3390/s22208027 |
_version_ | 1784818511107850240 |
---|---|
author | Huang, Wenjun Li, Wenbo Tang, Luqi Zhu, Xiaoming Zou, Bin |
author_facet | Huang, Wenjun Li, Wenbo Tang, Luqi Zhu, Xiaoming Zou, Bin |
author_sort | Huang, Wenjun |
collection | PubMed |
description | An accurate object pose is essential to assess its state and predict its movements. In recent years, scholars have often predicted object poses by matching an image with a virtual 3D model or by regressing the six-degree-of-freedom pose of the target directly from the pixel data via deep learning methods. However, these approaches may ignore a fact that was proposed in the early days of computer vision research, i.e., that object parts are strongly represented in the object pose. In this study, we propose a novel and lightweight deep learning framework, YAEN (yaw angle estimation network), for accurate object yaw angle prediction from a monocular camera based on the arrangement of parts. YAEN uses an encoding–decoding structure for vehicle yaw angle prediction. The vehicle part arrangement information is extracted by the part-encoding network, and the yaw angle is extracted from vehicle part arrangement information by the yaw angle decoding network. Because vehicle part information is refined by the encoder, the decoding network structure is lightweight; the YAEN model has low hardware requirements and can reach a detection speed of [Formula: see text] on a [Formula: see text] graphics cards. To improve the performance of our model, we used asymmetric convolution and SSE (sum of squared errors) loss functions of adding the sign. To verify the effectiveness of this model, we constructed an accurate yaw angle dataset under real-world conditions with two vehicles equipped with high-precision positioning devices. Experimental results prove that our method can achieve satisfactory prediction performance in scenarios in which vehicles do not obscure each other, with an average prediction error of less than 3.1° and an accuracy of 96.45% for prediction errors of less than 10° in real driving scenarios. |
format | Online Article Text |
id | pubmed-9607309 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-96073092022-10-28 A Deep Learning Framework for Accurate Vehicle Yaw Angle Estimation from a Monocular Camera Based on Part Arrangement Huang, Wenjun Li, Wenbo Tang, Luqi Zhu, Xiaoming Zou, Bin Sensors (Basel) Article An accurate object pose is essential to assess its state and predict its movements. In recent years, scholars have often predicted object poses by matching an image with a virtual 3D model or by regressing the six-degree-of-freedom pose of the target directly from the pixel data via deep learning methods. However, these approaches may ignore a fact that was proposed in the early days of computer vision research, i.e., that object parts are strongly represented in the object pose. In this study, we propose a novel and lightweight deep learning framework, YAEN (yaw angle estimation network), for accurate object yaw angle prediction from a monocular camera based on the arrangement of parts. YAEN uses an encoding–decoding structure for vehicle yaw angle prediction. The vehicle part arrangement information is extracted by the part-encoding network, and the yaw angle is extracted from vehicle part arrangement information by the yaw angle decoding network. Because vehicle part information is refined by the encoder, the decoding network structure is lightweight; the YAEN model has low hardware requirements and can reach a detection speed of [Formula: see text] on a [Formula: see text] graphics cards. To improve the performance of our model, we used asymmetric convolution and SSE (sum of squared errors) loss functions of adding the sign. To verify the effectiveness of this model, we constructed an accurate yaw angle dataset under real-world conditions with two vehicles equipped with high-precision positioning devices. Experimental results prove that our method can achieve satisfactory prediction performance in scenarios in which vehicles do not obscure each other, with an average prediction error of less than 3.1° and an accuracy of 96.45% for prediction errors of less than 10° in real driving scenarios. MDPI 2022-10-20 /pmc/articles/PMC9607309/ /pubmed/36298375 http://dx.doi.org/10.3390/s22208027 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Huang, Wenjun Li, Wenbo Tang, Luqi Zhu, Xiaoming Zou, Bin A Deep Learning Framework for Accurate Vehicle Yaw Angle Estimation from a Monocular Camera Based on Part Arrangement |
title | A Deep Learning Framework for Accurate Vehicle Yaw Angle Estimation from a Monocular Camera Based on Part Arrangement |
title_full | A Deep Learning Framework for Accurate Vehicle Yaw Angle Estimation from a Monocular Camera Based on Part Arrangement |
title_fullStr | A Deep Learning Framework for Accurate Vehicle Yaw Angle Estimation from a Monocular Camera Based on Part Arrangement |
title_full_unstemmed | A Deep Learning Framework for Accurate Vehicle Yaw Angle Estimation from a Monocular Camera Based on Part Arrangement |
title_short | A Deep Learning Framework for Accurate Vehicle Yaw Angle Estimation from a Monocular Camera Based on Part Arrangement |
title_sort | deep learning framework for accurate vehicle yaw angle estimation from a monocular camera based on part arrangement |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9607309/ https://www.ncbi.nlm.nih.gov/pubmed/36298375 http://dx.doi.org/10.3390/s22208027 |
work_keys_str_mv | AT huangwenjun adeeplearningframeworkforaccuratevehicleyawangleestimationfromamonocularcamerabasedonpartarrangement AT liwenbo adeeplearningframeworkforaccuratevehicleyawangleestimationfromamonocularcamerabasedonpartarrangement AT tangluqi adeeplearningframeworkforaccuratevehicleyawangleestimationfromamonocularcamerabasedonpartarrangement AT zhuxiaoming adeeplearningframeworkforaccuratevehicleyawangleestimationfromamonocularcamerabasedonpartarrangement AT zoubin adeeplearningframeworkforaccuratevehicleyawangleestimationfromamonocularcamerabasedonpartarrangement AT huangwenjun deeplearningframeworkforaccuratevehicleyawangleestimationfromamonocularcamerabasedonpartarrangement AT liwenbo deeplearningframeworkforaccuratevehicleyawangleestimationfromamonocularcamerabasedonpartarrangement AT tangluqi deeplearningframeworkforaccuratevehicleyawangleestimationfromamonocularcamerabasedonpartarrangement AT zhuxiaoming deeplearningframeworkforaccuratevehicleyawangleestimationfromamonocularcamerabasedonpartarrangement AT zoubin deeplearningframeworkforaccuratevehicleyawangleestimationfromamonocularcamerabasedonpartarrangement |