Cargando…
RD-YOLO: An Effective and Efficient Object Detector for Roadside Perception System
In recent years, intelligent driving technology based on vehicle–road cooperation has gradually become a research hotspot in the field of intelligent transportation. There are many studies regarding vehicle perception, but fewer studies regarding roadside perception. As sensors are installed at diff...
Autores principales: | , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9658208/ https://www.ncbi.nlm.nih.gov/pubmed/36365793 http://dx.doi.org/10.3390/s22218097 |
_version_ | 1784829894006407168 |
---|---|
author | Huang, Lei Huang, Wenzhun |
author_facet | Huang, Lei Huang, Wenzhun |
author_sort | Huang, Lei |
collection | PubMed |
description | In recent years, intelligent driving technology based on vehicle–road cooperation has gradually become a research hotspot in the field of intelligent transportation. There are many studies regarding vehicle perception, but fewer studies regarding roadside perception. As sensors are installed at different heights, the roadside object scale varies violently, which burdens the optimization of networks. Moreover, there is a large amount of overlapping and occlusion in complex road environments, which leads to a great challenge of object distinction. To solve the two problems raised above, we propose RD-YOLO. Based on YOLOv5s, we reconstructed the feature fusion layer to increase effective feature extraction and improve the detection capability of small targets. Then, we replaced the original pyramid network with a generalized feature pyramid network (GFPN) to improve the adaptability of the network to different scale features. We also integrated a coordinate attention (CA) mechanism to find attention regions in scenarios with dense objects. Finally, we replaced the original Loss with Focal-EIOU Loss to improve the speed of the bounding box regression and the positioning accuracy of the anchor box. Compared to the YOLOv5s, the RD-YOLO improves the mean average precision (mAP) by 5.5% on the Rope3D dataset and 2.9% on the UA-DETRAC dataset. Meanwhile, by modifying the feature fusion layer, the weight of RD-YOLO is decreased by 55.9% while the detection speed is almost unchanged. Nevertheless, the proposed algorithm is capable of real-time detection at faster than 71.9 frames/s (FPS) and achieves higher accuracy than the previous approaches with a similar FPS. |
format | Online Article Text |
id | pubmed-9658208 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-96582082022-11-15 RD-YOLO: An Effective and Efficient Object Detector for Roadside Perception System Huang, Lei Huang, Wenzhun Sensors (Basel) Article In recent years, intelligent driving technology based on vehicle–road cooperation has gradually become a research hotspot in the field of intelligent transportation. There are many studies regarding vehicle perception, but fewer studies regarding roadside perception. As sensors are installed at different heights, the roadside object scale varies violently, which burdens the optimization of networks. Moreover, there is a large amount of overlapping and occlusion in complex road environments, which leads to a great challenge of object distinction. To solve the two problems raised above, we propose RD-YOLO. Based on YOLOv5s, we reconstructed the feature fusion layer to increase effective feature extraction and improve the detection capability of small targets. Then, we replaced the original pyramid network with a generalized feature pyramid network (GFPN) to improve the adaptability of the network to different scale features. We also integrated a coordinate attention (CA) mechanism to find attention regions in scenarios with dense objects. Finally, we replaced the original Loss with Focal-EIOU Loss to improve the speed of the bounding box regression and the positioning accuracy of the anchor box. Compared to the YOLOv5s, the RD-YOLO improves the mean average precision (mAP) by 5.5% on the Rope3D dataset and 2.9% on the UA-DETRAC dataset. Meanwhile, by modifying the feature fusion layer, the weight of RD-YOLO is decreased by 55.9% while the detection speed is almost unchanged. Nevertheless, the proposed algorithm is capable of real-time detection at faster than 71.9 frames/s (FPS) and achieves higher accuracy than the previous approaches with a similar FPS. MDPI 2022-10-22 /pmc/articles/PMC9658208/ /pubmed/36365793 http://dx.doi.org/10.3390/s22218097 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Huang, Lei Huang, Wenzhun RD-YOLO: An Effective and Efficient Object Detector for Roadside Perception System |
title | RD-YOLO: An Effective and Efficient Object Detector for Roadside Perception System |
title_full | RD-YOLO: An Effective and Efficient Object Detector for Roadside Perception System |
title_fullStr | RD-YOLO: An Effective and Efficient Object Detector for Roadside Perception System |
title_full_unstemmed | RD-YOLO: An Effective and Efficient Object Detector for Roadside Perception System |
title_short | RD-YOLO: An Effective and Efficient Object Detector for Roadside Perception System |
title_sort | rd-yolo: an effective and efficient object detector for roadside perception system |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9658208/ https://www.ncbi.nlm.nih.gov/pubmed/36365793 http://dx.doi.org/10.3390/s22218097 |
work_keys_str_mv | AT huanglei rdyoloaneffectiveandefficientobjectdetectorforroadsideperceptionsystem AT huangwenzhun rdyoloaneffectiveandefficientobjectdetectorforroadsideperceptionsystem |