Cargando…

Towards High Accuracy Pedestrian Detection on Edge GPUs

Despite the rapid development of pedestrian detection algorithms, the balance between detection accuracy and efficiency is still far from being achieved due to edge GPUs (low computing power) limiting the parameters of the model. To address this issue, we propose the YOLOv4-TP-Tiny based on the YOLO...

Descripción completa

Detalles Bibliográficos
Autores principales: Zhou, Huaping, Wu, Tao, Sun, Kelei, Zhang, Chunjiong
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9413106/
https://www.ncbi.nlm.nih.gov/pubmed/36015741
http://dx.doi.org/10.3390/s22165980
_version_ 1784775654888177664
author Zhou, Huaping
Wu, Tao
Sun, Kelei
Zhang, Chunjiong
author_facet Zhou, Huaping
Wu, Tao
Sun, Kelei
Zhang, Chunjiong
author_sort Zhou, Huaping
collection PubMed
description Despite the rapid development of pedestrian detection algorithms, the balance between detection accuracy and efficiency is still far from being achieved due to edge GPUs (low computing power) limiting the parameters of the model. To address this issue, we propose the YOLOv4-TP-Tiny based on the YOLOv4 model, which mainly includes two modules, two-dimensional attention (TA) and pedestrian-based feature extraction (PFM). First, we integrate the TA mechanism into the backbone network, which increases the attention of the network to the visible area of pedestrians and improves the accuracy of pedestrian detection. Then, the PFM is used to replace the original spatial pyramid pooling (SPP) structure in the YOLOv4 to obtain the YOLOv4-TP algorithm, which can adapt to different sizes of people to obtain higher detection accuracy. To maintain detection speed, we replaced the normal convolution with a ghost network with a TA mechanism, resulting in more feature maps with fewer parameters. We constructed a one-way multi-scale feature fusion structure to replace the down-sampling process, thereby reducing network parameters to obtain the YOLOv4-TP-Tiny model. The experimental results show that the YOLOv4-TP-tiny has 58.3% AP and 31 FPS in the winder person pedestrian dataset. With the same hardware conditions and dataset, the AP of the YOLOv4-tiny is 55.9%, and the FPS is 29.
format Online
Article
Text
id pubmed-9413106
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-94131062022-08-27 Towards High Accuracy Pedestrian Detection on Edge GPUs Zhou, Huaping Wu, Tao Sun, Kelei Zhang, Chunjiong Sensors (Basel) Article Despite the rapid development of pedestrian detection algorithms, the balance between detection accuracy and efficiency is still far from being achieved due to edge GPUs (low computing power) limiting the parameters of the model. To address this issue, we propose the YOLOv4-TP-Tiny based on the YOLOv4 model, which mainly includes two modules, two-dimensional attention (TA) and pedestrian-based feature extraction (PFM). First, we integrate the TA mechanism into the backbone network, which increases the attention of the network to the visible area of pedestrians and improves the accuracy of pedestrian detection. Then, the PFM is used to replace the original spatial pyramid pooling (SPP) structure in the YOLOv4 to obtain the YOLOv4-TP algorithm, which can adapt to different sizes of people to obtain higher detection accuracy. To maintain detection speed, we replaced the normal convolution with a ghost network with a TA mechanism, resulting in more feature maps with fewer parameters. We constructed a one-way multi-scale feature fusion structure to replace the down-sampling process, thereby reducing network parameters to obtain the YOLOv4-TP-Tiny model. The experimental results show that the YOLOv4-TP-tiny has 58.3% AP and 31 FPS in the winder person pedestrian dataset. With the same hardware conditions and dataset, the AP of the YOLOv4-tiny is 55.9%, and the FPS is 29. MDPI 2022-08-10 /pmc/articles/PMC9413106/ /pubmed/36015741 http://dx.doi.org/10.3390/s22165980 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Zhou, Huaping
Wu, Tao
Sun, Kelei
Zhang, Chunjiong
Towards High Accuracy Pedestrian Detection on Edge GPUs
title Towards High Accuracy Pedestrian Detection on Edge GPUs
title_full Towards High Accuracy Pedestrian Detection on Edge GPUs
title_fullStr Towards High Accuracy Pedestrian Detection on Edge GPUs
title_full_unstemmed Towards High Accuracy Pedestrian Detection on Edge GPUs
title_short Towards High Accuracy Pedestrian Detection on Edge GPUs
title_sort towards high accuracy pedestrian detection on edge gpus
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9413106/
https://www.ncbi.nlm.nih.gov/pubmed/36015741
http://dx.doi.org/10.3390/s22165980
work_keys_str_mv AT zhouhuaping towardshighaccuracypedestriandetectiononedgegpus
AT wutao towardshighaccuracypedestriandetectiononedgegpus
AT sunkelei towardshighaccuracypedestriandetectiononedgegpus
AT zhangchunjiong towardshighaccuracypedestriandetectiononedgegpus