Cargando…

Alpha-SGANet: A multi-attention-scale feature pyramid network combined with lightweight network based on Alpha-IoU loss

The design of deep convolutional neural networks has resulted in significant advances and successes in the field of object detection. However, despite these achievements, the high computational and memory costs of such object detection networks on the edge or in mobile scenarios are one of the most...

Descripción completa

Detalles Bibliográficos
Autores principales: Li, Hong, Zhou, Qian, Mao, Yao, Zhang, Bing, Liu, Chao
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Public Library of Science 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9612525/
https://www.ncbi.nlm.nih.gov/pubmed/36301900
http://dx.doi.org/10.1371/journal.pone.0276581
_version_ 1784819791186362368
author Li, Hong
Zhou, Qian
Mao, Yao
Zhang, Bing
Liu, Chao
author_facet Li, Hong
Zhou, Qian
Mao, Yao
Zhang, Bing
Liu, Chao
author_sort Li, Hong
collection PubMed
description The design of deep convolutional neural networks has resulted in significant advances and successes in the field of object detection. However, despite these achievements, the high computational and memory costs of such object detection networks on the edge or in mobile scenarios are one of the most significant barriers to their broad adoption. To solve this problem, this paper introduces an improved lightweight real-time convolutional neural network based on YOLOv5, called Alpha-SGANet: A multi-attention-scale feature pyramid network combined with a lightweight network based on Alpha-IoU loss. Firstly, we add one more prediction head to detect different-scale objects, design a lightweight and efficient feature extraction network using ShuffleNetV2 in the backbone, and reduce information loss using the SPP module with a smaller convolutional nucleus. Then, cleverly, employ GAFPN to improve feature transition processing in the neck region, including the usage of the Ghost module to construct efficient feature maps to help prediction. The CBAM module was further integrated to find areas of interest in the scene; finally, combined with Alpha-IOU loss for model supervision training, the biggest performance improvement was achieved. The experiment results show that, compared with YOLOv5s, our proposed method can achieve higher accuracy with fewer parameters and has real-time speed through verification on the PASCAL VOC dataset and MS COCO dataset.
format Online
Article
Text
id pubmed-9612525
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Public Library of Science
record_format MEDLINE/PubMed
spelling pubmed-96125252022-10-28 Alpha-SGANet: A multi-attention-scale feature pyramid network combined with lightweight network based on Alpha-IoU loss Li, Hong Zhou, Qian Mao, Yao Zhang, Bing Liu, Chao PLoS One Research Article The design of deep convolutional neural networks has resulted in significant advances and successes in the field of object detection. However, despite these achievements, the high computational and memory costs of such object detection networks on the edge or in mobile scenarios are one of the most significant barriers to their broad adoption. To solve this problem, this paper introduces an improved lightweight real-time convolutional neural network based on YOLOv5, called Alpha-SGANet: A multi-attention-scale feature pyramid network combined with a lightweight network based on Alpha-IoU loss. Firstly, we add one more prediction head to detect different-scale objects, design a lightweight and efficient feature extraction network using ShuffleNetV2 in the backbone, and reduce information loss using the SPP module with a smaller convolutional nucleus. Then, cleverly, employ GAFPN to improve feature transition processing in the neck region, including the usage of the Ghost module to construct efficient feature maps to help prediction. The CBAM module was further integrated to find areas of interest in the scene; finally, combined with Alpha-IOU loss for model supervision training, the biggest performance improvement was achieved. The experiment results show that, compared with YOLOv5s, our proposed method can achieve higher accuracy with fewer parameters and has real-time speed through verification on the PASCAL VOC dataset and MS COCO dataset. Public Library of Science 2022-10-27 /pmc/articles/PMC9612525/ /pubmed/36301900 http://dx.doi.org/10.1371/journal.pone.0276581 Text en © 2022 Li et al https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
spellingShingle Research Article
Li, Hong
Zhou, Qian
Mao, Yao
Zhang, Bing
Liu, Chao
Alpha-SGANet: A multi-attention-scale feature pyramid network combined with lightweight network based on Alpha-IoU loss
title Alpha-SGANet: A multi-attention-scale feature pyramid network combined with lightweight network based on Alpha-IoU loss
title_full Alpha-SGANet: A multi-attention-scale feature pyramid network combined with lightweight network based on Alpha-IoU loss
title_fullStr Alpha-SGANet: A multi-attention-scale feature pyramid network combined with lightweight network based on Alpha-IoU loss
title_full_unstemmed Alpha-SGANet: A multi-attention-scale feature pyramid network combined with lightweight network based on Alpha-IoU loss
title_short Alpha-SGANet: A multi-attention-scale feature pyramid network combined with lightweight network based on Alpha-IoU loss
title_sort alpha-sganet: a multi-attention-scale feature pyramid network combined with lightweight network based on alpha-iou loss
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9612525/
https://www.ncbi.nlm.nih.gov/pubmed/36301900
http://dx.doi.org/10.1371/journal.pone.0276581
work_keys_str_mv AT lihong alphasganetamultiattentionscalefeaturepyramidnetworkcombinedwithlightweightnetworkbasedonalphaiouloss
AT zhouqian alphasganetamultiattentionscalefeaturepyramidnetworkcombinedwithlightweightnetworkbasedonalphaiouloss
AT maoyao alphasganetamultiattentionscalefeaturepyramidnetworkcombinedwithlightweightnetworkbasedonalphaiouloss
AT zhangbing alphasganetamultiattentionscalefeaturepyramidnetworkcombinedwithlightweightnetworkbasedonalphaiouloss
AT liuchao alphasganetamultiattentionscalefeaturepyramidnetworkcombinedwithlightweightnetworkbasedonalphaiouloss