Cargando…

One Spatio-Temporal Sharpening Attention Mechanism for Light-Weight YOLO Models Based on Sharpening Spatial Attention

Attention mechanisms have demonstrated great potential in improving the performance of deep convolutional neural networks (CNNs). However, many existing methods dedicate to developing channel or spatial attention modules for CNNs with lots of parameters, and complex attention modules inevitably affe...

Descripción completa

Detalles Bibliográficos
Autores principales: Xue, Mengfan, Chen, Minghao, Peng, Dongliang, Guo, Yunfei, Chen, Huajie
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8659721/
https://www.ncbi.nlm.nih.gov/pubmed/34883953
http://dx.doi.org/10.3390/s21237949
_version_ 1784613031376846848
author Xue, Mengfan
Chen, Minghao
Peng, Dongliang
Guo, Yunfei
Chen, Huajie
author_facet Xue, Mengfan
Chen, Minghao
Peng, Dongliang
Guo, Yunfei
Chen, Huajie
author_sort Xue, Mengfan
collection PubMed
description Attention mechanisms have demonstrated great potential in improving the performance of deep convolutional neural networks (CNNs). However, many existing methods dedicate to developing channel or spatial attention modules for CNNs with lots of parameters, and complex attention modules inevitably affect the performance of CNNs. During our experiments of embedding Convolutional Block Attention Module (CBAM) in light-weight model YOLOv5s, CBAM does influence the speed and increase model complexity while reduce the average precision, but Squeeze-and-Excitation (SE) has a positive impact in the model as part of CBAM. To replace the spatial attention module in CBAM and offer a suitable scheme of channel and spatial attention modules, this paper proposes one Spatio-temporal Sharpening Attention Mechanism (SSAM), which sequentially infers intermediate maps along channel attention module and Sharpening Spatial Attention (SSA) module. By introducing sharpening filter in spatial attention module, we propose SSA module with low complexity. We try to find a scheme to combine our SSA module with SE module or Efficient Channel Attention (ECA) module and show best improvement in models such as YOLOv5s and YOLOv3-tiny. Therefore, we perform various replacement experiments and offer one best scheme that is to embed channel attention modules in backbone and neck of the model and integrate SSAM into YOLO head. We verify the positive effect of our SSAM on two general object detection datasets VOC2012 and MS COCO2017. One for obtaining a suitable scheme and the other for proving the versatility of our method in complex scenes. Experimental results on the two datasets show obvious promotion in terms of average precision and detection performance, which demonstrates the usefulness of our SSAM in light-weight YOLO models. Furthermore, visualization results also show the advantage of enhancing positioning ability with our SSAM.
format Online
Article
Text
id pubmed-8659721
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-86597212021-12-10 One Spatio-Temporal Sharpening Attention Mechanism for Light-Weight YOLO Models Based on Sharpening Spatial Attention Xue, Mengfan Chen, Minghao Peng, Dongliang Guo, Yunfei Chen, Huajie Sensors (Basel) Article Attention mechanisms have demonstrated great potential in improving the performance of deep convolutional neural networks (CNNs). However, many existing methods dedicate to developing channel or spatial attention modules for CNNs with lots of parameters, and complex attention modules inevitably affect the performance of CNNs. During our experiments of embedding Convolutional Block Attention Module (CBAM) in light-weight model YOLOv5s, CBAM does influence the speed and increase model complexity while reduce the average precision, but Squeeze-and-Excitation (SE) has a positive impact in the model as part of CBAM. To replace the spatial attention module in CBAM and offer a suitable scheme of channel and spatial attention modules, this paper proposes one Spatio-temporal Sharpening Attention Mechanism (SSAM), which sequentially infers intermediate maps along channel attention module and Sharpening Spatial Attention (SSA) module. By introducing sharpening filter in spatial attention module, we propose SSA module with low complexity. We try to find a scheme to combine our SSA module with SE module or Efficient Channel Attention (ECA) module and show best improvement in models such as YOLOv5s and YOLOv3-tiny. Therefore, we perform various replacement experiments and offer one best scheme that is to embed channel attention modules in backbone and neck of the model and integrate SSAM into YOLO head. We verify the positive effect of our SSAM on two general object detection datasets VOC2012 and MS COCO2017. One for obtaining a suitable scheme and the other for proving the versatility of our method in complex scenes. Experimental results on the two datasets show obvious promotion in terms of average precision and detection performance, which demonstrates the usefulness of our SSAM in light-weight YOLO models. Furthermore, visualization results also show the advantage of enhancing positioning ability with our SSAM. MDPI 2021-11-28 /pmc/articles/PMC8659721/ /pubmed/34883953 http://dx.doi.org/10.3390/s21237949 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Xue, Mengfan
Chen, Minghao
Peng, Dongliang
Guo, Yunfei
Chen, Huajie
One Spatio-Temporal Sharpening Attention Mechanism for Light-Weight YOLO Models Based on Sharpening Spatial Attention
title One Spatio-Temporal Sharpening Attention Mechanism for Light-Weight YOLO Models Based on Sharpening Spatial Attention
title_full One Spatio-Temporal Sharpening Attention Mechanism for Light-Weight YOLO Models Based on Sharpening Spatial Attention
title_fullStr One Spatio-Temporal Sharpening Attention Mechanism for Light-Weight YOLO Models Based on Sharpening Spatial Attention
title_full_unstemmed One Spatio-Temporal Sharpening Attention Mechanism for Light-Weight YOLO Models Based on Sharpening Spatial Attention
title_short One Spatio-Temporal Sharpening Attention Mechanism for Light-Weight YOLO Models Based on Sharpening Spatial Attention
title_sort one spatio-temporal sharpening attention mechanism for light-weight yolo models based on sharpening spatial attention
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8659721/
https://www.ncbi.nlm.nih.gov/pubmed/34883953
http://dx.doi.org/10.3390/s21237949
work_keys_str_mv AT xuemengfan onespatiotemporalsharpeningattentionmechanismforlightweightyolomodelsbasedonsharpeningspatialattention
AT chenminghao onespatiotemporalsharpeningattentionmechanismforlightweightyolomodelsbasedonsharpeningspatialattention
AT pengdongliang onespatiotemporalsharpeningattentionmechanismforlightweightyolomodelsbasedonsharpeningspatialattention
AT guoyunfei onespatiotemporalsharpeningattentionmechanismforlightweightyolomodelsbasedonsharpeningspatialattention
AT chenhuajie onespatiotemporalsharpeningattentionmechanismforlightweightyolomodelsbasedonsharpeningspatialattention