Cargando…

Multiscale Hybrid Convolutional Deep Neural Networks with Channel Attention

Attention mechanisms can improve the performance of neural networks, but the recent attention networks bring a greater computational overhead while improving network performance. How to maintain model performance while reducing complexity is a hot research topic. In this paper, a lightweight Mixture...

Descripción completa

Detalles Bibliográficos
Autores principales: Yang, Hua, Yang, Ming, He, Bitao, Qin, Tao, Yang, Jing
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9497715/
https://www.ncbi.nlm.nih.gov/pubmed/36141066
http://dx.doi.org/10.3390/e24091180
_version_ 1784794574639595520
author Yang, Hua
Yang, Ming
He, Bitao
Qin, Tao
Yang, Jing
author_facet Yang, Hua
Yang, Ming
He, Bitao
Qin, Tao
Yang, Jing
author_sort Yang, Hua
collection PubMed
description Attention mechanisms can improve the performance of neural networks, but the recent attention networks bring a greater computational overhead while improving network performance. How to maintain model performance while reducing complexity is a hot research topic. In this paper, a lightweight Mixture Attention (MA) module is proposed to improve network performance and reduce the complexity of the model. Firstly, the MA module uses multi-branch architecture to process the input feature map in order to extract the multi-scale feature information of the input image. Secondly, in order to reduce the number of parameters, each branch uses group convolution independently, and the feature maps extracted by different branches are fused along the channel dimension. Finally, the fused feature maps are processed using the channel attention module to extract statistical information on the channels. The proposed method is efficient yet effective, e.g., the network parameters and computational cost are reduced by 9.86% and 7.83%, respectively, and the Top-1 performance is improved by 1.99% compared with ResNet50. Experimental results on common-used benchmarks, including CIFAR-10 for classification and PASCAL-VOC for object detection, demonstrate that the proposed MA outperforms the current SOTA methods significantly by achieving higher accuracy while having lower model complexity.
format Online
Article
Text
id pubmed-9497715
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-94977152022-09-23 Multiscale Hybrid Convolutional Deep Neural Networks with Channel Attention Yang, Hua Yang, Ming He, Bitao Qin, Tao Yang, Jing Entropy (Basel) Article Attention mechanisms can improve the performance of neural networks, but the recent attention networks bring a greater computational overhead while improving network performance. How to maintain model performance while reducing complexity is a hot research topic. In this paper, a lightweight Mixture Attention (MA) module is proposed to improve network performance and reduce the complexity of the model. Firstly, the MA module uses multi-branch architecture to process the input feature map in order to extract the multi-scale feature information of the input image. Secondly, in order to reduce the number of parameters, each branch uses group convolution independently, and the feature maps extracted by different branches are fused along the channel dimension. Finally, the fused feature maps are processed using the channel attention module to extract statistical information on the channels. The proposed method is efficient yet effective, e.g., the network parameters and computational cost are reduced by 9.86% and 7.83%, respectively, and the Top-1 performance is improved by 1.99% compared with ResNet50. Experimental results on common-used benchmarks, including CIFAR-10 for classification and PASCAL-VOC for object detection, demonstrate that the proposed MA outperforms the current SOTA methods significantly by achieving higher accuracy while having lower model complexity. MDPI 2022-08-24 /pmc/articles/PMC9497715/ /pubmed/36141066 http://dx.doi.org/10.3390/e24091180 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Yang, Hua
Yang, Ming
He, Bitao
Qin, Tao
Yang, Jing
Multiscale Hybrid Convolutional Deep Neural Networks with Channel Attention
title Multiscale Hybrid Convolutional Deep Neural Networks with Channel Attention
title_full Multiscale Hybrid Convolutional Deep Neural Networks with Channel Attention
title_fullStr Multiscale Hybrid Convolutional Deep Neural Networks with Channel Attention
title_full_unstemmed Multiscale Hybrid Convolutional Deep Neural Networks with Channel Attention
title_short Multiscale Hybrid Convolutional Deep Neural Networks with Channel Attention
title_sort multiscale hybrid convolutional deep neural networks with channel attention
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9497715/
https://www.ncbi.nlm.nih.gov/pubmed/36141066
http://dx.doi.org/10.3390/e24091180
work_keys_str_mv AT yanghua multiscalehybridconvolutionaldeepneuralnetworkswithchannelattention
AT yangming multiscalehybridconvolutionaldeepneuralnetworkswithchannelattention
AT hebitao multiscalehybridconvolutionaldeepneuralnetworkswithchannelattention
AT qintao multiscalehybridconvolutionaldeepneuralnetworkswithchannelattention
AT yangjing multiscalehybridconvolutionaldeepneuralnetworkswithchannelattention