Cargando…
Multiscale Hybrid Convolutional Deep Neural Networks with Channel Attention
Attention mechanisms can improve the performance of neural networks, but the recent attention networks bring a greater computational overhead while improving network performance. How to maintain model performance while reducing complexity is a hot research topic. In this paper, a lightweight Mixture...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9497715/ https://www.ncbi.nlm.nih.gov/pubmed/36141066 http://dx.doi.org/10.3390/e24091180 |
_version_ | 1784794574639595520 |
---|---|
author | Yang, Hua Yang, Ming He, Bitao Qin, Tao Yang, Jing |
author_facet | Yang, Hua Yang, Ming He, Bitao Qin, Tao Yang, Jing |
author_sort | Yang, Hua |
collection | PubMed |
description | Attention mechanisms can improve the performance of neural networks, but the recent attention networks bring a greater computational overhead while improving network performance. How to maintain model performance while reducing complexity is a hot research topic. In this paper, a lightweight Mixture Attention (MA) module is proposed to improve network performance and reduce the complexity of the model. Firstly, the MA module uses multi-branch architecture to process the input feature map in order to extract the multi-scale feature information of the input image. Secondly, in order to reduce the number of parameters, each branch uses group convolution independently, and the feature maps extracted by different branches are fused along the channel dimension. Finally, the fused feature maps are processed using the channel attention module to extract statistical information on the channels. The proposed method is efficient yet effective, e.g., the network parameters and computational cost are reduced by 9.86% and 7.83%, respectively, and the Top-1 performance is improved by 1.99% compared with ResNet50. Experimental results on common-used benchmarks, including CIFAR-10 for classification and PASCAL-VOC for object detection, demonstrate that the proposed MA outperforms the current SOTA methods significantly by achieving higher accuracy while having lower model complexity. |
format | Online Article Text |
id | pubmed-9497715 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-94977152022-09-23 Multiscale Hybrid Convolutional Deep Neural Networks with Channel Attention Yang, Hua Yang, Ming He, Bitao Qin, Tao Yang, Jing Entropy (Basel) Article Attention mechanisms can improve the performance of neural networks, but the recent attention networks bring a greater computational overhead while improving network performance. How to maintain model performance while reducing complexity is a hot research topic. In this paper, a lightweight Mixture Attention (MA) module is proposed to improve network performance and reduce the complexity of the model. Firstly, the MA module uses multi-branch architecture to process the input feature map in order to extract the multi-scale feature information of the input image. Secondly, in order to reduce the number of parameters, each branch uses group convolution independently, and the feature maps extracted by different branches are fused along the channel dimension. Finally, the fused feature maps are processed using the channel attention module to extract statistical information on the channels. The proposed method is efficient yet effective, e.g., the network parameters and computational cost are reduced by 9.86% and 7.83%, respectively, and the Top-1 performance is improved by 1.99% compared with ResNet50. Experimental results on common-used benchmarks, including CIFAR-10 for classification and PASCAL-VOC for object detection, demonstrate that the proposed MA outperforms the current SOTA methods significantly by achieving higher accuracy while having lower model complexity. MDPI 2022-08-24 /pmc/articles/PMC9497715/ /pubmed/36141066 http://dx.doi.org/10.3390/e24091180 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Yang, Hua Yang, Ming He, Bitao Qin, Tao Yang, Jing Multiscale Hybrid Convolutional Deep Neural Networks with Channel Attention |
title | Multiscale Hybrid Convolutional Deep Neural Networks with Channel Attention |
title_full | Multiscale Hybrid Convolutional Deep Neural Networks with Channel Attention |
title_fullStr | Multiscale Hybrid Convolutional Deep Neural Networks with Channel Attention |
title_full_unstemmed | Multiscale Hybrid Convolutional Deep Neural Networks with Channel Attention |
title_short | Multiscale Hybrid Convolutional Deep Neural Networks with Channel Attention |
title_sort | multiscale hybrid convolutional deep neural networks with channel attention |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9497715/ https://www.ncbi.nlm.nih.gov/pubmed/36141066 http://dx.doi.org/10.3390/e24091180 |
work_keys_str_mv | AT yanghua multiscalehybridconvolutionaldeepneuralnetworkswithchannelattention AT yangming multiscalehybridconvolutionaldeepneuralnetworkswithchannelattention AT hebitao multiscalehybridconvolutionaldeepneuralnetworkswithchannelattention AT qintao multiscalehybridconvolutionaldeepneuralnetworkswithchannelattention AT yangjing multiscalehybridconvolutionaldeepneuralnetworkswithchannelattention |