Cargando…

Enhanced mechanisms of pooling and channel attention for deep learning feature maps

The pooling function is vital for deep neural networks (DNNs). The operation is to generalize the representation of feature maps and progressively cut down the spatial size of feature maps to optimize the computing consumption of the network. Furthermore, the function is also the basis for the compu...

Descripción completa

Detalles Bibliográficos
Autores principales: Li, Hengyi, Yue, Xuebin, Meng, Lin
Formato: Online Artículo Texto
Lenguaje:English
Publicado: PeerJ Inc. 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9748832/
https://www.ncbi.nlm.nih.gov/pubmed/36532804
http://dx.doi.org/10.7717/peerj-cs.1161
_version_ 1784849911839195136
author Li, Hengyi
Yue, Xuebin
Meng, Lin
author_facet Li, Hengyi
Yue, Xuebin
Meng, Lin
author_sort Li, Hengyi
collection PubMed
description The pooling function is vital for deep neural networks (DNNs). The operation is to generalize the representation of feature maps and progressively cut down the spatial size of feature maps to optimize the computing consumption of the network. Furthermore, the function is also the basis for the computer vision attention mechanism. However, as a matter of fact, pooling is a down-sampling operation, which makes the feature-map representation approximately to small translations with the summary statistic of adjacent pixels. As a result, the function inevitably leads to information loss more or less. In this article, we propose a fused max-average pooling (FMAPooling) operation as well as an improved channel attention mechanism (FMAttn) by utilizing the two pooling functions to enhance the feature representation for DNNs. Basically, the methods are to enhance multiple-level features extracted by max pooling and average pooling respectively. The effectiveness of the proposals is verified with VGG, ResNet, and MobileNetV2 architectures on CIFAR10/100 and ImageNet100. According to the experimental results, the FMAPooling brings up to 1.63% accuracy improvement compared with the baseline model; the FMAttn achieves up to 2.21% accuracy improvement compared with the previous channel attention mechanism. Furthermore, the proposals are extensible and could be embedded into various DNN models easily, or take the place of certain structures of DNNs. The computation burden introduced by the proposals is negligible.
format Online
Article
Text
id pubmed-9748832
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher PeerJ Inc.
record_format MEDLINE/PubMed
spelling pubmed-97488322022-12-15 Enhanced mechanisms of pooling and channel attention for deep learning feature maps Li, Hengyi Yue, Xuebin Meng, Lin PeerJ Comput Sci Artificial Intelligence The pooling function is vital for deep neural networks (DNNs). The operation is to generalize the representation of feature maps and progressively cut down the spatial size of feature maps to optimize the computing consumption of the network. Furthermore, the function is also the basis for the computer vision attention mechanism. However, as a matter of fact, pooling is a down-sampling operation, which makes the feature-map representation approximately to small translations with the summary statistic of adjacent pixels. As a result, the function inevitably leads to information loss more or less. In this article, we propose a fused max-average pooling (FMAPooling) operation as well as an improved channel attention mechanism (FMAttn) by utilizing the two pooling functions to enhance the feature representation for DNNs. Basically, the methods are to enhance multiple-level features extracted by max pooling and average pooling respectively. The effectiveness of the proposals is verified with VGG, ResNet, and MobileNetV2 architectures on CIFAR10/100 and ImageNet100. According to the experimental results, the FMAPooling brings up to 1.63% accuracy improvement compared with the baseline model; the FMAttn achieves up to 2.21% accuracy improvement compared with the previous channel attention mechanism. Furthermore, the proposals are extensible and could be embedded into various DNN models easily, or take the place of certain structures of DNNs. The computation burden introduced by the proposals is negligible. PeerJ Inc. 2022-11-21 /pmc/articles/PMC9748832/ /pubmed/36532804 http://dx.doi.org/10.7717/peerj-cs.1161 Text en © 2022 Li et al. https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, reproduction and adaptation in any medium and for any purpose provided that it is properly attributed. For attribution, the original author(s), title, publication source (PeerJ Computer Science) and either DOI or URL of the article must be cited.
spellingShingle Artificial Intelligence
Li, Hengyi
Yue, Xuebin
Meng, Lin
Enhanced mechanisms of pooling and channel attention for deep learning feature maps
title Enhanced mechanisms of pooling and channel attention for deep learning feature maps
title_full Enhanced mechanisms of pooling and channel attention for deep learning feature maps
title_fullStr Enhanced mechanisms of pooling and channel attention for deep learning feature maps
title_full_unstemmed Enhanced mechanisms of pooling and channel attention for deep learning feature maps
title_short Enhanced mechanisms of pooling and channel attention for deep learning feature maps
title_sort enhanced mechanisms of pooling and channel attention for deep learning feature maps
topic Artificial Intelligence
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9748832/
https://www.ncbi.nlm.nih.gov/pubmed/36532804
http://dx.doi.org/10.7717/peerj-cs.1161
work_keys_str_mv AT lihengyi enhancedmechanismsofpoolingandchannelattentionfordeeplearningfeaturemaps
AT yuexuebin enhancedmechanismsofpoolingandchannelattentionfordeeplearningfeaturemaps
AT menglin enhancedmechanismsofpoolingandchannelattentionfordeeplearningfeaturemaps