Cargando…

A multi-scale gated multi-head attention depthwise separable CNN model for recognizing COVID-19

Coronavirus 2019 (COVID-19) is a new acute respiratory disease that has spread rapidly throughout the world. In this paper, a lightweight convolutional neural network (CNN) model named multi-scale gated multi-head attention depthwise separable CNN (MGMADS-CNN) is proposed, which is based on attentio...

Descripción completa

Detalles Bibliográficos
Autores principales: Hong, Geng, Chen, Xiaoyan, Chen, Jianyong, Zhang, Miao, Ren, Yumeng, Zhang, Xinyu
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Nature Publishing Group UK 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8433233/
https://www.ncbi.nlm.nih.gov/pubmed/34508120
http://dx.doi.org/10.1038/s41598-021-97428-8
_version_ 1783751335953301504
author Hong, Geng
Chen, Xiaoyan
Chen, Jianyong
Zhang, Miao
Ren, Yumeng
Zhang, Xinyu
author_facet Hong, Geng
Chen, Xiaoyan
Chen, Jianyong
Zhang, Miao
Ren, Yumeng
Zhang, Xinyu
author_sort Hong, Geng
collection PubMed
description Coronavirus 2019 (COVID-19) is a new acute respiratory disease that has spread rapidly throughout the world. In this paper, a lightweight convolutional neural network (CNN) model named multi-scale gated multi-head attention depthwise separable CNN (MGMADS-CNN) is proposed, which is based on attention mechanism and depthwise separable convolution. A multi-scale gated multi-head attention mechanism is designed to extract effective feature information from the COVID-19 X-ray and CT images for classification. Moreover, the depthwise separable convolution layers are adopted as MGMADS-CNN’s backbone to reduce the model size and parameters. The LeNet-5, AlexNet, GoogLeNet, ResNet, VGGNet-16, and three MGMADS-CNN models are trained, validated and tested with tenfold cross-validation on X-ray and CT images. The results show that MGMADS-CNN with three attention layers (MGMADS-3) has achieved accuracy of 96.75% on X-ray images and 98.25% on CT images. The specificity and sensitivity are 98.06% and 96.6% on X-ray images, and 98.17% and 98.05% on CT images. The size of MGMADS-3 model is only 43.6 M bytes. In addition, the detection speed of MGMADS-3 on X-ray images and CT images are 6.09 ms and 4.23 ms for per image, respectively. It is proved that the MGMADS-3 can detect and classify COVID-19 faster with higher accuracy and efficiency.
format Online
Article
Text
id pubmed-8433233
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher Nature Publishing Group UK
record_format MEDLINE/PubMed
spelling pubmed-84332332021-09-13 A multi-scale gated multi-head attention depthwise separable CNN model for recognizing COVID-19 Hong, Geng Chen, Xiaoyan Chen, Jianyong Zhang, Miao Ren, Yumeng Zhang, Xinyu Sci Rep Article Coronavirus 2019 (COVID-19) is a new acute respiratory disease that has spread rapidly throughout the world. In this paper, a lightweight convolutional neural network (CNN) model named multi-scale gated multi-head attention depthwise separable CNN (MGMADS-CNN) is proposed, which is based on attention mechanism and depthwise separable convolution. A multi-scale gated multi-head attention mechanism is designed to extract effective feature information from the COVID-19 X-ray and CT images for classification. Moreover, the depthwise separable convolution layers are adopted as MGMADS-CNN’s backbone to reduce the model size and parameters. The LeNet-5, AlexNet, GoogLeNet, ResNet, VGGNet-16, and three MGMADS-CNN models are trained, validated and tested with tenfold cross-validation on X-ray and CT images. The results show that MGMADS-CNN with three attention layers (MGMADS-3) has achieved accuracy of 96.75% on X-ray images and 98.25% on CT images. The specificity and sensitivity are 98.06% and 96.6% on X-ray images, and 98.17% and 98.05% on CT images. The size of MGMADS-3 model is only 43.6 M bytes. In addition, the detection speed of MGMADS-3 on X-ray images and CT images are 6.09 ms and 4.23 ms for per image, respectively. It is proved that the MGMADS-3 can detect and classify COVID-19 faster with higher accuracy and efficiency. Nature Publishing Group UK 2021-09-10 /pmc/articles/PMC8433233/ /pubmed/34508120 http://dx.doi.org/10.1038/s41598-021-97428-8 Text en © The Author(s) 2021 https://creativecommons.org/licenses/by/4.0/Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) .
spellingShingle Article
Hong, Geng
Chen, Xiaoyan
Chen, Jianyong
Zhang, Miao
Ren, Yumeng
Zhang, Xinyu
A multi-scale gated multi-head attention depthwise separable CNN model for recognizing COVID-19
title A multi-scale gated multi-head attention depthwise separable CNN model for recognizing COVID-19
title_full A multi-scale gated multi-head attention depthwise separable CNN model for recognizing COVID-19
title_fullStr A multi-scale gated multi-head attention depthwise separable CNN model for recognizing COVID-19
title_full_unstemmed A multi-scale gated multi-head attention depthwise separable CNN model for recognizing COVID-19
title_short A multi-scale gated multi-head attention depthwise separable CNN model for recognizing COVID-19
title_sort multi-scale gated multi-head attention depthwise separable cnn model for recognizing covid-19
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8433233/
https://www.ncbi.nlm.nih.gov/pubmed/34508120
http://dx.doi.org/10.1038/s41598-021-97428-8
work_keys_str_mv AT honggeng amultiscalegatedmultiheadattentiondepthwiseseparablecnnmodelforrecognizingcovid19
AT chenxiaoyan amultiscalegatedmultiheadattentiondepthwiseseparablecnnmodelforrecognizingcovid19
AT chenjianyong amultiscalegatedmultiheadattentiondepthwiseseparablecnnmodelforrecognizingcovid19
AT zhangmiao amultiscalegatedmultiheadattentiondepthwiseseparablecnnmodelforrecognizingcovid19
AT renyumeng amultiscalegatedmultiheadattentiondepthwiseseparablecnnmodelforrecognizingcovid19
AT zhangxinyu amultiscalegatedmultiheadattentiondepthwiseseparablecnnmodelforrecognizingcovid19
AT honggeng multiscalegatedmultiheadattentiondepthwiseseparablecnnmodelforrecognizingcovid19
AT chenxiaoyan multiscalegatedmultiheadattentiondepthwiseseparablecnnmodelforrecognizingcovid19
AT chenjianyong multiscalegatedmultiheadattentiondepthwiseseparablecnnmodelforrecognizingcovid19
AT zhangmiao multiscalegatedmultiheadattentiondepthwiseseparablecnnmodelforrecognizingcovid19
AT renyumeng multiscalegatedmultiheadattentiondepthwiseseparablecnnmodelforrecognizingcovid19
AT zhangxinyu multiscalegatedmultiheadattentiondepthwiseseparablecnnmodelforrecognizingcovid19