Cargando…
Deep learning attention-guided radiomics for COVID-19 chest radiograph classification
BACKGROUND: Accurate assessment of coronavirus disease 2019 (COVID-19) lung involvement through chest radiograph plays an important role in effective management of the infection. This study aims to develop a two-step feature merging method to integrate image features from deep learning and radiomics...
Autores principales: | , , , , , , , , , , , , , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
AME Publishing Company
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9929417/ https://www.ncbi.nlm.nih.gov/pubmed/36819269 http://dx.doi.org/10.21037/qims-22-531 |
_version_ | 1784888850349293568 |
---|---|
author | Yang, Dongrong Ren, Ge Ni, Ruiyan Huang, Yu-Hua Lam, Ngo Fung Daniel Sun, Hongfei Wan, Shiu Bun Nelson Wong, Man Fung Esther Chan, King Kwong Tsang, Hoi Ching Hailey Xu, Lu Wu, Tak Chiu Kong, Feng-Ming (Spring) Wáng, Yì Xiáng J. Qin, Jing Chan, Lawrence Wing Chi Ying, Michael Cai, Jing |
author_facet | Yang, Dongrong Ren, Ge Ni, Ruiyan Huang, Yu-Hua Lam, Ngo Fung Daniel Sun, Hongfei Wan, Shiu Bun Nelson Wong, Man Fung Esther Chan, King Kwong Tsang, Hoi Ching Hailey Xu, Lu Wu, Tak Chiu Kong, Feng-Ming (Spring) Wáng, Yì Xiáng J. Qin, Jing Chan, Lawrence Wing Chi Ying, Michael Cai, Jing |
author_sort | Yang, Dongrong |
collection | PubMed |
description | BACKGROUND: Accurate assessment of coronavirus disease 2019 (COVID-19) lung involvement through chest radiograph plays an important role in effective management of the infection. This study aims to develop a two-step feature merging method to integrate image features from deep learning and radiomics to differentiate COVID-19, non-COVID-19 pneumonia and normal chest radiographs (CXR). METHODS: In this study, a deformable convolutional neural network (deformable CNN) was developed and used as a feature extractor to obtain 1,024-dimensional deep learning latent representation (DLR) features. Then 1,069-dimensional radiomics features were extracted from the region of interest (ROI) guided by deformable CNN’s attention. The two feature sets were concatenated to generate a merged feature set for classification. For comparative experiments, the same process has been applied to the DLR-only feature set for verifying the effectiveness of feature concatenation. RESULTS: Using the merged feature set resulted in an overall average accuracy of 91.0% for three-class classification, representing a statistically significant improvement of 0.6% compared to the DLR-only classification. The recall and precision of classification into the COVID-19 class were 0.926 and 0.976, respectively. The feature merging method was shown to significantly improve the classification performance as compared to using only deep learning features, regardless of choice of classifier (P value <0.0001). Three classes’ F1-score were 0.892, 0.890, and 0.950 correspondingly (i.e., normal, non-COVID-19 pneumonia, COVID-19). CONCLUSIONS: A two-step COVID-19 classification framework integrating information from both DLR and radiomics features (guided by deep learning attention mechanism) has been developed. The proposed feature merging method has been shown to improve the performance of chest radiograph classification as compared to the case of using only deep learning features. |
format | Online Article Text |
id | pubmed-9929417 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | AME Publishing Company |
record_format | MEDLINE/PubMed |
spelling | pubmed-99294172023-02-16 Deep learning attention-guided radiomics for COVID-19 chest radiograph classification Yang, Dongrong Ren, Ge Ni, Ruiyan Huang, Yu-Hua Lam, Ngo Fung Daniel Sun, Hongfei Wan, Shiu Bun Nelson Wong, Man Fung Esther Chan, King Kwong Tsang, Hoi Ching Hailey Xu, Lu Wu, Tak Chiu Kong, Feng-Ming (Spring) Wáng, Yì Xiáng J. Qin, Jing Chan, Lawrence Wing Chi Ying, Michael Cai, Jing Quant Imaging Med Surg Original Article BACKGROUND: Accurate assessment of coronavirus disease 2019 (COVID-19) lung involvement through chest radiograph plays an important role in effective management of the infection. This study aims to develop a two-step feature merging method to integrate image features from deep learning and radiomics to differentiate COVID-19, non-COVID-19 pneumonia and normal chest radiographs (CXR). METHODS: In this study, a deformable convolutional neural network (deformable CNN) was developed and used as a feature extractor to obtain 1,024-dimensional deep learning latent representation (DLR) features. Then 1,069-dimensional radiomics features were extracted from the region of interest (ROI) guided by deformable CNN’s attention. The two feature sets were concatenated to generate a merged feature set for classification. For comparative experiments, the same process has been applied to the DLR-only feature set for verifying the effectiveness of feature concatenation. RESULTS: Using the merged feature set resulted in an overall average accuracy of 91.0% for three-class classification, representing a statistically significant improvement of 0.6% compared to the DLR-only classification. The recall and precision of classification into the COVID-19 class were 0.926 and 0.976, respectively. The feature merging method was shown to significantly improve the classification performance as compared to using only deep learning features, regardless of choice of classifier (P value <0.0001). Three classes’ F1-score were 0.892, 0.890, and 0.950 correspondingly (i.e., normal, non-COVID-19 pneumonia, COVID-19). CONCLUSIONS: A two-step COVID-19 classification framework integrating information from both DLR and radiomics features (guided by deep learning attention mechanism) has been developed. The proposed feature merging method has been shown to improve the performance of chest radiograph classification as compared to the case of using only deep learning features. AME Publishing Company 2022-11-21 2023-02-01 /pmc/articles/PMC9929417/ /pubmed/36819269 http://dx.doi.org/10.21037/qims-22-531 Text en 2023 Quantitative Imaging in Medicine and Surgery. All rights reserved. https://creativecommons.org/licenses/by-nc-nd/4.0/Open Access Statement: This is an Open Access article distributed in accordance with the Creative Commons Attribution-NonCommercial-NoDerivs 4.0 International License (CC BY-NC-ND 4.0), which permits the non-commercial replication and distribution of the article with the strict proviso that no changes or edits are made and the original work is properly cited (including links to both the formal publication through the relevant DOI and the license). See: https://creativecommons.org/licenses/by-nc-nd/4.0 (https://creativecommons.org/licenses/by-nc-nd/4.0/) . |
spellingShingle | Original Article Yang, Dongrong Ren, Ge Ni, Ruiyan Huang, Yu-Hua Lam, Ngo Fung Daniel Sun, Hongfei Wan, Shiu Bun Nelson Wong, Man Fung Esther Chan, King Kwong Tsang, Hoi Ching Hailey Xu, Lu Wu, Tak Chiu Kong, Feng-Ming (Spring) Wáng, Yì Xiáng J. Qin, Jing Chan, Lawrence Wing Chi Ying, Michael Cai, Jing Deep learning attention-guided radiomics for COVID-19 chest radiograph classification |
title | Deep learning attention-guided radiomics for COVID-19 chest radiograph classification |
title_full | Deep learning attention-guided radiomics for COVID-19 chest radiograph classification |
title_fullStr | Deep learning attention-guided radiomics for COVID-19 chest radiograph classification |
title_full_unstemmed | Deep learning attention-guided radiomics for COVID-19 chest radiograph classification |
title_short | Deep learning attention-guided radiomics for COVID-19 chest radiograph classification |
title_sort | deep learning attention-guided radiomics for covid-19 chest radiograph classification |
topic | Original Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9929417/ https://www.ncbi.nlm.nih.gov/pubmed/36819269 http://dx.doi.org/10.21037/qims-22-531 |
work_keys_str_mv | AT yangdongrong deeplearningattentionguidedradiomicsforcovid19chestradiographclassification AT renge deeplearningattentionguidedradiomicsforcovid19chestradiographclassification AT niruiyan deeplearningattentionguidedradiomicsforcovid19chestradiographclassification AT huangyuhua deeplearningattentionguidedradiomicsforcovid19chestradiographclassification AT lamngofungdaniel deeplearningattentionguidedradiomicsforcovid19chestradiographclassification AT sunhongfei deeplearningattentionguidedradiomicsforcovid19chestradiographclassification AT wanshiubunnelson deeplearningattentionguidedradiomicsforcovid19chestradiographclassification AT wongmanfungesther deeplearningattentionguidedradiomicsforcovid19chestradiographclassification AT chankingkwong deeplearningattentionguidedradiomicsforcovid19chestradiographclassification AT tsanghoichinghailey deeplearningattentionguidedradiomicsforcovid19chestradiographclassification AT xulu deeplearningattentionguidedradiomicsforcovid19chestradiographclassification AT wutakchiu deeplearningattentionguidedradiomicsforcovid19chestradiographclassification AT kongfengmingspring deeplearningattentionguidedradiomicsforcovid19chestradiographclassification AT wangyixiangj deeplearningattentionguidedradiomicsforcovid19chestradiographclassification AT qinjing deeplearningattentionguidedradiomicsforcovid19chestradiographclassification AT chanlawrencewingchi deeplearningattentionguidedradiomicsforcovid19chestradiographclassification AT yingmichael deeplearningattentionguidedradiomicsforcovid19chestradiographclassification AT caijing deeplearningattentionguidedradiomicsforcovid19chestradiographclassification |