Cargando…
Explainable COVID-19 detection using fractal dimension and vision transformer with Grad-CAM on cough sounds
The polymerase chain reaction (PCR) test is not only time-intensive but also a contact method that puts healthcare personnel at risk. Thus, contactless and fast detection tests are more valuable. Cough sound is an important indicator of COVID-19, and in this paper, a novel explainable scheme is deve...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Nalecz Institute of Biocybernetics and Biomedical Engineering of the Polish Academy of Sciences. Published by Elsevier B.V.
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9444505/ https://www.ncbi.nlm.nih.gov/pubmed/36092540 http://dx.doi.org/10.1016/j.bbe.2022.08.005 |
_version_ | 1784783240009089024 |
---|---|
author | Sobahi, Nebras Atila, Orhan Deniz, Erkan Sengur, Abdulkadir Acharya, U. Rajendra |
author_facet | Sobahi, Nebras Atila, Orhan Deniz, Erkan Sengur, Abdulkadir Acharya, U. Rajendra |
author_sort | Sobahi, Nebras |
collection | PubMed |
description | The polymerase chain reaction (PCR) test is not only time-intensive but also a contact method that puts healthcare personnel at risk. Thus, contactless and fast detection tests are more valuable. Cough sound is an important indicator of COVID-19, and in this paper, a novel explainable scheme is developed for cough sound-based COVID-19 detection. In the presented work, the cough sound is initially segmented into overlapping parts, and each segment is labeled as the input audio, which may contain other sounds. The deep Yet Another Mobile Network (YAMNet) model is considered in this work. After labeling, the segments labeled as cough are cropped and concatenated to reconstruct the pure cough sounds. Then, four fractal dimensions (FD) calculation methods are employed to acquire the FD coefficients on the cough sound with an overlapped sliding window that forms a matrix. The constructed matrixes are then used to form the fractal dimension images. Finally, a pretrained vision transformer (ViT) model is used to classify the constructed images into COVID-19, healthy and symptomatic classes. In this work, we demonstrate the performance of the ViT on cough sound-based COVID-19, and a visual explainability of the inner workings of the ViT model is shown. Three publically available cough sound datasets, namely COUGHVID, VIRUFY, and COSWARA, are used in this study. We have obtained 98.45%, 98.15%, and 97.59% accuracy for COUGHVID, VIRUFY, and COSWARA datasets, respectively. Our developed model obtained the highest performance compared to the state-of-the-art methods and is ready to be tested in real-world applications. |
format | Online Article Text |
id | pubmed-9444505 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Nalecz Institute of Biocybernetics and Biomedical Engineering of the Polish Academy of Sciences. Published by Elsevier B.V. |
record_format | MEDLINE/PubMed |
spelling | pubmed-94445052022-09-06 Explainable COVID-19 detection using fractal dimension and vision transformer with Grad-CAM on cough sounds Sobahi, Nebras Atila, Orhan Deniz, Erkan Sengur, Abdulkadir Acharya, U. Rajendra Biocybern Biomed Eng Original Research Article The polymerase chain reaction (PCR) test is not only time-intensive but also a contact method that puts healthcare personnel at risk. Thus, contactless and fast detection tests are more valuable. Cough sound is an important indicator of COVID-19, and in this paper, a novel explainable scheme is developed for cough sound-based COVID-19 detection. In the presented work, the cough sound is initially segmented into overlapping parts, and each segment is labeled as the input audio, which may contain other sounds. The deep Yet Another Mobile Network (YAMNet) model is considered in this work. After labeling, the segments labeled as cough are cropped and concatenated to reconstruct the pure cough sounds. Then, four fractal dimensions (FD) calculation methods are employed to acquire the FD coefficients on the cough sound with an overlapped sliding window that forms a matrix. The constructed matrixes are then used to form the fractal dimension images. Finally, a pretrained vision transformer (ViT) model is used to classify the constructed images into COVID-19, healthy and symptomatic classes. In this work, we demonstrate the performance of the ViT on cough sound-based COVID-19, and a visual explainability of the inner workings of the ViT model is shown. Three publically available cough sound datasets, namely COUGHVID, VIRUFY, and COSWARA, are used in this study. We have obtained 98.45%, 98.15%, and 97.59% accuracy for COUGHVID, VIRUFY, and COSWARA datasets, respectively. Our developed model obtained the highest performance compared to the state-of-the-art methods and is ready to be tested in real-world applications. Nalecz Institute of Biocybernetics and Biomedical Engineering of the Polish Academy of Sciences. Published by Elsevier B.V. 2022 2022-09-06 /pmc/articles/PMC9444505/ /pubmed/36092540 http://dx.doi.org/10.1016/j.bbe.2022.08.005 Text en © 2022 Nalecz Institute of Biocybernetics and Biomedical Engineering of the Polish Academy of Sciences. Published by Elsevier B.V. All rights reserved. Since January 2020 Elsevier has created a COVID-19 resource centre with free information in English and Mandarin on the novel coronavirus COVID-19. The COVID-19 resource centre is hosted on Elsevier Connect, the company's public news and information website. Elsevier hereby grants permission to make all its COVID-19-related research that is available on the COVID-19 resource centre - including this research content - immediately available in PubMed Central and other publicly funded repositories, such as the WHO COVID database with rights for unrestricted research re-use and analyses in any form or by any means with acknowledgement of the original source. These permissions are granted for free by Elsevier for as long as the COVID-19 resource centre remains active. |
spellingShingle | Original Research Article Sobahi, Nebras Atila, Orhan Deniz, Erkan Sengur, Abdulkadir Acharya, U. Rajendra Explainable COVID-19 detection using fractal dimension and vision transformer with Grad-CAM on cough sounds |
title | Explainable COVID-19 detection using fractal dimension and vision transformer with Grad-CAM on cough sounds |
title_full | Explainable COVID-19 detection using fractal dimension and vision transformer with Grad-CAM on cough sounds |
title_fullStr | Explainable COVID-19 detection using fractal dimension and vision transformer with Grad-CAM on cough sounds |
title_full_unstemmed | Explainable COVID-19 detection using fractal dimension and vision transformer with Grad-CAM on cough sounds |
title_short | Explainable COVID-19 detection using fractal dimension and vision transformer with Grad-CAM on cough sounds |
title_sort | explainable covid-19 detection using fractal dimension and vision transformer with grad-cam on cough sounds |
topic | Original Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9444505/ https://www.ncbi.nlm.nih.gov/pubmed/36092540 http://dx.doi.org/10.1016/j.bbe.2022.08.005 |
work_keys_str_mv | AT sobahinebras explainablecovid19detectionusingfractaldimensionandvisiontransformerwithgradcamoncoughsounds AT atilaorhan explainablecovid19detectionusingfractaldimensionandvisiontransformerwithgradcamoncoughsounds AT denizerkan explainablecovid19detectionusingfractaldimensionandvisiontransformerwithgradcamoncoughsounds AT sengurabdulkadir explainablecovid19detectionusingfractaldimensionandvisiontransformerwithgradcamoncoughsounds AT acharyaurajendra explainablecovid19detectionusingfractaldimensionandvisiontransformerwithgradcamoncoughsounds |