Cargando…

Explainable COVID-19 detection using fractal dimension and vision transformer with Grad-CAM on cough sounds

The polymerase chain reaction (PCR) test is not only time-intensive but also a contact method that puts healthcare personnel at risk. Thus, contactless and fast detection tests are more valuable. Cough sound is an important indicator of COVID-19, and in this paper, a novel explainable scheme is deve...

Descripción completa

Detalles Bibliográficos
Autores principales: Sobahi, Nebras, Atila, Orhan, Deniz, Erkan, Sengur, Abdulkadir, Acharya, U. Rajendra
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Nalecz Institute of Biocybernetics and Biomedical Engineering of the Polish Academy of Sciences. Published by Elsevier B.V. 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9444505/
https://www.ncbi.nlm.nih.gov/pubmed/36092540
http://dx.doi.org/10.1016/j.bbe.2022.08.005
_version_ 1784783240009089024
author Sobahi, Nebras
Atila, Orhan
Deniz, Erkan
Sengur, Abdulkadir
Acharya, U. Rajendra
author_facet Sobahi, Nebras
Atila, Orhan
Deniz, Erkan
Sengur, Abdulkadir
Acharya, U. Rajendra
author_sort Sobahi, Nebras
collection PubMed
description The polymerase chain reaction (PCR) test is not only time-intensive but also a contact method that puts healthcare personnel at risk. Thus, contactless and fast detection tests are more valuable. Cough sound is an important indicator of COVID-19, and in this paper, a novel explainable scheme is developed for cough sound-based COVID-19 detection. In the presented work, the cough sound is initially segmented into overlapping parts, and each segment is labeled as the input audio, which may contain other sounds. The deep Yet Another Mobile Network (YAMNet) model is considered in this work. After labeling, the segments labeled as cough are cropped and concatenated to reconstruct the pure cough sounds. Then, four fractal dimensions (FD) calculation methods are employed to acquire the FD coefficients on the cough sound with an overlapped sliding window that forms a matrix. The constructed matrixes are then used to form the fractal dimension images. Finally, a pretrained vision transformer (ViT) model is used to classify the constructed images into COVID-19, healthy and symptomatic classes. In this work, we demonstrate the performance of the ViT on cough sound-based COVID-19, and a visual explainability of the inner workings of the ViT model is shown. Three publically available cough sound datasets, namely COUGHVID, VIRUFY, and COSWARA, are used in this study. We have obtained 98.45%, 98.15%, and 97.59% accuracy for COUGHVID, VIRUFY, and COSWARA datasets, respectively. Our developed model obtained the highest performance compared to the state-of-the-art methods and is ready to be tested in real-world applications.
format Online
Article
Text
id pubmed-9444505
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Nalecz Institute of Biocybernetics and Biomedical Engineering of the Polish Academy of Sciences. Published by Elsevier B.V.
record_format MEDLINE/PubMed
spelling pubmed-94445052022-09-06 Explainable COVID-19 detection using fractal dimension and vision transformer with Grad-CAM on cough sounds Sobahi, Nebras Atila, Orhan Deniz, Erkan Sengur, Abdulkadir Acharya, U. Rajendra Biocybern Biomed Eng Original Research Article The polymerase chain reaction (PCR) test is not only time-intensive but also a contact method that puts healthcare personnel at risk. Thus, contactless and fast detection tests are more valuable. Cough sound is an important indicator of COVID-19, and in this paper, a novel explainable scheme is developed for cough sound-based COVID-19 detection. In the presented work, the cough sound is initially segmented into overlapping parts, and each segment is labeled as the input audio, which may contain other sounds. The deep Yet Another Mobile Network (YAMNet) model is considered in this work. After labeling, the segments labeled as cough are cropped and concatenated to reconstruct the pure cough sounds. Then, four fractal dimensions (FD) calculation methods are employed to acquire the FD coefficients on the cough sound with an overlapped sliding window that forms a matrix. The constructed matrixes are then used to form the fractal dimension images. Finally, a pretrained vision transformer (ViT) model is used to classify the constructed images into COVID-19, healthy and symptomatic classes. In this work, we demonstrate the performance of the ViT on cough sound-based COVID-19, and a visual explainability of the inner workings of the ViT model is shown. Three publically available cough sound datasets, namely COUGHVID, VIRUFY, and COSWARA, are used in this study. We have obtained 98.45%, 98.15%, and 97.59% accuracy for COUGHVID, VIRUFY, and COSWARA datasets, respectively. Our developed model obtained the highest performance compared to the state-of-the-art methods and is ready to be tested in real-world applications. Nalecz Institute of Biocybernetics and Biomedical Engineering of the Polish Academy of Sciences. Published by Elsevier B.V. 2022 2022-09-06 /pmc/articles/PMC9444505/ /pubmed/36092540 http://dx.doi.org/10.1016/j.bbe.2022.08.005 Text en © 2022 Nalecz Institute of Biocybernetics and Biomedical Engineering of the Polish Academy of Sciences. Published by Elsevier B.V. All rights reserved. Since January 2020 Elsevier has created a COVID-19 resource centre with free information in English and Mandarin on the novel coronavirus COVID-19. The COVID-19 resource centre is hosted on Elsevier Connect, the company's public news and information website. Elsevier hereby grants permission to make all its COVID-19-related research that is available on the COVID-19 resource centre - including this research content - immediately available in PubMed Central and other publicly funded repositories, such as the WHO COVID database with rights for unrestricted research re-use and analyses in any form or by any means with acknowledgement of the original source. These permissions are granted for free by Elsevier for as long as the COVID-19 resource centre remains active.
spellingShingle Original Research Article
Sobahi, Nebras
Atila, Orhan
Deniz, Erkan
Sengur, Abdulkadir
Acharya, U. Rajendra
Explainable COVID-19 detection using fractal dimension and vision transformer with Grad-CAM on cough sounds
title Explainable COVID-19 detection using fractal dimension and vision transformer with Grad-CAM on cough sounds
title_full Explainable COVID-19 detection using fractal dimension and vision transformer with Grad-CAM on cough sounds
title_fullStr Explainable COVID-19 detection using fractal dimension and vision transformer with Grad-CAM on cough sounds
title_full_unstemmed Explainable COVID-19 detection using fractal dimension and vision transformer with Grad-CAM on cough sounds
title_short Explainable COVID-19 detection using fractal dimension and vision transformer with Grad-CAM on cough sounds
title_sort explainable covid-19 detection using fractal dimension and vision transformer with grad-cam on cough sounds
topic Original Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9444505/
https://www.ncbi.nlm.nih.gov/pubmed/36092540
http://dx.doi.org/10.1016/j.bbe.2022.08.005
work_keys_str_mv AT sobahinebras explainablecovid19detectionusingfractaldimensionandvisiontransformerwithgradcamoncoughsounds
AT atilaorhan explainablecovid19detectionusingfractaldimensionandvisiontransformerwithgradcamoncoughsounds
AT denizerkan explainablecovid19detectionusingfractaldimensionandvisiontransformerwithgradcamoncoughsounds
AT sengurabdulkadir explainablecovid19detectionusingfractaldimensionandvisiontransformerwithgradcamoncoughsounds
AT acharyaurajendra explainablecovid19detectionusingfractaldimensionandvisiontransformerwithgradcamoncoughsounds