Cargando…

Feature-Based Fusion Using CNN for Lung and Heart Sound Classification †

Lung or heart sound classification is challenging due to the complex nature of audio data, its dynamic properties of time, and frequency domains. It is also very difficult to detect lung or heart conditions with small amounts of data or unbalanced and high noise in data. Furthermore, the quality of...

Descripción completa

Detalles Bibliográficos
Autores principales: Tariq, Zeenat, Shah, Sayed Khushal, Lee, Yugyung
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8875944/
https://www.ncbi.nlm.nih.gov/pubmed/35214424
http://dx.doi.org/10.3390/s22041521
_version_ 1784658051700097024
author Tariq, Zeenat
Shah, Sayed Khushal
Lee, Yugyung
author_facet Tariq, Zeenat
Shah, Sayed Khushal
Lee, Yugyung
author_sort Tariq, Zeenat
collection PubMed
description Lung or heart sound classification is challenging due to the complex nature of audio data, its dynamic properties of time, and frequency domains. It is also very difficult to detect lung or heart conditions with small amounts of data or unbalanced and high noise in data. Furthermore, the quality of data is a considerable pitfall for improving the performance of deep learning. In this paper, we propose a novel feature-based fusion network called FDC-FS for classifying heart and lung sounds. The FDC-FS framework aims to effectively transfer learning from three different deep neural network models built from audio datasets. The innovation of the proposed transfer learning relies on the transformation from audio data to image vectors and from three specific models to one fused model that would be more suitable for deep learning. We used two publicly available datasets for this study, i.e., lung sound data from ICHBI 2017 challenge and heart challenge data. We applied data augmentation techniques, such as noise distortion, pitch shift, and time stretching, dealing with some data issues in these datasets. Importantly, we extracted three unique features from the audio samples, i.e., Spectrogram, MFCC, and Chromagram. Finally, we built a fusion of three optimal convolutional neural network models by feeding the image feature vectors transformed from audio features. We confirmed the superiority of the proposed fusion model compared to the state-of-the-art works. The highest accuracy we achieved with FDC-FS is 99.1% with Spectrogram-based lung sound classification while 97% for Spectrogram and Chromagram based heart sound classification.
format Online
Article
Text
id pubmed-8875944
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-88759442022-02-26 Feature-Based Fusion Using CNN for Lung and Heart Sound Classification † Tariq, Zeenat Shah, Sayed Khushal Lee, Yugyung Sensors (Basel) Article Lung or heart sound classification is challenging due to the complex nature of audio data, its dynamic properties of time, and frequency domains. It is also very difficult to detect lung or heart conditions with small amounts of data or unbalanced and high noise in data. Furthermore, the quality of data is a considerable pitfall for improving the performance of deep learning. In this paper, we propose a novel feature-based fusion network called FDC-FS for classifying heart and lung sounds. The FDC-FS framework aims to effectively transfer learning from three different deep neural network models built from audio datasets. The innovation of the proposed transfer learning relies on the transformation from audio data to image vectors and from three specific models to one fused model that would be more suitable for deep learning. We used two publicly available datasets for this study, i.e., lung sound data from ICHBI 2017 challenge and heart challenge data. We applied data augmentation techniques, such as noise distortion, pitch shift, and time stretching, dealing with some data issues in these datasets. Importantly, we extracted three unique features from the audio samples, i.e., Spectrogram, MFCC, and Chromagram. Finally, we built a fusion of three optimal convolutional neural network models by feeding the image feature vectors transformed from audio features. We confirmed the superiority of the proposed fusion model compared to the state-of-the-art works. The highest accuracy we achieved with FDC-FS is 99.1% with Spectrogram-based lung sound classification while 97% for Spectrogram and Chromagram based heart sound classification. MDPI 2022-02-16 /pmc/articles/PMC8875944/ /pubmed/35214424 http://dx.doi.org/10.3390/s22041521 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Tariq, Zeenat
Shah, Sayed Khushal
Lee, Yugyung
Feature-Based Fusion Using CNN for Lung and Heart Sound Classification †
title Feature-Based Fusion Using CNN for Lung and Heart Sound Classification †
title_full Feature-Based Fusion Using CNN for Lung and Heart Sound Classification †
title_fullStr Feature-Based Fusion Using CNN for Lung and Heart Sound Classification †
title_full_unstemmed Feature-Based Fusion Using CNN for Lung and Heart Sound Classification †
title_short Feature-Based Fusion Using CNN for Lung and Heart Sound Classification †
title_sort feature-based fusion using cnn for lung and heart sound classification †
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8875944/
https://www.ncbi.nlm.nih.gov/pubmed/35214424
http://dx.doi.org/10.3390/s22041521
work_keys_str_mv AT tariqzeenat featurebasedfusionusingcnnforlungandheartsoundclassification
AT shahsayedkhushal featurebasedfusionusingcnnforlungandheartsoundclassification
AT leeyugyung featurebasedfusionusingcnnforlungandheartsoundclassification