Cargando…
Feature-Based Fusion Using CNN for Lung and Heart Sound Classification †
Lung or heart sound classification is challenging due to the complex nature of audio data, its dynamic properties of time, and frequency domains. It is also very difficult to detect lung or heart conditions with small amounts of data or unbalanced and high noise in data. Furthermore, the quality of...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8875944/ https://www.ncbi.nlm.nih.gov/pubmed/35214424 http://dx.doi.org/10.3390/s22041521 |
_version_ | 1784658051700097024 |
---|---|
author | Tariq, Zeenat Shah, Sayed Khushal Lee, Yugyung |
author_facet | Tariq, Zeenat Shah, Sayed Khushal Lee, Yugyung |
author_sort | Tariq, Zeenat |
collection | PubMed |
description | Lung or heart sound classification is challenging due to the complex nature of audio data, its dynamic properties of time, and frequency domains. It is also very difficult to detect lung or heart conditions with small amounts of data or unbalanced and high noise in data. Furthermore, the quality of data is a considerable pitfall for improving the performance of deep learning. In this paper, we propose a novel feature-based fusion network called FDC-FS for classifying heart and lung sounds. The FDC-FS framework aims to effectively transfer learning from three different deep neural network models built from audio datasets. The innovation of the proposed transfer learning relies on the transformation from audio data to image vectors and from three specific models to one fused model that would be more suitable for deep learning. We used two publicly available datasets for this study, i.e., lung sound data from ICHBI 2017 challenge and heart challenge data. We applied data augmentation techniques, such as noise distortion, pitch shift, and time stretching, dealing with some data issues in these datasets. Importantly, we extracted three unique features from the audio samples, i.e., Spectrogram, MFCC, and Chromagram. Finally, we built a fusion of three optimal convolutional neural network models by feeding the image feature vectors transformed from audio features. We confirmed the superiority of the proposed fusion model compared to the state-of-the-art works. The highest accuracy we achieved with FDC-FS is 99.1% with Spectrogram-based lung sound classification while 97% for Spectrogram and Chromagram based heart sound classification. |
format | Online Article Text |
id | pubmed-8875944 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-88759442022-02-26 Feature-Based Fusion Using CNN for Lung and Heart Sound Classification † Tariq, Zeenat Shah, Sayed Khushal Lee, Yugyung Sensors (Basel) Article Lung or heart sound classification is challenging due to the complex nature of audio data, its dynamic properties of time, and frequency domains. It is also very difficult to detect lung or heart conditions with small amounts of data or unbalanced and high noise in data. Furthermore, the quality of data is a considerable pitfall for improving the performance of deep learning. In this paper, we propose a novel feature-based fusion network called FDC-FS for classifying heart and lung sounds. The FDC-FS framework aims to effectively transfer learning from three different deep neural network models built from audio datasets. The innovation of the proposed transfer learning relies on the transformation from audio data to image vectors and from three specific models to one fused model that would be more suitable for deep learning. We used two publicly available datasets for this study, i.e., lung sound data from ICHBI 2017 challenge and heart challenge data. We applied data augmentation techniques, such as noise distortion, pitch shift, and time stretching, dealing with some data issues in these datasets. Importantly, we extracted three unique features from the audio samples, i.e., Spectrogram, MFCC, and Chromagram. Finally, we built a fusion of three optimal convolutional neural network models by feeding the image feature vectors transformed from audio features. We confirmed the superiority of the proposed fusion model compared to the state-of-the-art works. The highest accuracy we achieved with FDC-FS is 99.1% with Spectrogram-based lung sound classification while 97% for Spectrogram and Chromagram based heart sound classification. MDPI 2022-02-16 /pmc/articles/PMC8875944/ /pubmed/35214424 http://dx.doi.org/10.3390/s22041521 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Tariq, Zeenat Shah, Sayed Khushal Lee, Yugyung Feature-Based Fusion Using CNN for Lung and Heart Sound Classification † |
title | Feature-Based Fusion Using CNN for Lung and Heart Sound Classification † |
title_full | Feature-Based Fusion Using CNN for Lung and Heart Sound Classification † |
title_fullStr | Feature-Based Fusion Using CNN for Lung and Heart Sound Classification † |
title_full_unstemmed | Feature-Based Fusion Using CNN for Lung and Heart Sound Classification † |
title_short | Feature-Based Fusion Using CNN for Lung and Heart Sound Classification † |
title_sort | feature-based fusion using cnn for lung and heart sound classification † |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8875944/ https://www.ncbi.nlm.nih.gov/pubmed/35214424 http://dx.doi.org/10.3390/s22041521 |
work_keys_str_mv | AT tariqzeenat featurebasedfusionusingcnnforlungandheartsoundclassification AT shahsayedkhushal featurebasedfusionusingcnnforlungandheartsoundclassification AT leeyugyung featurebasedfusionusingcnnforlungandheartsoundclassification |