Cargando…

High-Level CNN and Machine Learning Methods for Speaker Recognition

Speaker Recognition (SR) is a common task in AI-based sound analysis, involving structurally different methodologies such as Deep Learning or “traditional” Machine Learning (ML). In this paper, we compared and explored the two methodologies on the DEMoS dataset consisting of 8869 audio files of 58 s...

Descripción completa

Detalles Bibliográficos
Autores principales: Costantini, Giovanni, Cesarini, Valerio, Brenna, Emanuele
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10098737/
https://www.ncbi.nlm.nih.gov/pubmed/37050521
http://dx.doi.org/10.3390/s23073461
_version_ 1785024885740797952
author Costantini, Giovanni
Cesarini, Valerio
Brenna, Emanuele
author_facet Costantini, Giovanni
Cesarini, Valerio
Brenna, Emanuele
author_sort Costantini, Giovanni
collection PubMed
description Speaker Recognition (SR) is a common task in AI-based sound analysis, involving structurally different methodologies such as Deep Learning or “traditional” Machine Learning (ML). In this paper, we compared and explored the two methodologies on the DEMoS dataset consisting of 8869 audio files of 58 speakers in different emotional states. A custom CNN is compared to several pre-trained nets using image inputs of spectrograms and Cepstral-temporal (MFCC) graphs. AML approach based on acoustic feature extraction, selection and multi-class classification by means of a Naïve Bayes model is also considered. Results show how a custom, less deep CNN trained on grayscale spectrogram images obtain the most accurate results, 90.15% on grayscale spectrograms and 83.17% on colored MFCC. AlexNet provides comparable results, reaching 89.28% on spectrograms and 83.43% on MFCC.The Naïve Bayes classifier provides a 87.09% accuracy and a 0.985 average AUC while being faster to train and more interpretable. Feature selection shows how F0, MFCC and voicing-related features are the most characterizing for this SR task. The high amount of training samples and the emotional content of the DEMoS dataset better reflect a real case scenario for speaker recognition, and account for the generalization power of the models.
format Online
Article
Text
id pubmed-10098737
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-100987372023-04-14 High-Level CNN and Machine Learning Methods for Speaker Recognition Costantini, Giovanni Cesarini, Valerio Brenna, Emanuele Sensors (Basel) Article Speaker Recognition (SR) is a common task in AI-based sound analysis, involving structurally different methodologies such as Deep Learning or “traditional” Machine Learning (ML). In this paper, we compared and explored the two methodologies on the DEMoS dataset consisting of 8869 audio files of 58 speakers in different emotional states. A custom CNN is compared to several pre-trained nets using image inputs of spectrograms and Cepstral-temporal (MFCC) graphs. AML approach based on acoustic feature extraction, selection and multi-class classification by means of a Naïve Bayes model is also considered. Results show how a custom, less deep CNN trained on grayscale spectrogram images obtain the most accurate results, 90.15% on grayscale spectrograms and 83.17% on colored MFCC. AlexNet provides comparable results, reaching 89.28% on spectrograms and 83.43% on MFCC.The Naïve Bayes classifier provides a 87.09% accuracy and a 0.985 average AUC while being faster to train and more interpretable. Feature selection shows how F0, MFCC and voicing-related features are the most characterizing for this SR task. The high amount of training samples and the emotional content of the DEMoS dataset better reflect a real case scenario for speaker recognition, and account for the generalization power of the models. MDPI 2023-03-25 /pmc/articles/PMC10098737/ /pubmed/37050521 http://dx.doi.org/10.3390/s23073461 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Costantini, Giovanni
Cesarini, Valerio
Brenna, Emanuele
High-Level CNN and Machine Learning Methods for Speaker Recognition
title High-Level CNN and Machine Learning Methods for Speaker Recognition
title_full High-Level CNN and Machine Learning Methods for Speaker Recognition
title_fullStr High-Level CNN and Machine Learning Methods for Speaker Recognition
title_full_unstemmed High-Level CNN and Machine Learning Methods for Speaker Recognition
title_short High-Level CNN and Machine Learning Methods for Speaker Recognition
title_sort high-level cnn and machine learning methods for speaker recognition
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10098737/
https://www.ncbi.nlm.nih.gov/pubmed/37050521
http://dx.doi.org/10.3390/s23073461
work_keys_str_mv AT costantinigiovanni highlevelcnnandmachinelearningmethodsforspeakerrecognition
AT cesarinivalerio highlevelcnnandmachinelearningmethodsforspeakerrecognition
AT brennaemanuele highlevelcnnandmachinelearningmethodsforspeakerrecognition