Cargando…

A Deep Learning Method Using Gender-Specific Features for Emotion Recognition

Speech reflects people’s mental state and using a microphone sensor is a potential method for human–computer interaction. Speech recognition using this sensor is conducive to the diagnosis of mental illnesses. The gender difference of speakers affects the process of speech emotion recognition based...

Descripción completa

Detalles Bibliográficos
Autores principales: Zhang, Li-Min, Li, Yang, Zhang, Yue-Ting, Ng, Giap Weng, Leau, Yu-Beng, Yan, Hao
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9921859/
https://www.ncbi.nlm.nih.gov/pubmed/36772395
http://dx.doi.org/10.3390/s23031355
_version_ 1784887412950827008
author Zhang, Li-Min
Li, Yang
Zhang, Yue-Ting
Ng, Giap Weng
Leau, Yu-Beng
Yan, Hao
author_facet Zhang, Li-Min
Li, Yang
Zhang, Yue-Ting
Ng, Giap Weng
Leau, Yu-Beng
Yan, Hao
author_sort Zhang, Li-Min
collection PubMed
description Speech reflects people’s mental state and using a microphone sensor is a potential method for human–computer interaction. Speech recognition using this sensor is conducive to the diagnosis of mental illnesses. The gender difference of speakers affects the process of speech emotion recognition based on specific acoustic features, resulting in the decline of emotion recognition accuracy. Therefore, we believe that the accuracy of speech emotion recognition can be effectively improved by selecting different features of speech for emotion recognition based on the speech representations of different genders. In this paper, we propose a speech emotion recognition method based on gender classification. First, we use MLP to classify the original speech by gender. Second, based on the different acoustic features of male and female speech, we analyze the influence weights of multiple speech emotion features in male and female speech, and establish the optimal feature sets for male and female emotion recognition, respectively. Finally, we train and test CNN and BiLSTM, respectively, by using the male and the female speech emotion feature sets. The results show that the proposed emotion recognition models have an advantage in terms of average recognition accuracy compared with gender-mixed recognition models.
format Online
Article
Text
id pubmed-9921859
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-99218592023-02-12 A Deep Learning Method Using Gender-Specific Features for Emotion Recognition Zhang, Li-Min Li, Yang Zhang, Yue-Ting Ng, Giap Weng Leau, Yu-Beng Yan, Hao Sensors (Basel) Article Speech reflects people’s mental state and using a microphone sensor is a potential method for human–computer interaction. Speech recognition using this sensor is conducive to the diagnosis of mental illnesses. The gender difference of speakers affects the process of speech emotion recognition based on specific acoustic features, resulting in the decline of emotion recognition accuracy. Therefore, we believe that the accuracy of speech emotion recognition can be effectively improved by selecting different features of speech for emotion recognition based on the speech representations of different genders. In this paper, we propose a speech emotion recognition method based on gender classification. First, we use MLP to classify the original speech by gender. Second, based on the different acoustic features of male and female speech, we analyze the influence weights of multiple speech emotion features in male and female speech, and establish the optimal feature sets for male and female emotion recognition, respectively. Finally, we train and test CNN and BiLSTM, respectively, by using the male and the female speech emotion feature sets. The results show that the proposed emotion recognition models have an advantage in terms of average recognition accuracy compared with gender-mixed recognition models. MDPI 2023-01-25 /pmc/articles/PMC9921859/ /pubmed/36772395 http://dx.doi.org/10.3390/s23031355 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Zhang, Li-Min
Li, Yang
Zhang, Yue-Ting
Ng, Giap Weng
Leau, Yu-Beng
Yan, Hao
A Deep Learning Method Using Gender-Specific Features for Emotion Recognition
title A Deep Learning Method Using Gender-Specific Features for Emotion Recognition
title_full A Deep Learning Method Using Gender-Specific Features for Emotion Recognition
title_fullStr A Deep Learning Method Using Gender-Specific Features for Emotion Recognition
title_full_unstemmed A Deep Learning Method Using Gender-Specific Features for Emotion Recognition
title_short A Deep Learning Method Using Gender-Specific Features for Emotion Recognition
title_sort deep learning method using gender-specific features for emotion recognition
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9921859/
https://www.ncbi.nlm.nih.gov/pubmed/36772395
http://dx.doi.org/10.3390/s23031355
work_keys_str_mv AT zhanglimin adeeplearningmethodusinggenderspecificfeaturesforemotionrecognition
AT liyang adeeplearningmethodusinggenderspecificfeaturesforemotionrecognition
AT zhangyueting adeeplearningmethodusinggenderspecificfeaturesforemotionrecognition
AT nggiapweng adeeplearningmethodusinggenderspecificfeaturesforemotionrecognition
AT leauyubeng adeeplearningmethodusinggenderspecificfeaturesforemotionrecognition
AT yanhao adeeplearningmethodusinggenderspecificfeaturesforemotionrecognition
AT zhanglimin deeplearningmethodusinggenderspecificfeaturesforemotionrecognition
AT liyang deeplearningmethodusinggenderspecificfeaturesforemotionrecognition
AT zhangyueting deeplearningmethodusinggenderspecificfeaturesforemotionrecognition
AT nggiapweng deeplearningmethodusinggenderspecificfeaturesforemotionrecognition
AT leauyubeng deeplearningmethodusinggenderspecificfeaturesforemotionrecognition
AT yanhao deeplearningmethodusinggenderspecificfeaturesforemotionrecognition