Cargando…
End-to-end emotional speech recognition using acoustic model adaptation based on knowledge distillation
The end-to-end approach provides better performance in speech recognition compared to the traditional hidden Markov model-deep neural network (HMM-DNN)-based approach, but still shows poor performance in abnormal speech, especially emotional speech. The optimal solution is to build an acoustic model...
Autores principales: | , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Springer US
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9923643/ https://www.ncbi.nlm.nih.gov/pubmed/36817556 http://dx.doi.org/10.1007/s11042-023-14680-y |
_version_ | 1784887758998732800 |
---|---|
author | Yun, Hong-In Park, Jeong-Sik |
author_facet | Yun, Hong-In Park, Jeong-Sik |
author_sort | Yun, Hong-In |
collection | PubMed |
description | The end-to-end approach provides better performance in speech recognition compared to the traditional hidden Markov model-deep neural network (HMM-DNN)-based approach, but still shows poor performance in abnormal speech, especially emotional speech. The optimal solution is to build an acoustic model suitable for emotional speech recognition using only emotional speech data for each emotion, but it is impossible because it is difficult to collect sufficient amount of emotional speech data for each emotion. In this study, we propose a method to improve the emotional speech recognition performance by using the knowledge distillation technique that was originally introduced to decrease computational intensity of deep learning-based approaches by reducing the number of model parameters. In addition to its use as model compression, we employ this technique for model adaptation to emotional speech. The proposed method builds a basic model (referred to as a teacher model) with a number of model parameters using an amount of normal speech data, and then constructs a target model (referred to as a student model) with fewer model parameters using a small amount of emotional speech data (i.e., adaptation data). Since the student model is built with emotional speech data, it is expected to reflect the emotional characteristics of each emotion well. In the emotional speech recognition experiment, the student model maintained recognition performance regardless of the number of model parameters, whereas the teacher model degraded performance significantly as the number of parameters decreased, showing performance degradation of about 10% in word error rate. This result demonstrates that the student model serves as an acoustic model suitable for emotional speech recognition even though it does not require much emotional speech data. |
format | Online Article Text |
id | pubmed-9923643 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | Springer US |
record_format | MEDLINE/PubMed |
spelling | pubmed-99236432023-02-13 End-to-end emotional speech recognition using acoustic model adaptation based on knowledge distillation Yun, Hong-In Park, Jeong-Sik Multimed Tools Appl Article The end-to-end approach provides better performance in speech recognition compared to the traditional hidden Markov model-deep neural network (HMM-DNN)-based approach, but still shows poor performance in abnormal speech, especially emotional speech. The optimal solution is to build an acoustic model suitable for emotional speech recognition using only emotional speech data for each emotion, but it is impossible because it is difficult to collect sufficient amount of emotional speech data for each emotion. In this study, we propose a method to improve the emotional speech recognition performance by using the knowledge distillation technique that was originally introduced to decrease computational intensity of deep learning-based approaches by reducing the number of model parameters. In addition to its use as model compression, we employ this technique for model adaptation to emotional speech. The proposed method builds a basic model (referred to as a teacher model) with a number of model parameters using an amount of normal speech data, and then constructs a target model (referred to as a student model) with fewer model parameters using a small amount of emotional speech data (i.e., adaptation data). Since the student model is built with emotional speech data, it is expected to reflect the emotional characteristics of each emotion well. In the emotional speech recognition experiment, the student model maintained recognition performance regardless of the number of model parameters, whereas the teacher model degraded performance significantly as the number of parameters decreased, showing performance degradation of about 10% in word error rate. This result demonstrates that the student model serves as an acoustic model suitable for emotional speech recognition even though it does not require much emotional speech data. Springer US 2023-02-13 2023 /pmc/articles/PMC9923643/ /pubmed/36817556 http://dx.doi.org/10.1007/s11042-023-14680-y Text en © The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2023. Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law. This article is made available via the PMC Open Access Subset for unrestricted research re-use and secondary analysis in any form or by any means with acknowledgement of the original source. These permissions are granted for the duration of the World Health Organization (WHO) declaration of COVID-19 as a global pandemic. |
spellingShingle | Article Yun, Hong-In Park, Jeong-Sik End-to-end emotional speech recognition using acoustic model adaptation based on knowledge distillation |
title | End-to-end emotional speech recognition using acoustic model adaptation based on knowledge distillation |
title_full | End-to-end emotional speech recognition using acoustic model adaptation based on knowledge distillation |
title_fullStr | End-to-end emotional speech recognition using acoustic model adaptation based on knowledge distillation |
title_full_unstemmed | End-to-end emotional speech recognition using acoustic model adaptation based on knowledge distillation |
title_short | End-to-end emotional speech recognition using acoustic model adaptation based on knowledge distillation |
title_sort | end-to-end emotional speech recognition using acoustic model adaptation based on knowledge distillation |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9923643/ https://www.ncbi.nlm.nih.gov/pubmed/36817556 http://dx.doi.org/10.1007/s11042-023-14680-y |
work_keys_str_mv | AT yunhongin endtoendemotionalspeechrecognitionusingacousticmodeladaptationbasedonknowledgedistillation AT parkjeongsik endtoendemotionalspeechrecognitionusingacousticmodeladaptationbasedonknowledgedistillation |