Cargando…

Improving Speech Recognition Performance in Noisy Environments by Enhancing Lip Reading Accuracy

The current accuracy of speech recognition can reach over 97% on different datasets, but in noisy environments, it is greatly reduced. Improving speech recognition performance in noisy environments is a challenging task. Due to the fact that visual information is not affected by noise, researchers o...

Descripción completa

Detalles Bibliográficos
Autores principales: Li, Dengshi, Gao, Yu, Zhu, Chenyi, Wang, Qianrui, Wang, Ruoxi
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9963439/
https://www.ncbi.nlm.nih.gov/pubmed/36850648
http://dx.doi.org/10.3390/s23042053
_version_ 1784896254297243648
author Li, Dengshi
Gao, Yu
Zhu, Chenyi
Wang, Qianrui
Wang, Ruoxi
author_facet Li, Dengshi
Gao, Yu
Zhu, Chenyi
Wang, Qianrui
Wang, Ruoxi
author_sort Li, Dengshi
collection PubMed
description The current accuracy of speech recognition can reach over 97% on different datasets, but in noisy environments, it is greatly reduced. Improving speech recognition performance in noisy environments is a challenging task. Due to the fact that visual information is not affected by noise, researchers often use lip information to help to improve speech recognition performance. This is where the performance of lip recognition and the effect of cross-modal fusion are particularly important. In this paper, we try to improve the accuracy of speech recognition in noisy environments by improving the lip reading performance and the cross-modal fusion effect. First, due to the same lip possibly containing multiple meanings, we constructed a one-to-many mapping relationship model between lips and speech allowing for the lip reading model to consider which articulations are represented from the input lip movements. Audio representations are also preserved by modeling the inter-relationships between paired audiovisual representations. At the inference stage, the preserved audio representations could be extracted from memory by the learned inter-relationships using only video input. Second, a joint cross-fusion model using the attention mechanism could effectively exploit complementary intermodal relationships, and the model calculates cross-attention weights on the basis of the correlations between joint feature representations and individual modalities. Lastly, our proposed model achieved a 4.0% reduction in WER in a −15 dB SNR environment compared to the baseline method, and a 10.1% reduction in WER compared to speech recognition. The experimental results show that our method could achieve a significant improvement over speech recognition models in different noise environments.
format Online
Article
Text
id pubmed-9963439
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-99634392023-02-26 Improving Speech Recognition Performance in Noisy Environments by Enhancing Lip Reading Accuracy Li, Dengshi Gao, Yu Zhu, Chenyi Wang, Qianrui Wang, Ruoxi Sensors (Basel) Article The current accuracy of speech recognition can reach over 97% on different datasets, but in noisy environments, it is greatly reduced. Improving speech recognition performance in noisy environments is a challenging task. Due to the fact that visual information is not affected by noise, researchers often use lip information to help to improve speech recognition performance. This is where the performance of lip recognition and the effect of cross-modal fusion are particularly important. In this paper, we try to improve the accuracy of speech recognition in noisy environments by improving the lip reading performance and the cross-modal fusion effect. First, due to the same lip possibly containing multiple meanings, we constructed a one-to-many mapping relationship model between lips and speech allowing for the lip reading model to consider which articulations are represented from the input lip movements. Audio representations are also preserved by modeling the inter-relationships between paired audiovisual representations. At the inference stage, the preserved audio representations could be extracted from memory by the learned inter-relationships using only video input. Second, a joint cross-fusion model using the attention mechanism could effectively exploit complementary intermodal relationships, and the model calculates cross-attention weights on the basis of the correlations between joint feature representations and individual modalities. Lastly, our proposed model achieved a 4.0% reduction in WER in a −15 dB SNR environment compared to the baseline method, and a 10.1% reduction in WER compared to speech recognition. The experimental results show that our method could achieve a significant improvement over speech recognition models in different noise environments. MDPI 2023-02-11 /pmc/articles/PMC9963439/ /pubmed/36850648 http://dx.doi.org/10.3390/s23042053 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Li, Dengshi
Gao, Yu
Zhu, Chenyi
Wang, Qianrui
Wang, Ruoxi
Improving Speech Recognition Performance in Noisy Environments by Enhancing Lip Reading Accuracy
title Improving Speech Recognition Performance in Noisy Environments by Enhancing Lip Reading Accuracy
title_full Improving Speech Recognition Performance in Noisy Environments by Enhancing Lip Reading Accuracy
title_fullStr Improving Speech Recognition Performance in Noisy Environments by Enhancing Lip Reading Accuracy
title_full_unstemmed Improving Speech Recognition Performance in Noisy Environments by Enhancing Lip Reading Accuracy
title_short Improving Speech Recognition Performance in Noisy Environments by Enhancing Lip Reading Accuracy
title_sort improving speech recognition performance in noisy environments by enhancing lip reading accuracy
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9963439/
https://www.ncbi.nlm.nih.gov/pubmed/36850648
http://dx.doi.org/10.3390/s23042053
work_keys_str_mv AT lidengshi improvingspeechrecognitionperformanceinnoisyenvironmentsbyenhancinglipreadingaccuracy
AT gaoyu improvingspeechrecognitionperformanceinnoisyenvironmentsbyenhancinglipreadingaccuracy
AT zhuchenyi improvingspeechrecognitionperformanceinnoisyenvironmentsbyenhancinglipreadingaccuracy
AT wangqianrui improvingspeechrecognitionperformanceinnoisyenvironmentsbyenhancinglipreadingaccuracy
AT wangruoxi improvingspeechrecognitionperformanceinnoisyenvironmentsbyenhancinglipreadingaccuracy