Cargando…
Lipreading Architecture Based on Multiple Convolutional Neural Networks for Sentence-Level Visual Speech Recognition
In visual speech recognition (VSR), speech is transcribed using only visual information to interpret tongue and teeth movements. Recently, deep learning has shown outstanding performance in VSR, with accuracy exceeding that of lipreaders on benchmark datasets. However, several problems still exist w...
Autores principales: | Jeon, Sanghun, Elsharkawy, Ahmed, Kim, Mun Sang |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8747278/ https://www.ncbi.nlm.nih.gov/pubmed/35009612 http://dx.doi.org/10.3390/s22010072 |
Ejemplares similares
-
End-to-End Sentence-Level Multi-View Lipreading Architecture with Spatial Attention Module Integrated Multiple CNNs and Cascaded Local Self-Attention-CTC
por: Jeon, Sanghun, et al.
Publicado: (2022) -
Noise-Robust Multimodal Audio-Visual Speech Recognition System for Speech-Based Interaction Applications
por: Jeon, Sanghun, et al.
Publicado: (2022) -
Visual Cortical Entrainment to Motion and Categorical Speech Features during Silent Lipreading
por: O’Sullivan, Aisling E., et al.
Publicado: (2017) -
End-to-End Lip-Reading Open Cloud-Based Speech Architecture
por: Jeon, Sanghun, et al.
Publicado: (2022) -
The Neural Basis of Speech Perception through Lipreading and Manual Cues: Evidence from Deaf Native Users of Cued Speech
por: Aparicio, Mario, et al.
Publicado: (2017)