Cargando…
Real-Time Control of an Articulatory-Based Speech Synthesizer for Brain Computer Interfaces
Restoring natural speech in paralyzed and aphasic people could be achieved using a Brain-Computer Interface (BCI) controlling a speech synthesizer in real-time. To reach this goal, a prerequisite is to develop a speech synthesizer producing intelligible speech in real-time with a reasonable number o...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Public Library of Science
2016
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5120792/ https://www.ncbi.nlm.nih.gov/pubmed/27880768 http://dx.doi.org/10.1371/journal.pcbi.1005119 |
_version_ | 1782469298279677952 |
---|---|
author | Bocquelet, Florent Hueber, Thomas Girin, Laurent Savariaux, Christophe Yvert, Blaise |
author_facet | Bocquelet, Florent Hueber, Thomas Girin, Laurent Savariaux, Christophe Yvert, Blaise |
author_sort | Bocquelet, Florent |
collection | PubMed |
description | Restoring natural speech in paralyzed and aphasic people could be achieved using a Brain-Computer Interface (BCI) controlling a speech synthesizer in real-time. To reach this goal, a prerequisite is to develop a speech synthesizer producing intelligible speech in real-time with a reasonable number of control parameters. We present here an articulatory-based speech synthesizer that can be controlled in real-time for future BCI applications. This synthesizer converts movements of the main speech articulators (tongue, jaw, velum, and lips) into intelligible speech. The articulatory-to-acoustic mapping is performed using a deep neural network (DNN) trained on electromagnetic articulography (EMA) data recorded on a reference speaker synchronously with the produced speech signal. This DNN is then used in both offline and online modes to map the position of sensors glued on different speech articulators into acoustic parameters that are further converted into an audio signal using a vocoder. In offline mode, highly intelligible speech could be obtained as assessed by perceptual evaluation performed by 12 listeners. Then, to anticipate future BCI applications, we further assessed the real-time control of the synthesizer by both the reference speaker and new speakers, in a closed-loop paradigm using EMA data recorded in real time. A short calibration period was used to compensate for differences in sensor positions and articulatory differences between new speakers and the reference speaker. We found that real-time synthesis of vowels and consonants was possible with good intelligibility. In conclusion, these results open to future speech BCI applications using such articulatory-based speech synthesizer. |
format | Online Article Text |
id | pubmed-5120792 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2016 |
publisher | Public Library of Science |
record_format | MEDLINE/PubMed |
spelling | pubmed-51207922016-12-15 Real-Time Control of an Articulatory-Based Speech Synthesizer for Brain Computer Interfaces Bocquelet, Florent Hueber, Thomas Girin, Laurent Savariaux, Christophe Yvert, Blaise PLoS Comput Biol Research Article Restoring natural speech in paralyzed and aphasic people could be achieved using a Brain-Computer Interface (BCI) controlling a speech synthesizer in real-time. To reach this goal, a prerequisite is to develop a speech synthesizer producing intelligible speech in real-time with a reasonable number of control parameters. We present here an articulatory-based speech synthesizer that can be controlled in real-time for future BCI applications. This synthesizer converts movements of the main speech articulators (tongue, jaw, velum, and lips) into intelligible speech. The articulatory-to-acoustic mapping is performed using a deep neural network (DNN) trained on electromagnetic articulography (EMA) data recorded on a reference speaker synchronously with the produced speech signal. This DNN is then used in both offline and online modes to map the position of sensors glued on different speech articulators into acoustic parameters that are further converted into an audio signal using a vocoder. In offline mode, highly intelligible speech could be obtained as assessed by perceptual evaluation performed by 12 listeners. Then, to anticipate future BCI applications, we further assessed the real-time control of the synthesizer by both the reference speaker and new speakers, in a closed-loop paradigm using EMA data recorded in real time. A short calibration period was used to compensate for differences in sensor positions and articulatory differences between new speakers and the reference speaker. We found that real-time synthesis of vowels and consonants was possible with good intelligibility. In conclusion, these results open to future speech BCI applications using such articulatory-based speech synthesizer. Public Library of Science 2016-11-23 /pmc/articles/PMC5120792/ /pubmed/27880768 http://dx.doi.org/10.1371/journal.pcbi.1005119 Text en © 2016 Bocquelet et al http://creativecommons.org/licenses/by/4.0/ This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. |
spellingShingle | Research Article Bocquelet, Florent Hueber, Thomas Girin, Laurent Savariaux, Christophe Yvert, Blaise Real-Time Control of an Articulatory-Based Speech Synthesizer for Brain Computer Interfaces |
title | Real-Time Control of an Articulatory-Based Speech Synthesizer for Brain Computer Interfaces |
title_full | Real-Time Control of an Articulatory-Based Speech Synthesizer for Brain Computer Interfaces |
title_fullStr | Real-Time Control of an Articulatory-Based Speech Synthesizer for Brain Computer Interfaces |
title_full_unstemmed | Real-Time Control of an Articulatory-Based Speech Synthesizer for Brain Computer Interfaces |
title_short | Real-Time Control of an Articulatory-Based Speech Synthesizer for Brain Computer Interfaces |
title_sort | real-time control of an articulatory-based speech synthesizer for brain computer interfaces |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5120792/ https://www.ncbi.nlm.nih.gov/pubmed/27880768 http://dx.doi.org/10.1371/journal.pcbi.1005119 |
work_keys_str_mv | AT bocqueletflorent realtimecontrolofanarticulatorybasedspeechsynthesizerforbraincomputerinterfaces AT hueberthomas realtimecontrolofanarticulatorybasedspeechsynthesizerforbraincomputerinterfaces AT girinlaurent realtimecontrolofanarticulatorybasedspeechsynthesizerforbraincomputerinterfaces AT savariauxchristophe realtimecontrolofanarticulatorybasedspeechsynthesizerforbraincomputerinterfaces AT yvertblaise realtimecontrolofanarticulatorybasedspeechsynthesizerforbraincomputerinterfaces |