Cargando…

Contributions of local speech encoding and functional connectivity to audio-visual speech perception

Seeing a speaker’s face enhances speech intelligibility in adverse environments. We investigated the underlying network mechanisms by quantifying local speech representations and directed connectivity in MEG data obtained while human participants listened to speech of varying acoustic SNR and visual...

Descripción completa

Detalles Bibliográficos
Autores principales: Giordano, Bruno L, Ince, Robin A A, Gross, Joachim, Schyns, Philippe G, Panzeri, Stefano, Kayser, Christoph
Formato: Online Artículo Texto
Lenguaje:English
Publicado: eLife Sciences Publications, Ltd 2017
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5462535/
https://www.ncbi.nlm.nih.gov/pubmed/28590903
http://dx.doi.org/10.7554/eLife.24763
_version_ 1783242530250293248
author Giordano, Bruno L
Ince, Robin A A
Gross, Joachim
Schyns, Philippe G
Panzeri, Stefano
Kayser, Christoph
author_facet Giordano, Bruno L
Ince, Robin A A
Gross, Joachim
Schyns, Philippe G
Panzeri, Stefano
Kayser, Christoph
author_sort Giordano, Bruno L
collection PubMed
description Seeing a speaker’s face enhances speech intelligibility in adverse environments. We investigated the underlying network mechanisms by quantifying local speech representations and directed connectivity in MEG data obtained while human participants listened to speech of varying acoustic SNR and visual context. During high acoustic SNR speech encoding by temporally entrained brain activity was strong in temporal and inferior frontal cortex, while during low SNR strong entrainment emerged in premotor and superior frontal cortex. These changes in local encoding were accompanied by changes in directed connectivity along the ventral stream and the auditory-premotor axis. Importantly, the behavioral benefit arising from seeing the speaker’s face was not predicted by changes in local encoding but rather by enhanced functional connectivity between temporal and inferior frontal cortex. Our results demonstrate a role of auditory-frontal interactions in visual speech representations and suggest that functional connectivity along the ventral pathway facilitates speech comprehension in multisensory environments. DOI: http://dx.doi.org/10.7554/eLife.24763.001
format Online
Article
Text
id pubmed-5462535
institution National Center for Biotechnology Information
language English
publishDate 2017
publisher eLife Sciences Publications, Ltd
record_format MEDLINE/PubMed
spelling pubmed-54625352017-06-09 Contributions of local speech encoding and functional connectivity to audio-visual speech perception Giordano, Bruno L Ince, Robin A A Gross, Joachim Schyns, Philippe G Panzeri, Stefano Kayser, Christoph eLife Neuroscience Seeing a speaker’s face enhances speech intelligibility in adverse environments. We investigated the underlying network mechanisms by quantifying local speech representations and directed connectivity in MEG data obtained while human participants listened to speech of varying acoustic SNR and visual context. During high acoustic SNR speech encoding by temporally entrained brain activity was strong in temporal and inferior frontal cortex, while during low SNR strong entrainment emerged in premotor and superior frontal cortex. These changes in local encoding were accompanied by changes in directed connectivity along the ventral stream and the auditory-premotor axis. Importantly, the behavioral benefit arising from seeing the speaker’s face was not predicted by changes in local encoding but rather by enhanced functional connectivity between temporal and inferior frontal cortex. Our results demonstrate a role of auditory-frontal interactions in visual speech representations and suggest that functional connectivity along the ventral pathway facilitates speech comprehension in multisensory environments. DOI: http://dx.doi.org/10.7554/eLife.24763.001 eLife Sciences Publications, Ltd 2017-06-07 /pmc/articles/PMC5462535/ /pubmed/28590903 http://dx.doi.org/10.7554/eLife.24763 Text en © 2017, Giordano et al http://creativecommons.org/licenses/by/4.0/ This article is distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use and redistribution provided that the original author and source are credited.
spellingShingle Neuroscience
Giordano, Bruno L
Ince, Robin A A
Gross, Joachim
Schyns, Philippe G
Panzeri, Stefano
Kayser, Christoph
Contributions of local speech encoding and functional connectivity to audio-visual speech perception
title Contributions of local speech encoding and functional connectivity to audio-visual speech perception
title_full Contributions of local speech encoding and functional connectivity to audio-visual speech perception
title_fullStr Contributions of local speech encoding and functional connectivity to audio-visual speech perception
title_full_unstemmed Contributions of local speech encoding and functional connectivity to audio-visual speech perception
title_short Contributions of local speech encoding and functional connectivity to audio-visual speech perception
title_sort contributions of local speech encoding and functional connectivity to audio-visual speech perception
topic Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5462535/
https://www.ncbi.nlm.nih.gov/pubmed/28590903
http://dx.doi.org/10.7554/eLife.24763
work_keys_str_mv AT giordanobrunol contributionsoflocalspeechencodingandfunctionalconnectivitytoaudiovisualspeechperception
AT incerobinaa contributionsoflocalspeechencodingandfunctionalconnectivitytoaudiovisualspeechperception
AT grossjoachim contributionsoflocalspeechencodingandfunctionalconnectivitytoaudiovisualspeechperception
AT schynsphilippeg contributionsoflocalspeechencodingandfunctionalconnectivitytoaudiovisualspeechperception
AT panzeristefano contributionsoflocalspeechencodingandfunctionalconnectivitytoaudiovisualspeechperception
AT kayserchristoph contributionsoflocalspeechencodingandfunctionalconnectivitytoaudiovisualspeechperception