Cargando…

Top-Down Predictions of Familiarity and Congruency in Audio-Visual Speech Perception at Neural Level

During speech perception, listeners rely on multimodal input and make use of both auditory and visual information. When presented with speech, for example syllables, the differences in brain responses to distinct stimuli are not, however, caused merely by the acoustic or visual features of the stimu...

Descripción completa

Detalles Bibliográficos
Autores principales: Kolozsvári, Orsolya B., Xu, Weiyong, Leppänen, Paavo H. T., Hämäläinen, Jarmo A.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2019
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6639789/
https://www.ncbi.nlm.nih.gov/pubmed/31354459
http://dx.doi.org/10.3389/fnhum.2019.00243
_version_ 1783436532675248128
author Kolozsvári, Orsolya B.
Xu, Weiyong
Leppänen, Paavo H. T.
Hämäläinen, Jarmo A.
author_facet Kolozsvári, Orsolya B.
Xu, Weiyong
Leppänen, Paavo H. T.
Hämäläinen, Jarmo A.
author_sort Kolozsvári, Orsolya B.
collection PubMed
description During speech perception, listeners rely on multimodal input and make use of both auditory and visual information. When presented with speech, for example syllables, the differences in brain responses to distinct stimuli are not, however, caused merely by the acoustic or visual features of the stimuli. The congruency of the auditory and visual information and the familiarity of a syllable, that is, whether it appears in the listener’s native language or not, also modulates brain responses. We investigated how the congruency and familiarity of the presented stimuli affect brain responses to audio-visual (AV) speech in 12 adult Finnish native speakers and 12 adult Chinese native speakers. They watched videos of a Chinese speaker pronouncing syllables (/pa/, /pha/, /ta/, /tha/, /fa/) during a magnetoencephalography (MEG) measurement where only /pa/ and /ta/ were part of Finnish phonology while all the stimuli were part of Chinese phonology. The stimuli were presented in audio-visual (congruent or incongruent), audio only, or visual only conditions. The brain responses were examined in five time-windows: 75–125, 150–200, 200–300, 300–400, and 400–600 ms. We found significant differences for the congruency comparison in the fourth time-window (300–400 ms) in both sensor and source level analysis. Larger responses were observed for the incongruent stimuli than for the congruent stimuli. For the familiarity comparisons no significant differences were found. The results are in line with earlier studies reporting on the modulation of brain responses for audio-visual congruency around 250–500 ms. This suggests a much stronger process for the general detection of a mismatch between predictions based on lip movements and the auditory signal than for the top-down modulation of brain responses based on phonological information.
format Online
Article
Text
id pubmed-6639789
institution National Center for Biotechnology Information
language English
publishDate 2019
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-66397892019-07-26 Top-Down Predictions of Familiarity and Congruency in Audio-Visual Speech Perception at Neural Level Kolozsvári, Orsolya B. Xu, Weiyong Leppänen, Paavo H. T. Hämäläinen, Jarmo A. Front Hum Neurosci Neuroscience During speech perception, listeners rely on multimodal input and make use of both auditory and visual information. When presented with speech, for example syllables, the differences in brain responses to distinct stimuli are not, however, caused merely by the acoustic or visual features of the stimuli. The congruency of the auditory and visual information and the familiarity of a syllable, that is, whether it appears in the listener’s native language or not, also modulates brain responses. We investigated how the congruency and familiarity of the presented stimuli affect brain responses to audio-visual (AV) speech in 12 adult Finnish native speakers and 12 adult Chinese native speakers. They watched videos of a Chinese speaker pronouncing syllables (/pa/, /pha/, /ta/, /tha/, /fa/) during a magnetoencephalography (MEG) measurement where only /pa/ and /ta/ were part of Finnish phonology while all the stimuli were part of Chinese phonology. The stimuli were presented in audio-visual (congruent or incongruent), audio only, or visual only conditions. The brain responses were examined in five time-windows: 75–125, 150–200, 200–300, 300–400, and 400–600 ms. We found significant differences for the congruency comparison in the fourth time-window (300–400 ms) in both sensor and source level analysis. Larger responses were observed for the incongruent stimuli than for the congruent stimuli. For the familiarity comparisons no significant differences were found. The results are in line with earlier studies reporting on the modulation of brain responses for audio-visual congruency around 250–500 ms. This suggests a much stronger process for the general detection of a mismatch between predictions based on lip movements and the auditory signal than for the top-down modulation of brain responses based on phonological information. Frontiers Media S.A. 2019-07-12 /pmc/articles/PMC6639789/ /pubmed/31354459 http://dx.doi.org/10.3389/fnhum.2019.00243 Text en Copyright © 2019 Kolozsvári, Xu, Leppänen and Hämäläinen. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Neuroscience
Kolozsvári, Orsolya B.
Xu, Weiyong
Leppänen, Paavo H. T.
Hämäläinen, Jarmo A.
Top-Down Predictions of Familiarity and Congruency in Audio-Visual Speech Perception at Neural Level
title Top-Down Predictions of Familiarity and Congruency in Audio-Visual Speech Perception at Neural Level
title_full Top-Down Predictions of Familiarity and Congruency in Audio-Visual Speech Perception at Neural Level
title_fullStr Top-Down Predictions of Familiarity and Congruency in Audio-Visual Speech Perception at Neural Level
title_full_unstemmed Top-Down Predictions of Familiarity and Congruency in Audio-Visual Speech Perception at Neural Level
title_short Top-Down Predictions of Familiarity and Congruency in Audio-Visual Speech Perception at Neural Level
title_sort top-down predictions of familiarity and congruency in audio-visual speech perception at neural level
topic Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6639789/
https://www.ncbi.nlm.nih.gov/pubmed/31354459
http://dx.doi.org/10.3389/fnhum.2019.00243
work_keys_str_mv AT kolozsvariorsolyab topdownpredictionsoffamiliarityandcongruencyinaudiovisualspeechperceptionatneurallevel
AT xuweiyong topdownpredictionsoffamiliarityandcongruencyinaudiovisualspeechperceptionatneurallevel
AT leppanenpaavoht topdownpredictionsoffamiliarityandcongruencyinaudiovisualspeechperceptionatneurallevel
AT hamalainenjarmoa topdownpredictionsoffamiliarityandcongruencyinaudiovisualspeechperceptionatneurallevel