Cargando…

A Causal Inference Model Explains Perception of the McGurk Effect and Other Incongruent Audiovisual Speech

Audiovisual speech integration combines information from auditory speech (talker’s voice) and visual speech (talker’s mouth movements) to improve perceptual accuracy. However, if the auditory and visual speech emanate from different talkers, integration decreases accuracy. Therefore, a key step in a...

Descripción completa

Detalles Bibliográficos
Autores principales: Magnotti, John F., Beauchamp, Michael S.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Public Library of Science 2017
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5312805/
https://www.ncbi.nlm.nih.gov/pubmed/28207734
http://dx.doi.org/10.1371/journal.pcbi.1005229
_version_ 1782508256178995200
author Magnotti, John F.
Beauchamp, Michael S.
author_facet Magnotti, John F.
Beauchamp, Michael S.
author_sort Magnotti, John F.
collection PubMed
description Audiovisual speech integration combines information from auditory speech (talker’s voice) and visual speech (talker’s mouth movements) to improve perceptual accuracy. However, if the auditory and visual speech emanate from different talkers, integration decreases accuracy. Therefore, a key step in audiovisual speech perception is deciding whether auditory and visual speech have the same source, a process known as causal inference. A well-known illusion, the McGurk Effect, consists of incongruent audiovisual syllables, such as auditory “ba” + visual “ga” (AbaVga), that are integrated to produce a fused percept (“da”). This illusion raises two fundamental questions: first, given the incongruence between the auditory and visual syllables in the McGurk stimulus, why are they integrated; and second, why does the McGurk effect not occur for other, very similar syllables (e.g., AgaVba). We describe a simplified model of causal inference in multisensory speech perception (CIMS) that predicts the perception of arbitrary combinations of auditory and visual speech. We applied this model to behavioral data collected from 60 subjects perceiving both McGurk and non-McGurk incongruent speech stimuli. The CIMS model successfully predicted both the audiovisual integration observed for McGurk stimuli and the lack of integration observed for non-McGurk stimuli. An identical model without causal inference failed to accurately predict perception for either form of incongruent speech. The CIMS model uses causal inference to provide a computational framework for studying how the brain performs one of its most important tasks, integrating auditory and visual speech cues to allow us to communicate with others.
format Online
Article
Text
id pubmed-5312805
institution National Center for Biotechnology Information
language English
publishDate 2017
publisher Public Library of Science
record_format MEDLINE/PubMed
spelling pubmed-53128052017-03-03 A Causal Inference Model Explains Perception of the McGurk Effect and Other Incongruent Audiovisual Speech Magnotti, John F. Beauchamp, Michael S. PLoS Comput Biol Research Article Audiovisual speech integration combines information from auditory speech (talker’s voice) and visual speech (talker’s mouth movements) to improve perceptual accuracy. However, if the auditory and visual speech emanate from different talkers, integration decreases accuracy. Therefore, a key step in audiovisual speech perception is deciding whether auditory and visual speech have the same source, a process known as causal inference. A well-known illusion, the McGurk Effect, consists of incongruent audiovisual syllables, such as auditory “ba” + visual “ga” (AbaVga), that are integrated to produce a fused percept (“da”). This illusion raises two fundamental questions: first, given the incongruence between the auditory and visual syllables in the McGurk stimulus, why are they integrated; and second, why does the McGurk effect not occur for other, very similar syllables (e.g., AgaVba). We describe a simplified model of causal inference in multisensory speech perception (CIMS) that predicts the perception of arbitrary combinations of auditory and visual speech. We applied this model to behavioral data collected from 60 subjects perceiving both McGurk and non-McGurk incongruent speech stimuli. The CIMS model successfully predicted both the audiovisual integration observed for McGurk stimuli and the lack of integration observed for non-McGurk stimuli. An identical model without causal inference failed to accurately predict perception for either form of incongruent speech. The CIMS model uses causal inference to provide a computational framework for studying how the brain performs one of its most important tasks, integrating auditory and visual speech cues to allow us to communicate with others. Public Library of Science 2017-02-16 /pmc/articles/PMC5312805/ /pubmed/28207734 http://dx.doi.org/10.1371/journal.pcbi.1005229 Text en © 2017 Magnotti, Beauchamp http://creativecommons.org/licenses/by/4.0/ This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
spellingShingle Research Article
Magnotti, John F.
Beauchamp, Michael S.
A Causal Inference Model Explains Perception of the McGurk Effect and Other Incongruent Audiovisual Speech
title A Causal Inference Model Explains Perception of the McGurk Effect and Other Incongruent Audiovisual Speech
title_full A Causal Inference Model Explains Perception of the McGurk Effect and Other Incongruent Audiovisual Speech
title_fullStr A Causal Inference Model Explains Perception of the McGurk Effect and Other Incongruent Audiovisual Speech
title_full_unstemmed A Causal Inference Model Explains Perception of the McGurk Effect and Other Incongruent Audiovisual Speech
title_short A Causal Inference Model Explains Perception of the McGurk Effect and Other Incongruent Audiovisual Speech
title_sort causal inference model explains perception of the mcgurk effect and other incongruent audiovisual speech
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5312805/
https://www.ncbi.nlm.nih.gov/pubmed/28207734
http://dx.doi.org/10.1371/journal.pcbi.1005229
work_keys_str_mv AT magnottijohnf acausalinferencemodelexplainsperceptionofthemcgurkeffectandotherincongruentaudiovisualspeech
AT beauchampmichaels acausalinferencemodelexplainsperceptionofthemcgurkeffectandotherincongruentaudiovisualspeech
AT magnottijohnf causalinferencemodelexplainsperceptionofthemcgurkeffectandotherincongruentaudiovisualspeech
AT beauchampmichaels causalinferencemodelexplainsperceptionofthemcgurkeffectandotherincongruentaudiovisualspeech