Cargando…
A Causal Inference Model Explains Perception of the McGurk Effect and Other Incongruent Audiovisual Speech
Audiovisual speech integration combines information from auditory speech (talker’s voice) and visual speech (talker’s mouth movements) to improve perceptual accuracy. However, if the auditory and visual speech emanate from different talkers, integration decreases accuracy. Therefore, a key step in a...
Autores principales: | , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Public Library of Science
2017
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5312805/ https://www.ncbi.nlm.nih.gov/pubmed/28207734 http://dx.doi.org/10.1371/journal.pcbi.1005229 |
_version_ | 1782508256178995200 |
---|---|
author | Magnotti, John F. Beauchamp, Michael S. |
author_facet | Magnotti, John F. Beauchamp, Michael S. |
author_sort | Magnotti, John F. |
collection | PubMed |
description | Audiovisual speech integration combines information from auditory speech (talker’s voice) and visual speech (talker’s mouth movements) to improve perceptual accuracy. However, if the auditory and visual speech emanate from different talkers, integration decreases accuracy. Therefore, a key step in audiovisual speech perception is deciding whether auditory and visual speech have the same source, a process known as causal inference. A well-known illusion, the McGurk Effect, consists of incongruent audiovisual syllables, such as auditory “ba” + visual “ga” (AbaVga), that are integrated to produce a fused percept (“da”). This illusion raises two fundamental questions: first, given the incongruence between the auditory and visual syllables in the McGurk stimulus, why are they integrated; and second, why does the McGurk effect not occur for other, very similar syllables (e.g., AgaVba). We describe a simplified model of causal inference in multisensory speech perception (CIMS) that predicts the perception of arbitrary combinations of auditory and visual speech. We applied this model to behavioral data collected from 60 subjects perceiving both McGurk and non-McGurk incongruent speech stimuli. The CIMS model successfully predicted both the audiovisual integration observed for McGurk stimuli and the lack of integration observed for non-McGurk stimuli. An identical model without causal inference failed to accurately predict perception for either form of incongruent speech. The CIMS model uses causal inference to provide a computational framework for studying how the brain performs one of its most important tasks, integrating auditory and visual speech cues to allow us to communicate with others. |
format | Online Article Text |
id | pubmed-5312805 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2017 |
publisher | Public Library of Science |
record_format | MEDLINE/PubMed |
spelling | pubmed-53128052017-03-03 A Causal Inference Model Explains Perception of the McGurk Effect and Other Incongruent Audiovisual Speech Magnotti, John F. Beauchamp, Michael S. PLoS Comput Biol Research Article Audiovisual speech integration combines information from auditory speech (talker’s voice) and visual speech (talker’s mouth movements) to improve perceptual accuracy. However, if the auditory and visual speech emanate from different talkers, integration decreases accuracy. Therefore, a key step in audiovisual speech perception is deciding whether auditory and visual speech have the same source, a process known as causal inference. A well-known illusion, the McGurk Effect, consists of incongruent audiovisual syllables, such as auditory “ba” + visual “ga” (AbaVga), that are integrated to produce a fused percept (“da”). This illusion raises two fundamental questions: first, given the incongruence between the auditory and visual syllables in the McGurk stimulus, why are they integrated; and second, why does the McGurk effect not occur for other, very similar syllables (e.g., AgaVba). We describe a simplified model of causal inference in multisensory speech perception (CIMS) that predicts the perception of arbitrary combinations of auditory and visual speech. We applied this model to behavioral data collected from 60 subjects perceiving both McGurk and non-McGurk incongruent speech stimuli. The CIMS model successfully predicted both the audiovisual integration observed for McGurk stimuli and the lack of integration observed for non-McGurk stimuli. An identical model without causal inference failed to accurately predict perception for either form of incongruent speech. The CIMS model uses causal inference to provide a computational framework for studying how the brain performs one of its most important tasks, integrating auditory and visual speech cues to allow us to communicate with others. Public Library of Science 2017-02-16 /pmc/articles/PMC5312805/ /pubmed/28207734 http://dx.doi.org/10.1371/journal.pcbi.1005229 Text en © 2017 Magnotti, Beauchamp http://creativecommons.org/licenses/by/4.0/ This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. |
spellingShingle | Research Article Magnotti, John F. Beauchamp, Michael S. A Causal Inference Model Explains Perception of the McGurk Effect and Other Incongruent Audiovisual Speech |
title | A Causal Inference Model Explains Perception of the McGurk Effect and Other Incongruent Audiovisual Speech |
title_full | A Causal Inference Model Explains Perception of the McGurk Effect and Other Incongruent Audiovisual Speech |
title_fullStr | A Causal Inference Model Explains Perception of the McGurk Effect and Other Incongruent Audiovisual Speech |
title_full_unstemmed | A Causal Inference Model Explains Perception of the McGurk Effect and Other Incongruent Audiovisual Speech |
title_short | A Causal Inference Model Explains Perception of the McGurk Effect and Other Incongruent Audiovisual Speech |
title_sort | causal inference model explains perception of the mcgurk effect and other incongruent audiovisual speech |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5312805/ https://www.ncbi.nlm.nih.gov/pubmed/28207734 http://dx.doi.org/10.1371/journal.pcbi.1005229 |
work_keys_str_mv | AT magnottijohnf acausalinferencemodelexplainsperceptionofthemcgurkeffectandotherincongruentaudiovisualspeech AT beauchampmichaels acausalinferencemodelexplainsperceptionofthemcgurkeffectandotherincongruentaudiovisualspeech AT magnottijohnf causalinferencemodelexplainsperceptionofthemcgurkeffectandotherincongruentaudiovisualspeech AT beauchampmichaels causalinferencemodelexplainsperceptionofthemcgurkeffectandotherincongruentaudiovisualspeech |