Cargando…
Audio-Visual Speech Cue Combination
BACKGROUND: Different sources of sensory information can interact, often shaping what we think we have seen or heard. This can enhance the precision of perceptual decisions relative to those made on the basis of a single source of information. From a computational perspective, there are multiple rea...
Autores principales: | , , , |
---|---|
Formato: | Texto |
Lenguaje: | English |
Publicado: |
Public Library of Science
2010
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC2855706/ https://www.ncbi.nlm.nih.gov/pubmed/20419130 http://dx.doi.org/10.1371/journal.pone.0010217 |
_version_ | 1782180206138621952 |
---|---|
author | Arnold, Derek H. Tear, Morgan Schindel, Ryan Roseboom, Warrick |
author_facet | Arnold, Derek H. Tear, Morgan Schindel, Ryan Roseboom, Warrick |
author_sort | Arnold, Derek H. |
collection | PubMed |
description | BACKGROUND: Different sources of sensory information can interact, often shaping what we think we have seen or heard. This can enhance the precision of perceptual decisions relative to those made on the basis of a single source of information. From a computational perspective, there are multiple reasons why this might happen, and each predicts a different degree of enhanced precision. Relatively slight improvements can arise when perceptual decisions are made on the basis of multiple independent sensory estimates, as opposed to just one. These improvements can arise as a consequence of probability summation. Greater improvements can occur if two initially independent estimates are summated to form a single integrated code, especially if the summation is weighted in accordance with the variance associated with each independent estimate. This form of combination is often described as a Bayesian maximum likelihood estimate. Still greater improvements are possible if the two sources of information are encoded via a common physiological process. PRINCIPAL FINDINGS: Here we show that the provision of simultaneous audio and visual speech cues can result in substantial sensitivity improvements, relative to single sensory modality based decisions. The magnitude of the improvements is greater than can be predicted on the basis of either a Bayesian maximum likelihood estimate or a probability summation. CONCLUSION: Our data suggest that primary estimates of speech content are determined by a physiological process that takes input from both visual and auditory processing, resulting in greater sensitivity than would be possible if initially independent audio and visual estimates were formed and then subsequently combined. |
format | Text |
id | pubmed-2855706 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2010 |
publisher | Public Library of Science |
record_format | MEDLINE/PubMed |
spelling | pubmed-28557062010-04-23 Audio-Visual Speech Cue Combination Arnold, Derek H. Tear, Morgan Schindel, Ryan Roseboom, Warrick PLoS One Research Article BACKGROUND: Different sources of sensory information can interact, often shaping what we think we have seen or heard. This can enhance the precision of perceptual decisions relative to those made on the basis of a single source of information. From a computational perspective, there are multiple reasons why this might happen, and each predicts a different degree of enhanced precision. Relatively slight improvements can arise when perceptual decisions are made on the basis of multiple independent sensory estimates, as opposed to just one. These improvements can arise as a consequence of probability summation. Greater improvements can occur if two initially independent estimates are summated to form a single integrated code, especially if the summation is weighted in accordance with the variance associated with each independent estimate. This form of combination is often described as a Bayesian maximum likelihood estimate. Still greater improvements are possible if the two sources of information are encoded via a common physiological process. PRINCIPAL FINDINGS: Here we show that the provision of simultaneous audio and visual speech cues can result in substantial sensitivity improvements, relative to single sensory modality based decisions. The magnitude of the improvements is greater than can be predicted on the basis of either a Bayesian maximum likelihood estimate or a probability summation. CONCLUSION: Our data suggest that primary estimates of speech content are determined by a physiological process that takes input from both visual and auditory processing, resulting in greater sensitivity than would be possible if initially independent audio and visual estimates were formed and then subsequently combined. Public Library of Science 2010-04-16 /pmc/articles/PMC2855706/ /pubmed/20419130 http://dx.doi.org/10.1371/journal.pone.0010217 Text en Arnold et al. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are properly credited. |
spellingShingle | Research Article Arnold, Derek H. Tear, Morgan Schindel, Ryan Roseboom, Warrick Audio-Visual Speech Cue Combination |
title | Audio-Visual Speech Cue Combination |
title_full | Audio-Visual Speech Cue Combination |
title_fullStr | Audio-Visual Speech Cue Combination |
title_full_unstemmed | Audio-Visual Speech Cue Combination |
title_short | Audio-Visual Speech Cue Combination |
title_sort | audio-visual speech cue combination |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC2855706/ https://www.ncbi.nlm.nih.gov/pubmed/20419130 http://dx.doi.org/10.1371/journal.pone.0010217 |
work_keys_str_mv | AT arnoldderekh audiovisualspeechcuecombination AT tearmorgan audiovisualspeechcuecombination AT schindelryan audiovisualspeechcuecombination AT roseboomwarrick audiovisualspeechcuecombination |