Cargando…
Audio-Visual Speech Timing Sensitivity Is Enhanced in Cluttered Conditions
Events encoded in separate sensory modalities, such as audition and vision, can seem to be synchronous across a relatively broad range of physical timing differences. This may suggest that the precision of audio-visual timing judgments is inherently poor. Here we show that this is not necessarily tr...
Autores principales: | , , , |
---|---|
Formato: | Texto |
Lenguaje: | English |
Publicado: |
Public Library of Science
2011
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3071827/ https://www.ncbi.nlm.nih.gov/pubmed/21494684 http://dx.doi.org/10.1371/journal.pone.0018309 |
_version_ | 1782201478682771456 |
---|---|
author | Roseboom, Warrick Nishida, Shin'ya Fujisaki, Waka Arnold, Derek H. |
author_facet | Roseboom, Warrick Nishida, Shin'ya Fujisaki, Waka Arnold, Derek H. |
author_sort | Roseboom, Warrick |
collection | PubMed |
description | Events encoded in separate sensory modalities, such as audition and vision, can seem to be synchronous across a relatively broad range of physical timing differences. This may suggest that the precision of audio-visual timing judgments is inherently poor. Here we show that this is not necessarily true. We contrast timing sensitivity for isolated streams of audio and visual speech, and for streams of audio and visual speech accompanied by additional, temporally offset, visual speech streams. We find that the precision with which synchronous streams of audio and visual speech are identified is enhanced by the presence of additional streams of asynchronous visual speech. Our data suggest that timing perception is shaped by selective grouping processes, which can result in enhanced precision in temporally cluttered environments. The imprecision suggested by previous studies might therefore be a consequence of examining isolated pairs of audio and visual events. We argue that when an isolated pair of cross-modal events is presented, they tend to group perceptually and to seem synchronous as a consequence. We have revealed greater precision by providing multiple visual signals, possibly allowing a single auditory speech stream to group selectively with the most synchronous visual candidate. The grouping processes we have identified might be important in daily life, such as when we attempt to follow a conversation in a crowded room. |
format | Text |
id | pubmed-3071827 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2011 |
publisher | Public Library of Science |
record_format | MEDLINE/PubMed |
spelling | pubmed-30718272011-04-14 Audio-Visual Speech Timing Sensitivity Is Enhanced in Cluttered Conditions Roseboom, Warrick Nishida, Shin'ya Fujisaki, Waka Arnold, Derek H. PLoS One Research Article Events encoded in separate sensory modalities, such as audition and vision, can seem to be synchronous across a relatively broad range of physical timing differences. This may suggest that the precision of audio-visual timing judgments is inherently poor. Here we show that this is not necessarily true. We contrast timing sensitivity for isolated streams of audio and visual speech, and for streams of audio and visual speech accompanied by additional, temporally offset, visual speech streams. We find that the precision with which synchronous streams of audio and visual speech are identified is enhanced by the presence of additional streams of asynchronous visual speech. Our data suggest that timing perception is shaped by selective grouping processes, which can result in enhanced precision in temporally cluttered environments. The imprecision suggested by previous studies might therefore be a consequence of examining isolated pairs of audio and visual events. We argue that when an isolated pair of cross-modal events is presented, they tend to group perceptually and to seem synchronous as a consequence. We have revealed greater precision by providing multiple visual signals, possibly allowing a single auditory speech stream to group selectively with the most synchronous visual candidate. The grouping processes we have identified might be important in daily life, such as when we attempt to follow a conversation in a crowded room. Public Library of Science 2011-04-06 /pmc/articles/PMC3071827/ /pubmed/21494684 http://dx.doi.org/10.1371/journal.pone.0018309 Text en Roseboom et al. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are properly credited. |
spellingShingle | Research Article Roseboom, Warrick Nishida, Shin'ya Fujisaki, Waka Arnold, Derek H. Audio-Visual Speech Timing Sensitivity Is Enhanced in Cluttered Conditions |
title | Audio-Visual Speech Timing Sensitivity Is Enhanced in Cluttered Conditions |
title_full | Audio-Visual Speech Timing Sensitivity Is Enhanced in Cluttered Conditions |
title_fullStr | Audio-Visual Speech Timing Sensitivity Is Enhanced in Cluttered Conditions |
title_full_unstemmed | Audio-Visual Speech Timing Sensitivity Is Enhanced in Cluttered Conditions |
title_short | Audio-Visual Speech Timing Sensitivity Is Enhanced in Cluttered Conditions |
title_sort | audio-visual speech timing sensitivity is enhanced in cluttered conditions |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3071827/ https://www.ncbi.nlm.nih.gov/pubmed/21494684 http://dx.doi.org/10.1371/journal.pone.0018309 |
work_keys_str_mv | AT roseboomwarrick audiovisualspeechtimingsensitivityisenhancedinclutteredconditions AT nishidashinya audiovisualspeechtimingsensitivityisenhancedinclutteredconditions AT fujisakiwaka audiovisualspeechtimingsensitivityisenhancedinclutteredconditions AT arnoldderekh audiovisualspeechtimingsensitivityisenhancedinclutteredconditions |