Cargando…

Audio-Visual Speech Timing Sensitivity Is Enhanced in Cluttered Conditions

Events encoded in separate sensory modalities, such as audition and vision, can seem to be synchronous across a relatively broad range of physical timing differences. This may suggest that the precision of audio-visual timing judgments is inherently poor. Here we show that this is not necessarily tr...

Descripción completa

Detalles Bibliográficos
Autores principales: Roseboom, Warrick, Nishida, Shin'ya, Fujisaki, Waka, Arnold, Derek H.
Formato: Texto
Lenguaje:English
Publicado: Public Library of Science 2011
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3071827/
https://www.ncbi.nlm.nih.gov/pubmed/21494684
http://dx.doi.org/10.1371/journal.pone.0018309
_version_ 1782201478682771456
author Roseboom, Warrick
Nishida, Shin'ya
Fujisaki, Waka
Arnold, Derek H.
author_facet Roseboom, Warrick
Nishida, Shin'ya
Fujisaki, Waka
Arnold, Derek H.
author_sort Roseboom, Warrick
collection PubMed
description Events encoded in separate sensory modalities, such as audition and vision, can seem to be synchronous across a relatively broad range of physical timing differences. This may suggest that the precision of audio-visual timing judgments is inherently poor. Here we show that this is not necessarily true. We contrast timing sensitivity for isolated streams of audio and visual speech, and for streams of audio and visual speech accompanied by additional, temporally offset, visual speech streams. We find that the precision with which synchronous streams of audio and visual speech are identified is enhanced by the presence of additional streams of asynchronous visual speech. Our data suggest that timing perception is shaped by selective grouping processes, which can result in enhanced precision in temporally cluttered environments. The imprecision suggested by previous studies might therefore be a consequence of examining isolated pairs of audio and visual events. We argue that when an isolated pair of cross-modal events is presented, they tend to group perceptually and to seem synchronous as a consequence. We have revealed greater precision by providing multiple visual signals, possibly allowing a single auditory speech stream to group selectively with the most synchronous visual candidate. The grouping processes we have identified might be important in daily life, such as when we attempt to follow a conversation in a crowded room.
format Text
id pubmed-3071827
institution National Center for Biotechnology Information
language English
publishDate 2011
publisher Public Library of Science
record_format MEDLINE/PubMed
spelling pubmed-30718272011-04-14 Audio-Visual Speech Timing Sensitivity Is Enhanced in Cluttered Conditions Roseboom, Warrick Nishida, Shin'ya Fujisaki, Waka Arnold, Derek H. PLoS One Research Article Events encoded in separate sensory modalities, such as audition and vision, can seem to be synchronous across a relatively broad range of physical timing differences. This may suggest that the precision of audio-visual timing judgments is inherently poor. Here we show that this is not necessarily true. We contrast timing sensitivity for isolated streams of audio and visual speech, and for streams of audio and visual speech accompanied by additional, temporally offset, visual speech streams. We find that the precision with which synchronous streams of audio and visual speech are identified is enhanced by the presence of additional streams of asynchronous visual speech. Our data suggest that timing perception is shaped by selective grouping processes, which can result in enhanced precision in temporally cluttered environments. The imprecision suggested by previous studies might therefore be a consequence of examining isolated pairs of audio and visual events. We argue that when an isolated pair of cross-modal events is presented, they tend to group perceptually and to seem synchronous as a consequence. We have revealed greater precision by providing multiple visual signals, possibly allowing a single auditory speech stream to group selectively with the most synchronous visual candidate. The grouping processes we have identified might be important in daily life, such as when we attempt to follow a conversation in a crowded room. Public Library of Science 2011-04-06 /pmc/articles/PMC3071827/ /pubmed/21494684 http://dx.doi.org/10.1371/journal.pone.0018309 Text en Roseboom et al. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are properly credited.
spellingShingle Research Article
Roseboom, Warrick
Nishida, Shin'ya
Fujisaki, Waka
Arnold, Derek H.
Audio-Visual Speech Timing Sensitivity Is Enhanced in Cluttered Conditions
title Audio-Visual Speech Timing Sensitivity Is Enhanced in Cluttered Conditions
title_full Audio-Visual Speech Timing Sensitivity Is Enhanced in Cluttered Conditions
title_fullStr Audio-Visual Speech Timing Sensitivity Is Enhanced in Cluttered Conditions
title_full_unstemmed Audio-Visual Speech Timing Sensitivity Is Enhanced in Cluttered Conditions
title_short Audio-Visual Speech Timing Sensitivity Is Enhanced in Cluttered Conditions
title_sort audio-visual speech timing sensitivity is enhanced in cluttered conditions
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3071827/
https://www.ncbi.nlm.nih.gov/pubmed/21494684
http://dx.doi.org/10.1371/journal.pone.0018309
work_keys_str_mv AT roseboomwarrick audiovisualspeechtimingsensitivityisenhancedinclutteredconditions
AT nishidashinya audiovisualspeechtimingsensitivityisenhancedinclutteredconditions
AT fujisakiwaka audiovisualspeechtimingsensitivityisenhancedinclutteredconditions
AT arnoldderekh audiovisualspeechtimingsensitivityisenhancedinclutteredconditions