Cargando…
Experience with crossmodal statistics reduces the sensitivity for audio-visual temporal asynchrony
Bayesian models propose that multisensory integration depends on both sensory evidence (the likelihood) and priors indicating whether or not two inputs belong to the same event. The present study manipulated the prior for dynamic auditory and visual stimuli to co-occur and tested the predicted enhan...
Autores principales: | Habets, Boukje, Bruns, Patrick, Röder, Brigitte |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Nature Publishing Group UK
2017
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5431144/ https://www.ncbi.nlm.nih.gov/pubmed/28469137 http://dx.doi.org/10.1038/s41598-017-01252-y |
Ejemplares similares
-
Infants are superior in implicit crossmodal learning and use other learning mechanisms than adults
por: Rohlf, Sophie, et al.
Publicado: (2017) -
Feedback Modulates Audio-Visual Spatial Recalibration
por: Kramer, Alexander, et al.
Publicado: (2020) -
Crossmodal associations modulate multisensory spatial integration
por: Tong, Jonathan, et al.
Publicado: (2020) -
Audio–visual interaction in visual motion detection: Synchrony versus Asynchrony
por: Rosemann, Stephanie, et al.
Publicado: (2017) -
Asynchrony adaptation reveals neural population code for audio-visual timing
por: Roach, Neil W., et al.
Publicado: (2011)