Cargando…

Rapid Audiovisual Integration Guides Predictive Actions

Natural movements, such as catching a ball or capturing prey, typically involve multiple senses. Yet, laboratory studies on human movements commonly focus solely on vision and ignore sound. Here, we ask how visual and auditory signals are integrated to guide interceptive movements. Human observers t...

Descripción completa

Detalles Bibliográficos
Autores principales: Kreyenmeier, Philipp, Schroeger, Anna, Cañal-Bruland, Rouwen, Raab, Markus, Spering, Miriam
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Society for Neuroscience 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10464656/
https://www.ncbi.nlm.nih.gov/pubmed/37591732
http://dx.doi.org/10.1523/ENEURO.0134-23.2023
_version_ 1785098516236861440
author Kreyenmeier, Philipp
Schroeger, Anna
Cañal-Bruland, Rouwen
Raab, Markus
Spering, Miriam
author_facet Kreyenmeier, Philipp
Schroeger, Anna
Cañal-Bruland, Rouwen
Raab, Markus
Spering, Miriam
author_sort Kreyenmeier, Philipp
collection PubMed
description Natural movements, such as catching a ball or capturing prey, typically involve multiple senses. Yet, laboratory studies on human movements commonly focus solely on vision and ignore sound. Here, we ask how visual and auditory signals are integrated to guide interceptive movements. Human observers tracked the brief launch of a simulated baseball, randomly paired with batting sounds of varying intensities, and made a quick pointing movement at the ball. Movement end points revealed systematic overestimation of target speed when the ball launch was paired with a loud versus a quiet sound, although sound was never informative. This effect was modulated by the availability of visual information; sounds biased interception when the visual presentation duration of the ball was short. Amplitude of the first catch-up saccade, occurring ∼125 ms after target launch, revealed early integration of audiovisual information for trajectory estimation. This sound-induced bias was reversed during later predictive saccades when more visual information was available. Our findings suggest that auditory and visual signals are integrated to guide interception and that this integration process must occur early at a neural site that receives auditory and visual signals within an ultrashort time span.
format Online
Article
Text
id pubmed-10464656
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Society for Neuroscience
record_format MEDLINE/PubMed
spelling pubmed-104646562023-08-30 Rapid Audiovisual Integration Guides Predictive Actions Kreyenmeier, Philipp Schroeger, Anna Cañal-Bruland, Rouwen Raab, Markus Spering, Miriam eNeuro Research Article: New Research Natural movements, such as catching a ball or capturing prey, typically involve multiple senses. Yet, laboratory studies on human movements commonly focus solely on vision and ignore sound. Here, we ask how visual and auditory signals are integrated to guide interceptive movements. Human observers tracked the brief launch of a simulated baseball, randomly paired with batting sounds of varying intensities, and made a quick pointing movement at the ball. Movement end points revealed systematic overestimation of target speed when the ball launch was paired with a loud versus a quiet sound, although sound was never informative. This effect was modulated by the availability of visual information; sounds biased interception when the visual presentation duration of the ball was short. Amplitude of the first catch-up saccade, occurring ∼125 ms after target launch, revealed early integration of audiovisual information for trajectory estimation. This sound-induced bias was reversed during later predictive saccades when more visual information was available. Our findings suggest that auditory and visual signals are integrated to guide interception and that this integration process must occur early at a neural site that receives auditory and visual signals within an ultrashort time span. Society for Neuroscience 2023-08-25 /pmc/articles/PMC10464656/ /pubmed/37591732 http://dx.doi.org/10.1523/ENEURO.0134-23.2023 Text en Copyright © 2023 Kreyenmeier et al. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution 4.0 International license (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution and reproduction in any medium provided that the original work is properly attributed.
spellingShingle Research Article: New Research
Kreyenmeier, Philipp
Schroeger, Anna
Cañal-Bruland, Rouwen
Raab, Markus
Spering, Miriam
Rapid Audiovisual Integration Guides Predictive Actions
title Rapid Audiovisual Integration Guides Predictive Actions
title_full Rapid Audiovisual Integration Guides Predictive Actions
title_fullStr Rapid Audiovisual Integration Guides Predictive Actions
title_full_unstemmed Rapid Audiovisual Integration Guides Predictive Actions
title_short Rapid Audiovisual Integration Guides Predictive Actions
title_sort rapid audiovisual integration guides predictive actions
topic Research Article: New Research
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10464656/
https://www.ncbi.nlm.nih.gov/pubmed/37591732
http://dx.doi.org/10.1523/ENEURO.0134-23.2023
work_keys_str_mv AT kreyenmeierphilipp rapidaudiovisualintegrationguidespredictiveactions
AT schroegeranna rapidaudiovisualintegrationguidespredictiveactions
AT canalbrulandrouwen rapidaudiovisualintegrationguidespredictiveactions
AT raabmarkus rapidaudiovisualintegrationguidespredictiveactions
AT speringmiriam rapidaudiovisualintegrationguidespredictiveactions