Cargando…

Inferring visual space from ultra-fine extra-retinal knowledge of gaze position

It has long been debated how humans resolve fine details and perceive a stable visual world despite the incessant fixational motion of their eyes. Current theories assume these processes to rely solely on the visual input to the retina, without contributions from motor and/or proprioceptive sources....

Descripción completa

Detalles Bibliográficos
Autores principales: Zhao, Zhetuo, Ahissar, Ehud, Victor, Jonathan D., Rucci, Michele
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Nature Publishing Group UK 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9845343/
https://www.ncbi.nlm.nih.gov/pubmed/36650146
http://dx.doi.org/10.1038/s41467-023-35834-4
_version_ 1784870885688082432
author Zhao, Zhetuo
Ahissar, Ehud
Victor, Jonathan D.
Rucci, Michele
author_facet Zhao, Zhetuo
Ahissar, Ehud
Victor, Jonathan D.
Rucci, Michele
author_sort Zhao, Zhetuo
collection PubMed
description It has long been debated how humans resolve fine details and perceive a stable visual world despite the incessant fixational motion of their eyes. Current theories assume these processes to rely solely on the visual input to the retina, without contributions from motor and/or proprioceptive sources. Here we show that contrary to this widespread assumption, the visual system has access to high-resolution extra-retinal knowledge of fixational eye motion and uses it to deduce spatial relations. Building on recent advances in gaze-contingent display control, we created a spatial discrimination task in which the stimulus configuration was entirely determined by oculomotor activity. Our results show that humans correctly infer geometrical relations in the absence of spatial information on the retina and accurately combine high-resolution extraretinal monitoring of gaze displacement with retinal signals. These findings reveal a sensory-motor strategy for encoding space, in which fine oculomotor knowledge is used to interpret the fixational input to the retina.
format Online
Article
Text
id pubmed-9845343
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Nature Publishing Group UK
record_format MEDLINE/PubMed
spelling pubmed-98453432023-01-19 Inferring visual space from ultra-fine extra-retinal knowledge of gaze position Zhao, Zhetuo Ahissar, Ehud Victor, Jonathan D. Rucci, Michele Nat Commun Article It has long been debated how humans resolve fine details and perceive a stable visual world despite the incessant fixational motion of their eyes. Current theories assume these processes to rely solely on the visual input to the retina, without contributions from motor and/or proprioceptive sources. Here we show that contrary to this widespread assumption, the visual system has access to high-resolution extra-retinal knowledge of fixational eye motion and uses it to deduce spatial relations. Building on recent advances in gaze-contingent display control, we created a spatial discrimination task in which the stimulus configuration was entirely determined by oculomotor activity. Our results show that humans correctly infer geometrical relations in the absence of spatial information on the retina and accurately combine high-resolution extraretinal monitoring of gaze displacement with retinal signals. These findings reveal a sensory-motor strategy for encoding space, in which fine oculomotor knowledge is used to interpret the fixational input to the retina. Nature Publishing Group UK 2023-01-17 /pmc/articles/PMC9845343/ /pubmed/36650146 http://dx.doi.org/10.1038/s41467-023-35834-4 Text en © The Author(s) 2023 https://creativecommons.org/licenses/by/4.0/Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) .
spellingShingle Article
Zhao, Zhetuo
Ahissar, Ehud
Victor, Jonathan D.
Rucci, Michele
Inferring visual space from ultra-fine extra-retinal knowledge of gaze position
title Inferring visual space from ultra-fine extra-retinal knowledge of gaze position
title_full Inferring visual space from ultra-fine extra-retinal knowledge of gaze position
title_fullStr Inferring visual space from ultra-fine extra-retinal knowledge of gaze position
title_full_unstemmed Inferring visual space from ultra-fine extra-retinal knowledge of gaze position
title_short Inferring visual space from ultra-fine extra-retinal knowledge of gaze position
title_sort inferring visual space from ultra-fine extra-retinal knowledge of gaze position
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9845343/
https://www.ncbi.nlm.nih.gov/pubmed/36650146
http://dx.doi.org/10.1038/s41467-023-35834-4
work_keys_str_mv AT zhaozhetuo inferringvisualspacefromultrafineextraretinalknowledgeofgazeposition
AT ahissarehud inferringvisualspacefromultrafineextraretinalknowledgeofgazeposition
AT victorjonathand inferringvisualspacefromultrafineextraretinalknowledgeofgazeposition
AT ruccimichele inferringvisualspacefromultrafineextraretinalknowledgeofgazeposition