Cargando…
Inferring visual space from ultra-fine extra-retinal knowledge of gaze position
It has long been debated how humans resolve fine details and perceive a stable visual world despite the incessant fixational motion of their eyes. Current theories assume these processes to rely solely on the visual input to the retina, without contributions from motor and/or proprioceptive sources....
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Nature Publishing Group UK
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9845343/ https://www.ncbi.nlm.nih.gov/pubmed/36650146 http://dx.doi.org/10.1038/s41467-023-35834-4 |
_version_ | 1784870885688082432 |
---|---|
author | Zhao, Zhetuo Ahissar, Ehud Victor, Jonathan D. Rucci, Michele |
author_facet | Zhao, Zhetuo Ahissar, Ehud Victor, Jonathan D. Rucci, Michele |
author_sort | Zhao, Zhetuo |
collection | PubMed |
description | It has long been debated how humans resolve fine details and perceive a stable visual world despite the incessant fixational motion of their eyes. Current theories assume these processes to rely solely on the visual input to the retina, without contributions from motor and/or proprioceptive sources. Here we show that contrary to this widespread assumption, the visual system has access to high-resolution extra-retinal knowledge of fixational eye motion and uses it to deduce spatial relations. Building on recent advances in gaze-contingent display control, we created a spatial discrimination task in which the stimulus configuration was entirely determined by oculomotor activity. Our results show that humans correctly infer geometrical relations in the absence of spatial information on the retina and accurately combine high-resolution extraretinal monitoring of gaze displacement with retinal signals. These findings reveal a sensory-motor strategy for encoding space, in which fine oculomotor knowledge is used to interpret the fixational input to the retina. |
format | Online Article Text |
id | pubmed-9845343 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | Nature Publishing Group UK |
record_format | MEDLINE/PubMed |
spelling | pubmed-98453432023-01-19 Inferring visual space from ultra-fine extra-retinal knowledge of gaze position Zhao, Zhetuo Ahissar, Ehud Victor, Jonathan D. Rucci, Michele Nat Commun Article It has long been debated how humans resolve fine details and perceive a stable visual world despite the incessant fixational motion of their eyes. Current theories assume these processes to rely solely on the visual input to the retina, without contributions from motor and/or proprioceptive sources. Here we show that contrary to this widespread assumption, the visual system has access to high-resolution extra-retinal knowledge of fixational eye motion and uses it to deduce spatial relations. Building on recent advances in gaze-contingent display control, we created a spatial discrimination task in which the stimulus configuration was entirely determined by oculomotor activity. Our results show that humans correctly infer geometrical relations in the absence of spatial information on the retina and accurately combine high-resolution extraretinal monitoring of gaze displacement with retinal signals. These findings reveal a sensory-motor strategy for encoding space, in which fine oculomotor knowledge is used to interpret the fixational input to the retina. Nature Publishing Group UK 2023-01-17 /pmc/articles/PMC9845343/ /pubmed/36650146 http://dx.doi.org/10.1038/s41467-023-35834-4 Text en © The Author(s) 2023 https://creativecommons.org/licenses/by/4.0/Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) . |
spellingShingle | Article Zhao, Zhetuo Ahissar, Ehud Victor, Jonathan D. Rucci, Michele Inferring visual space from ultra-fine extra-retinal knowledge of gaze position |
title | Inferring visual space from ultra-fine extra-retinal knowledge of gaze position |
title_full | Inferring visual space from ultra-fine extra-retinal knowledge of gaze position |
title_fullStr | Inferring visual space from ultra-fine extra-retinal knowledge of gaze position |
title_full_unstemmed | Inferring visual space from ultra-fine extra-retinal knowledge of gaze position |
title_short | Inferring visual space from ultra-fine extra-retinal knowledge of gaze position |
title_sort | inferring visual space from ultra-fine extra-retinal knowledge of gaze position |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9845343/ https://www.ncbi.nlm.nih.gov/pubmed/36650146 http://dx.doi.org/10.1038/s41467-023-35834-4 |
work_keys_str_mv | AT zhaozhetuo inferringvisualspacefromultrafineextraretinalknowledgeofgazeposition AT ahissarehud inferringvisualspacefromultrafineextraretinalknowledgeofgazeposition AT victorjonathand inferringvisualspacefromultrafineextraretinalknowledgeofgazeposition AT ruccimichele inferringvisualspacefromultrafineextraretinalknowledgeofgazeposition |