Cargando…

A dual foveal-peripheral visual processing model implements efficient saccade selection

We develop a visuomotor model that implements visual search as a focal accuracy-seeking policy, with the target's position and category drawn independently from a common generative process. Consistently with the anatomical separation between the ventral versus dorsal pathways, the model is comp...

Descripción completa

Detalles Bibliográficos
Autores principales: Daucé, Emmanuel, Albiges, Pierre, Perrinet, Laurent U.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: The Association for Research in Vision and Ophthalmology 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7443118/
http://dx.doi.org/10.1167/jov.20.8.22
_version_ 1783573571212148736
author Daucé, Emmanuel
Albiges, Pierre
Perrinet, Laurent U.
author_facet Daucé, Emmanuel
Albiges, Pierre
Perrinet, Laurent U.
author_sort Daucé, Emmanuel
collection PubMed
description We develop a visuomotor model that implements visual search as a focal accuracy-seeking policy, with the target's position and category drawn independently from a common generative process. Consistently with the anatomical separation between the ventral versus dorsal pathways, the model is composed of two pathways that respectively infer what to see and where to look. The “What” network is a classical deep learning classifier that only processes a small region around the center of fixation, providing a “foveal” accuracy. In contrast, the “Where” network processes the full visual field in a biomimetic fashion, using a log-polar retinotopic encoding, which is preserved up to the action selection level. In our model, the foveal accuracy is used as a monitoring signal to train the “Where” network, much like in the “actor/critic” framework. After training, the “Where” network provides an “accuracy map” that serves to guide the eye toward peripheral objects. Finally, the comparison of both networks’ accuracies amounts to either selecting a saccade or keeping the eye focused at the center to identify the target. We test this setup on a simple task of finding a digit in a large, cluttered image. Our simulation results demonstrate the effectiveness of this approach, increasing by one order of magnitude the radius of the visual field toward which the agent can detect and recognize a target, either through a single saccade or with multiple ones. Importantly, our log-polar treatment of the visual information exploits the strong compression rate performed at the sensory level, providing ways to implement visual search in a sublinear fashion, in contrast with mainstream computer vision.
format Online
Article
Text
id pubmed-7443118
institution National Center for Biotechnology Information
language English
publishDate 2020
publisher The Association for Research in Vision and Ophthalmology
record_format MEDLINE/PubMed
spelling pubmed-74431182020-09-01 A dual foveal-peripheral visual processing model implements efficient saccade selection Daucé, Emmanuel Albiges, Pierre Perrinet, Laurent U. J Vis Article We develop a visuomotor model that implements visual search as a focal accuracy-seeking policy, with the target's position and category drawn independently from a common generative process. Consistently with the anatomical separation between the ventral versus dorsal pathways, the model is composed of two pathways that respectively infer what to see and where to look. The “What” network is a classical deep learning classifier that only processes a small region around the center of fixation, providing a “foveal” accuracy. In contrast, the “Where” network processes the full visual field in a biomimetic fashion, using a log-polar retinotopic encoding, which is preserved up to the action selection level. In our model, the foveal accuracy is used as a monitoring signal to train the “Where” network, much like in the “actor/critic” framework. After training, the “Where” network provides an “accuracy map” that serves to guide the eye toward peripheral objects. Finally, the comparison of both networks’ accuracies amounts to either selecting a saccade or keeping the eye focused at the center to identify the target. We test this setup on a simple task of finding a digit in a large, cluttered image. Our simulation results demonstrate the effectiveness of this approach, increasing by one order of magnitude the radius of the visual field toward which the agent can detect and recognize a target, either through a single saccade or with multiple ones. Importantly, our log-polar treatment of the visual information exploits the strong compression rate performed at the sensory level, providing ways to implement visual search in a sublinear fashion, in contrast with mainstream computer vision. The Association for Research in Vision and Ophthalmology 2020-08-20 /pmc/articles/PMC7443118/ http://dx.doi.org/10.1167/jov.20.8.22 Text en Copyright 2020 The Authors http://creativecommons.org/licenses/by/4.0/ This work is licensed under a Creative Commons Attribution 4.0 International License.
spellingShingle Article
Daucé, Emmanuel
Albiges, Pierre
Perrinet, Laurent U.
A dual foveal-peripheral visual processing model implements efficient saccade selection
title A dual foveal-peripheral visual processing model implements efficient saccade selection
title_full A dual foveal-peripheral visual processing model implements efficient saccade selection
title_fullStr A dual foveal-peripheral visual processing model implements efficient saccade selection
title_full_unstemmed A dual foveal-peripheral visual processing model implements efficient saccade selection
title_short A dual foveal-peripheral visual processing model implements efficient saccade selection
title_sort dual foveal-peripheral visual processing model implements efficient saccade selection
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7443118/
http://dx.doi.org/10.1167/jov.20.8.22
work_keys_str_mv AT dauceemmanuel adualfovealperipheralvisualprocessingmodelimplementsefficientsaccadeselection
AT albigespierre adualfovealperipheralvisualprocessingmodelimplementsefficientsaccadeselection
AT perrinetlaurentu adualfovealperipheralvisualprocessingmodelimplementsefficientsaccadeselection
AT dauceemmanuel dualfovealperipheralvisualprocessingmodelimplementsefficientsaccadeselection
AT albigespierre dualfovealperipheralvisualprocessingmodelimplementsefficientsaccadeselection
AT perrinetlaurentu dualfovealperipheralvisualprocessingmodelimplementsefficientsaccadeselection