Cargando…
A dual foveal-peripheral visual processing model implements efficient saccade selection
We develop a visuomotor model that implements visual search as a focal accuracy-seeking policy, with the target's position and category drawn independently from a common generative process. Consistently with the anatomical separation between the ventral versus dorsal pathways, the model is comp...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
The Association for Research in Vision and Ophthalmology
2020
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7443118/ http://dx.doi.org/10.1167/jov.20.8.22 |
_version_ | 1783573571212148736 |
---|---|
author | Daucé, Emmanuel Albiges, Pierre Perrinet, Laurent U. |
author_facet | Daucé, Emmanuel Albiges, Pierre Perrinet, Laurent U. |
author_sort | Daucé, Emmanuel |
collection | PubMed |
description | We develop a visuomotor model that implements visual search as a focal accuracy-seeking policy, with the target's position and category drawn independently from a common generative process. Consistently with the anatomical separation between the ventral versus dorsal pathways, the model is composed of two pathways that respectively infer what to see and where to look. The “What” network is a classical deep learning classifier that only processes a small region around the center of fixation, providing a “foveal” accuracy. In contrast, the “Where” network processes the full visual field in a biomimetic fashion, using a log-polar retinotopic encoding, which is preserved up to the action selection level. In our model, the foveal accuracy is used as a monitoring signal to train the “Where” network, much like in the “actor/critic” framework. After training, the “Where” network provides an “accuracy map” that serves to guide the eye toward peripheral objects. Finally, the comparison of both networks’ accuracies amounts to either selecting a saccade or keeping the eye focused at the center to identify the target. We test this setup on a simple task of finding a digit in a large, cluttered image. Our simulation results demonstrate the effectiveness of this approach, increasing by one order of magnitude the radius of the visual field toward which the agent can detect and recognize a target, either through a single saccade or with multiple ones. Importantly, our log-polar treatment of the visual information exploits the strong compression rate performed at the sensory level, providing ways to implement visual search in a sublinear fashion, in contrast with mainstream computer vision. |
format | Online Article Text |
id | pubmed-7443118 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2020 |
publisher | The Association for Research in Vision and Ophthalmology |
record_format | MEDLINE/PubMed |
spelling | pubmed-74431182020-09-01 A dual foveal-peripheral visual processing model implements efficient saccade selection Daucé, Emmanuel Albiges, Pierre Perrinet, Laurent U. J Vis Article We develop a visuomotor model that implements visual search as a focal accuracy-seeking policy, with the target's position and category drawn independently from a common generative process. Consistently with the anatomical separation between the ventral versus dorsal pathways, the model is composed of two pathways that respectively infer what to see and where to look. The “What” network is a classical deep learning classifier that only processes a small region around the center of fixation, providing a “foveal” accuracy. In contrast, the “Where” network processes the full visual field in a biomimetic fashion, using a log-polar retinotopic encoding, which is preserved up to the action selection level. In our model, the foveal accuracy is used as a monitoring signal to train the “Where” network, much like in the “actor/critic” framework. After training, the “Where” network provides an “accuracy map” that serves to guide the eye toward peripheral objects. Finally, the comparison of both networks’ accuracies amounts to either selecting a saccade or keeping the eye focused at the center to identify the target. We test this setup on a simple task of finding a digit in a large, cluttered image. Our simulation results demonstrate the effectiveness of this approach, increasing by one order of magnitude the radius of the visual field toward which the agent can detect and recognize a target, either through a single saccade or with multiple ones. Importantly, our log-polar treatment of the visual information exploits the strong compression rate performed at the sensory level, providing ways to implement visual search in a sublinear fashion, in contrast with mainstream computer vision. The Association for Research in Vision and Ophthalmology 2020-08-20 /pmc/articles/PMC7443118/ http://dx.doi.org/10.1167/jov.20.8.22 Text en Copyright 2020 The Authors http://creativecommons.org/licenses/by/4.0/ This work is licensed under a Creative Commons Attribution 4.0 International License. |
spellingShingle | Article Daucé, Emmanuel Albiges, Pierre Perrinet, Laurent U. A dual foveal-peripheral visual processing model implements efficient saccade selection |
title | A dual foveal-peripheral visual processing model implements efficient saccade selection |
title_full | A dual foveal-peripheral visual processing model implements efficient saccade selection |
title_fullStr | A dual foveal-peripheral visual processing model implements efficient saccade selection |
title_full_unstemmed | A dual foveal-peripheral visual processing model implements efficient saccade selection |
title_short | A dual foveal-peripheral visual processing model implements efficient saccade selection |
title_sort | dual foveal-peripheral visual processing model implements efficient saccade selection |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7443118/ http://dx.doi.org/10.1167/jov.20.8.22 |
work_keys_str_mv | AT dauceemmanuel adualfovealperipheralvisualprocessingmodelimplementsefficientsaccadeselection AT albigespierre adualfovealperipheralvisualprocessingmodelimplementsefficientsaccadeselection AT perrinetlaurentu adualfovealperipheralvisualprocessingmodelimplementsefficientsaccadeselection AT dauceemmanuel dualfovealperipheralvisualprocessingmodelimplementsefficientsaccadeselection AT albigespierre dualfovealperipheralvisualprocessingmodelimplementsefficientsaccadeselection AT perrinetlaurentu dualfovealperipheralvisualprocessingmodelimplementsefficientsaccadeselection |