Cargando…

Contrast sensitivity functions in autoencoders

Three decades ago, Atick et al. suggested that human frequency sensitivity may emerge from the enhancement required for a more efficient analysis of retinal images. Here we reassess the relevance of low-level vision tasks in the explanation of the contrast sensitivity functions (CSFs) in light of 1)...

Descripción completa

Detalles Bibliográficos
Autores principales: Li, Qiang, Gomez-Villa, Alex, Bertalmío, Marcelo, Malo, Jesús
Formato: Online Artículo Texto
Lenguaje:English
Publicado: The Association for Research in Vision and Ophthalmology 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9145138/
https://www.ncbi.nlm.nih.gov/pubmed/35587354
http://dx.doi.org/10.1167/jov.22.6.8
_version_ 1784716219321942016
author Li, Qiang
Gomez-Villa, Alex
Bertalmío, Marcelo
Malo, Jesús
author_facet Li, Qiang
Gomez-Villa, Alex
Bertalmío, Marcelo
Malo, Jesús
author_sort Li, Qiang
collection PubMed
description Three decades ago, Atick et al. suggested that human frequency sensitivity may emerge from the enhancement required for a more efficient analysis of retinal images. Here we reassess the relevance of low-level vision tasks in the explanation of the contrast sensitivity functions (CSFs) in light of 1) the current trend of using artificial neural networks for studying vision, and 2) the current knowledge of retinal image representations. As a first contribution, we show that a very popular type of convolutional neural networks (CNNs), called autoencoders, may develop human-like CSFs in the spatiotemporal and chromatic dimensions when trained to perform some basic low-level vision tasks (like retinal noise and optical blur removal), but not others (like chromatic) adaptation or pure reconstruction after simple bottlenecks). As an illustrative example, the best CNN (in the considered set of simple architectures for enhancement of the retinal signal) reproduces the CSFs with a root mean square error of 11% of the maximum sensitivity. As a second contribution, we provide experimental evidence of the fact that, for some functional goals (at low abstraction level), deeper CNNs that are better in reaching the quantitative goal are actually worse in replicating human-like phenomena (such as the CSFs). This low-level result (for the explored networks) is not necessarily in contradiction with other works that report advantages of deeper nets in modeling higher level vision goals. However, in line with a growing body of literature, our results suggests another word of caution about CNNs in vision science because the use of simplified units or unrealistic architectures in goal optimization may be a limitation for the modeling and understanding of human vision.
format Online
Article
Text
id pubmed-9145138
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher The Association for Research in Vision and Ophthalmology
record_format MEDLINE/PubMed
spelling pubmed-91451382022-05-29 Contrast sensitivity functions in autoencoders Li, Qiang Gomez-Villa, Alex Bertalmío, Marcelo Malo, Jesús J Vis Article Three decades ago, Atick et al. suggested that human frequency sensitivity may emerge from the enhancement required for a more efficient analysis of retinal images. Here we reassess the relevance of low-level vision tasks in the explanation of the contrast sensitivity functions (CSFs) in light of 1) the current trend of using artificial neural networks for studying vision, and 2) the current knowledge of retinal image representations. As a first contribution, we show that a very popular type of convolutional neural networks (CNNs), called autoencoders, may develop human-like CSFs in the spatiotemporal and chromatic dimensions when trained to perform some basic low-level vision tasks (like retinal noise and optical blur removal), but not others (like chromatic) adaptation or pure reconstruction after simple bottlenecks). As an illustrative example, the best CNN (in the considered set of simple architectures for enhancement of the retinal signal) reproduces the CSFs with a root mean square error of 11% of the maximum sensitivity. As a second contribution, we provide experimental evidence of the fact that, for some functional goals (at low abstraction level), deeper CNNs that are better in reaching the quantitative goal are actually worse in replicating human-like phenomena (such as the CSFs). This low-level result (for the explored networks) is not necessarily in contradiction with other works that report advantages of deeper nets in modeling higher level vision goals. However, in line with a growing body of literature, our results suggests another word of caution about CNNs in vision science because the use of simplified units or unrealistic architectures in goal optimization may be a limitation for the modeling and understanding of human vision. The Association for Research in Vision and Ophthalmology 2022-05-19 /pmc/articles/PMC9145138/ /pubmed/35587354 http://dx.doi.org/10.1167/jov.22.6.8 Text en Copyright 2022 The Authors https://creativecommons.org/licenses/by-nc-nd/4.0/This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.
spellingShingle Article
Li, Qiang
Gomez-Villa, Alex
Bertalmío, Marcelo
Malo, Jesús
Contrast sensitivity functions in autoencoders
title Contrast sensitivity functions in autoencoders
title_full Contrast sensitivity functions in autoencoders
title_fullStr Contrast sensitivity functions in autoencoders
title_full_unstemmed Contrast sensitivity functions in autoencoders
title_short Contrast sensitivity functions in autoencoders
title_sort contrast sensitivity functions in autoencoders
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9145138/
https://www.ncbi.nlm.nih.gov/pubmed/35587354
http://dx.doi.org/10.1167/jov.22.6.8
work_keys_str_mv AT liqiang contrastsensitivityfunctionsinautoencoders
AT gomezvillaalex contrastsensitivityfunctionsinautoencoders
AT bertalmiomarcelo contrastsensitivityfunctionsinautoencoders
AT malojesus contrastsensitivityfunctionsinautoencoders