Cargando…

Disrupted visual input unveils the computational details of artificial neural networks for face perception

BACKGROUND: Convolutional Neural Network (DCNN), with its great performance, has attracted attention of researchers from many disciplines. The studies of the DCNN and that of biological neural systems have inspired each other reciprocally. The brain-inspired neural networks not only achieve great pe...

Descripción completa

Detalles Bibliográficos
Autores principales: Li, Yi-Fan, Ying, Haojiang
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9744930/
https://www.ncbi.nlm.nih.gov/pubmed/36523327
http://dx.doi.org/10.3389/fncom.2022.1054421
_version_ 1784849037535477760
author Li, Yi-Fan
Ying, Haojiang
author_facet Li, Yi-Fan
Ying, Haojiang
author_sort Li, Yi-Fan
collection PubMed
description BACKGROUND: Convolutional Neural Network (DCNN), with its great performance, has attracted attention of researchers from many disciplines. The studies of the DCNN and that of biological neural systems have inspired each other reciprocally. The brain-inspired neural networks not only achieve great performance but also serve as a computational model of biological neural systems. METHODS: Here in this study, we trained and tested several typical DCNNs (AlexNet, VGG11, VGG13, VGG16, DenseNet, MobileNet, and EfficientNet) with a face ethnicity categorization task for experiment 1, and an emotion categorization task for experiment 2. We measured the performance of DCNNs by testing them with original and lossy visual inputs (various kinds of image occlusion) and compared their performance with human participants. Moreover, the class activation map (CAM) method allowed us to visualize the foci of the “attention” of these DCNNs. RESULTS: The results suggested that the VGG13 performed the best: Its performance closely resembled human participants in terms of psychophysics measurements, it utilized similar areas of visual inputs as humans, and it had the most consistent performance with inputs having various kinds of impairments. DISCUSSION: In general, we examined the processing mechanism of DCNNs using a new paradigm and found that VGG13 might be the most human-like DCNN in this task. This study also highlighted a possible paradigm to study and develop DCNNs using human perception as a benchmark.
format Online
Article
Text
id pubmed-9744930
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-97449302022-12-14 Disrupted visual input unveils the computational details of artificial neural networks for face perception Li, Yi-Fan Ying, Haojiang Front Comput Neurosci Neuroscience BACKGROUND: Convolutional Neural Network (DCNN), with its great performance, has attracted attention of researchers from many disciplines. The studies of the DCNN and that of biological neural systems have inspired each other reciprocally. The brain-inspired neural networks not only achieve great performance but also serve as a computational model of biological neural systems. METHODS: Here in this study, we trained and tested several typical DCNNs (AlexNet, VGG11, VGG13, VGG16, DenseNet, MobileNet, and EfficientNet) with a face ethnicity categorization task for experiment 1, and an emotion categorization task for experiment 2. We measured the performance of DCNNs by testing them with original and lossy visual inputs (various kinds of image occlusion) and compared their performance with human participants. Moreover, the class activation map (CAM) method allowed us to visualize the foci of the “attention” of these DCNNs. RESULTS: The results suggested that the VGG13 performed the best: Its performance closely resembled human participants in terms of psychophysics measurements, it utilized similar areas of visual inputs as humans, and it had the most consistent performance with inputs having various kinds of impairments. DISCUSSION: In general, we examined the processing mechanism of DCNNs using a new paradigm and found that VGG13 might be the most human-like DCNN in this task. This study also highlighted a possible paradigm to study and develop DCNNs using human perception as a benchmark. Frontiers Media S.A. 2022-11-29 /pmc/articles/PMC9744930/ /pubmed/36523327 http://dx.doi.org/10.3389/fncom.2022.1054421 Text en Copyright © 2022 Li and Ying. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Neuroscience
Li, Yi-Fan
Ying, Haojiang
Disrupted visual input unveils the computational details of artificial neural networks for face perception
title Disrupted visual input unveils the computational details of artificial neural networks for face perception
title_full Disrupted visual input unveils the computational details of artificial neural networks for face perception
title_fullStr Disrupted visual input unveils the computational details of artificial neural networks for face perception
title_full_unstemmed Disrupted visual input unveils the computational details of artificial neural networks for face perception
title_short Disrupted visual input unveils the computational details of artificial neural networks for face perception
title_sort disrupted visual input unveils the computational details of artificial neural networks for face perception
topic Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9744930/
https://www.ncbi.nlm.nih.gov/pubmed/36523327
http://dx.doi.org/10.3389/fncom.2022.1054421
work_keys_str_mv AT liyifan disruptedvisualinputunveilsthecomputationaldetailsofartificialneuralnetworksforfaceperception
AT yinghaojiang disruptedvisualinputunveilsthecomputationaldetailsofartificialneuralnetworksforfaceperception