Cargando…
Disrupted visual input unveils the computational details of artificial neural networks for face perception
BACKGROUND: Convolutional Neural Network (DCNN), with its great performance, has attracted attention of researchers from many disciplines. The studies of the DCNN and that of biological neural systems have inspired each other reciprocally. The brain-inspired neural networks not only achieve great pe...
Autores principales: | , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Frontiers Media S.A.
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9744930/ https://www.ncbi.nlm.nih.gov/pubmed/36523327 http://dx.doi.org/10.3389/fncom.2022.1054421 |
_version_ | 1784849037535477760 |
---|---|
author | Li, Yi-Fan Ying, Haojiang |
author_facet | Li, Yi-Fan Ying, Haojiang |
author_sort | Li, Yi-Fan |
collection | PubMed |
description | BACKGROUND: Convolutional Neural Network (DCNN), with its great performance, has attracted attention of researchers from many disciplines. The studies of the DCNN and that of biological neural systems have inspired each other reciprocally. The brain-inspired neural networks not only achieve great performance but also serve as a computational model of biological neural systems. METHODS: Here in this study, we trained and tested several typical DCNNs (AlexNet, VGG11, VGG13, VGG16, DenseNet, MobileNet, and EfficientNet) with a face ethnicity categorization task for experiment 1, and an emotion categorization task for experiment 2. We measured the performance of DCNNs by testing them with original and lossy visual inputs (various kinds of image occlusion) and compared their performance with human participants. Moreover, the class activation map (CAM) method allowed us to visualize the foci of the “attention” of these DCNNs. RESULTS: The results suggested that the VGG13 performed the best: Its performance closely resembled human participants in terms of psychophysics measurements, it utilized similar areas of visual inputs as humans, and it had the most consistent performance with inputs having various kinds of impairments. DISCUSSION: In general, we examined the processing mechanism of DCNNs using a new paradigm and found that VGG13 might be the most human-like DCNN in this task. This study also highlighted a possible paradigm to study and develop DCNNs using human perception as a benchmark. |
format | Online Article Text |
id | pubmed-9744930 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Frontiers Media S.A. |
record_format | MEDLINE/PubMed |
spelling | pubmed-97449302022-12-14 Disrupted visual input unveils the computational details of artificial neural networks for face perception Li, Yi-Fan Ying, Haojiang Front Comput Neurosci Neuroscience BACKGROUND: Convolutional Neural Network (DCNN), with its great performance, has attracted attention of researchers from many disciplines. The studies of the DCNN and that of biological neural systems have inspired each other reciprocally. The brain-inspired neural networks not only achieve great performance but also serve as a computational model of biological neural systems. METHODS: Here in this study, we trained and tested several typical DCNNs (AlexNet, VGG11, VGG13, VGG16, DenseNet, MobileNet, and EfficientNet) with a face ethnicity categorization task for experiment 1, and an emotion categorization task for experiment 2. We measured the performance of DCNNs by testing them with original and lossy visual inputs (various kinds of image occlusion) and compared their performance with human participants. Moreover, the class activation map (CAM) method allowed us to visualize the foci of the “attention” of these DCNNs. RESULTS: The results suggested that the VGG13 performed the best: Its performance closely resembled human participants in terms of psychophysics measurements, it utilized similar areas of visual inputs as humans, and it had the most consistent performance with inputs having various kinds of impairments. DISCUSSION: In general, we examined the processing mechanism of DCNNs using a new paradigm and found that VGG13 might be the most human-like DCNN in this task. This study also highlighted a possible paradigm to study and develop DCNNs using human perception as a benchmark. Frontiers Media S.A. 2022-11-29 /pmc/articles/PMC9744930/ /pubmed/36523327 http://dx.doi.org/10.3389/fncom.2022.1054421 Text en Copyright © 2022 Li and Ying. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. |
spellingShingle | Neuroscience Li, Yi-Fan Ying, Haojiang Disrupted visual input unveils the computational details of artificial neural networks for face perception |
title | Disrupted visual input unveils the computational details of artificial neural networks for face perception |
title_full | Disrupted visual input unveils the computational details of artificial neural networks for face perception |
title_fullStr | Disrupted visual input unveils the computational details of artificial neural networks for face perception |
title_full_unstemmed | Disrupted visual input unveils the computational details of artificial neural networks for face perception |
title_short | Disrupted visual input unveils the computational details of artificial neural networks for face perception |
title_sort | disrupted visual input unveils the computational details of artificial neural networks for face perception |
topic | Neuroscience |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9744930/ https://www.ncbi.nlm.nih.gov/pubmed/36523327 http://dx.doi.org/10.3389/fncom.2022.1054421 |
work_keys_str_mv | AT liyifan disruptedvisualinputunveilsthecomputationaldetailsofartificialneuralnetworksforfaceperception AT yinghaojiang disruptedvisualinputunveilsthecomputationaldetailsofartificialneuralnetworksforfaceperception |