Cargando…
Facial expression is retained in deep networks trained for face identification
Facial expressions distort visual cues for identification in two-dimensional images. Face processing systems in the brain must decouple image-based information from multiple sources to operate in the social world. Deep convolutional neural networks (DCNN) trained for face identification retain ident...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
The Association for Research in Vision and Ophthalmology
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8039571/ https://www.ncbi.nlm.nih.gov/pubmed/33821927 http://dx.doi.org/10.1167/jov.21.4.4 |
_version_ | 1783677621193670656 |
---|---|
author | Colón, Y. Ivette Castillo, Carlos D. O’Toole, Alice J. |
author_facet | Colón, Y. Ivette Castillo, Carlos D. O’Toole, Alice J. |
author_sort | Colón, Y. Ivette |
collection | PubMed |
description | Facial expressions distort visual cues for identification in two-dimensional images. Face processing systems in the brain must decouple image-based information from multiple sources to operate in the social world. Deep convolutional neural networks (DCNN) trained for face identification retain identity-irrelevant, image-based information (e.g., viewpoint). We asked whether a DCNN trained for identity also retains expression information that generalizes over viewpoint change. DCNN representations were generated for a controlled dataset containing images of 70 actors posing 7 facial expressions (happy, sad, angry, surprised, fearful, disgusted, neutral), from 5 viewpoints (frontal, 90° and 45° left and right profiles). Two-dimensional visualizations of the DCNN representations revealed hierarchical groupings by identity, followed by viewpoint, and then by facial expression. Linear discriminant analysis of full-dimensional representations predicted expressions accurately, mean 76.8% correct for happiness, followed by surprise, disgust, anger, neutral, sad, and fearful at 42.0%; chance [Formula: see text] 14.3%. Expression classification was stable across viewpoints. Representational similarity heatmaps indicated that image similarities within identities varied more by viewpoint than by expression. We conclude that an identity-trained, deep network retains shape-deformable information about expression and viewpoint, along with identity, in a unified form—consistent with a recent hypothesis for ventral visual stream processing. |
format | Online Article Text |
id | pubmed-8039571 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | The Association for Research in Vision and Ophthalmology |
record_format | MEDLINE/PubMed |
spelling | pubmed-80395712021-04-20 Facial expression is retained in deep networks trained for face identification Colón, Y. Ivette Castillo, Carlos D. O’Toole, Alice J. J Vis Article Facial expressions distort visual cues for identification in two-dimensional images. Face processing systems in the brain must decouple image-based information from multiple sources to operate in the social world. Deep convolutional neural networks (DCNN) trained for face identification retain identity-irrelevant, image-based information (e.g., viewpoint). We asked whether a DCNN trained for identity also retains expression information that generalizes over viewpoint change. DCNN representations were generated for a controlled dataset containing images of 70 actors posing 7 facial expressions (happy, sad, angry, surprised, fearful, disgusted, neutral), from 5 viewpoints (frontal, 90° and 45° left and right profiles). Two-dimensional visualizations of the DCNN representations revealed hierarchical groupings by identity, followed by viewpoint, and then by facial expression. Linear discriminant analysis of full-dimensional representations predicted expressions accurately, mean 76.8% correct for happiness, followed by surprise, disgust, anger, neutral, sad, and fearful at 42.0%; chance [Formula: see text] 14.3%. Expression classification was stable across viewpoints. Representational similarity heatmaps indicated that image similarities within identities varied more by viewpoint than by expression. We conclude that an identity-trained, deep network retains shape-deformable information about expression and viewpoint, along with identity, in a unified form—consistent with a recent hypothesis for ventral visual stream processing. The Association for Research in Vision and Ophthalmology 2021-04-06 /pmc/articles/PMC8039571/ /pubmed/33821927 http://dx.doi.org/10.1167/jov.21.4.4 Text en Copyright 2021, The Authors https://creativecommons.org/licenses/by-nc-nd/4.0/This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License. |
spellingShingle | Article Colón, Y. Ivette Castillo, Carlos D. O’Toole, Alice J. Facial expression is retained in deep networks trained for face identification |
title | Facial expression is retained in deep networks trained for face identification |
title_full | Facial expression is retained in deep networks trained for face identification |
title_fullStr | Facial expression is retained in deep networks trained for face identification |
title_full_unstemmed | Facial expression is retained in deep networks trained for face identification |
title_short | Facial expression is retained in deep networks trained for face identification |
title_sort | facial expression is retained in deep networks trained for face identification |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8039571/ https://www.ncbi.nlm.nih.gov/pubmed/33821927 http://dx.doi.org/10.1167/jov.21.4.4 |
work_keys_str_mv | AT colonyivette facialexpressionisretainedindeepnetworkstrainedforfaceidentification AT castillocarlosd facialexpressionisretainedindeepnetworkstrainedforfaceidentification AT otoolealicej facialexpressionisretainedindeepnetworkstrainedforfaceidentification |