Cargando…

Facial expression is retained in deep networks trained for face identification

Facial expressions distort visual cues for identification in two-dimensional images. Face processing systems in the brain must decouple image-based information from multiple sources to operate in the social world. Deep convolutional neural networks (DCNN) trained for face identification retain ident...

Descripción completa

Detalles Bibliográficos
Autores principales: Colón, Y. Ivette, Castillo, Carlos D., O’Toole, Alice J.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: The Association for Research in Vision and Ophthalmology 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8039571/
https://www.ncbi.nlm.nih.gov/pubmed/33821927
http://dx.doi.org/10.1167/jov.21.4.4
_version_ 1783677621193670656
author Colón, Y. Ivette
Castillo, Carlos D.
O’Toole, Alice J.
author_facet Colón, Y. Ivette
Castillo, Carlos D.
O’Toole, Alice J.
author_sort Colón, Y. Ivette
collection PubMed
description Facial expressions distort visual cues for identification in two-dimensional images. Face processing systems in the brain must decouple image-based information from multiple sources to operate in the social world. Deep convolutional neural networks (DCNN) trained for face identification retain identity-irrelevant, image-based information (e.g., viewpoint). We asked whether a DCNN trained for identity also retains expression information that generalizes over viewpoint change. DCNN representations were generated for a controlled dataset containing images of 70 actors posing 7 facial expressions (happy, sad, angry, surprised, fearful, disgusted, neutral), from 5 viewpoints (frontal, 90° and 45° left and right profiles). Two-dimensional visualizations of the DCNN representations revealed hierarchical groupings by identity, followed by viewpoint, and then by facial expression. Linear discriminant analysis of full-dimensional representations predicted expressions accurately, mean 76.8% correct for happiness, followed by surprise, disgust, anger, neutral, sad, and fearful at 42.0%; chance [Formula: see text] 14.3%. Expression classification was stable across viewpoints. Representational similarity heatmaps indicated that image similarities within identities varied more by viewpoint than by expression. We conclude that an identity-trained, deep network retains shape-deformable information about expression and viewpoint, along with identity, in a unified form—consistent with a recent hypothesis for ventral visual stream processing.
format Online
Article
Text
id pubmed-8039571
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher The Association for Research in Vision and Ophthalmology
record_format MEDLINE/PubMed
spelling pubmed-80395712021-04-20 Facial expression is retained in deep networks trained for face identification Colón, Y. Ivette Castillo, Carlos D. O’Toole, Alice J. J Vis Article Facial expressions distort visual cues for identification in two-dimensional images. Face processing systems in the brain must decouple image-based information from multiple sources to operate in the social world. Deep convolutional neural networks (DCNN) trained for face identification retain identity-irrelevant, image-based information (e.g., viewpoint). We asked whether a DCNN trained for identity also retains expression information that generalizes over viewpoint change. DCNN representations were generated for a controlled dataset containing images of 70 actors posing 7 facial expressions (happy, sad, angry, surprised, fearful, disgusted, neutral), from 5 viewpoints (frontal, 90° and 45° left and right profiles). Two-dimensional visualizations of the DCNN representations revealed hierarchical groupings by identity, followed by viewpoint, and then by facial expression. Linear discriminant analysis of full-dimensional representations predicted expressions accurately, mean 76.8% correct for happiness, followed by surprise, disgust, anger, neutral, sad, and fearful at 42.0%; chance [Formula: see text] 14.3%. Expression classification was stable across viewpoints. Representational similarity heatmaps indicated that image similarities within identities varied more by viewpoint than by expression. We conclude that an identity-trained, deep network retains shape-deformable information about expression and viewpoint, along with identity, in a unified form—consistent with a recent hypothesis for ventral visual stream processing. The Association for Research in Vision and Ophthalmology 2021-04-06 /pmc/articles/PMC8039571/ /pubmed/33821927 http://dx.doi.org/10.1167/jov.21.4.4 Text en Copyright 2021, The Authors https://creativecommons.org/licenses/by-nc-nd/4.0/This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.
spellingShingle Article
Colón, Y. Ivette
Castillo, Carlos D.
O’Toole, Alice J.
Facial expression is retained in deep networks trained for face identification
title Facial expression is retained in deep networks trained for face identification
title_full Facial expression is retained in deep networks trained for face identification
title_fullStr Facial expression is retained in deep networks trained for face identification
title_full_unstemmed Facial expression is retained in deep networks trained for face identification
title_short Facial expression is retained in deep networks trained for face identification
title_sort facial expression is retained in deep networks trained for face identification
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8039571/
https://www.ncbi.nlm.nih.gov/pubmed/33821927
http://dx.doi.org/10.1167/jov.21.4.4
work_keys_str_mv AT colonyivette facialexpressionisretainedindeepnetworkstrainedforfaceidentification
AT castillocarlosd facialexpressionisretainedindeepnetworkstrainedforfaceidentification
AT otoolealicej facialexpressionisretainedindeepnetworkstrainedforfaceidentification