Cargando…

Facial Expression Recognition Robust to Occlusion and to Intra-Similarity Problem Using Relevant Subsampling

This paper proposes facial expression recognition (FER) with the wild data set. In particular, this paper chiefly deals with two issues, occlusion and intra-similarity problems. The attention mechanism enables one to use the most relevant areas of facial images for specific expressions, and the trip...

Descripción completa

Detalles Bibliográficos
Autores principales: Kim, Jieun, Lee, Deokwoo
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10007059/
https://www.ncbi.nlm.nih.gov/pubmed/36904823
http://dx.doi.org/10.3390/s23052619
_version_ 1784905424873455616
author Kim, Jieun
Lee, Deokwoo
author_facet Kim, Jieun
Lee, Deokwoo
author_sort Kim, Jieun
collection PubMed
description This paper proposes facial expression recognition (FER) with the wild data set. In particular, this paper chiefly deals with two issues, occlusion and intra-similarity problems. The attention mechanism enables one to use the most relevant areas of facial images for specific expressions, and the triplet loss function solves the intra-similarity problem that sometimes fails to aggregate the same expression from different faces and vice versa. The proposed approach for the FER is robust to occlusion, and it uses a spatial transformer network (STN) with an attention mechanism to utilize specific facial region that dominantly contributes (or that is the most relevant) to particular facial expressions, e.g., anger, contempt, disgust, fear, joy, sadness, and surprise. In addition, the STN model is connected to the triplet loss function to improve the recognition rate which outperforms the existing approaches that employ cross-entropy or other approaches using only deep neural networks or classical methods. The triplet loss module alleviates limitations of the intra-similarity problem, leading to further improvement of the classification. Experimental results are provided to substantiate the proposed approach for FER, and the result outperforms the recognition rate in more practical cases, e.g., occlusion. The quantitative result provides FER results with more than [Formula: see text] higher accuracy compared to the existing FER results in CK+ data sets and [Formula: see text] higher than the accuracy of the results with the modified ResNet model in the FER2013 data set.
format Online
Article
Text
id pubmed-10007059
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-100070592023-03-12 Facial Expression Recognition Robust to Occlusion and to Intra-Similarity Problem Using Relevant Subsampling Kim, Jieun Lee, Deokwoo Sensors (Basel) Article This paper proposes facial expression recognition (FER) with the wild data set. In particular, this paper chiefly deals with two issues, occlusion and intra-similarity problems. The attention mechanism enables one to use the most relevant areas of facial images for specific expressions, and the triplet loss function solves the intra-similarity problem that sometimes fails to aggregate the same expression from different faces and vice versa. The proposed approach for the FER is robust to occlusion, and it uses a spatial transformer network (STN) with an attention mechanism to utilize specific facial region that dominantly contributes (or that is the most relevant) to particular facial expressions, e.g., anger, contempt, disgust, fear, joy, sadness, and surprise. In addition, the STN model is connected to the triplet loss function to improve the recognition rate which outperforms the existing approaches that employ cross-entropy or other approaches using only deep neural networks or classical methods. The triplet loss module alleviates limitations of the intra-similarity problem, leading to further improvement of the classification. Experimental results are provided to substantiate the proposed approach for FER, and the result outperforms the recognition rate in more practical cases, e.g., occlusion. The quantitative result provides FER results with more than [Formula: see text] higher accuracy compared to the existing FER results in CK+ data sets and [Formula: see text] higher than the accuracy of the results with the modified ResNet model in the FER2013 data set. MDPI 2023-02-27 /pmc/articles/PMC10007059/ /pubmed/36904823 http://dx.doi.org/10.3390/s23052619 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Kim, Jieun
Lee, Deokwoo
Facial Expression Recognition Robust to Occlusion and to Intra-Similarity Problem Using Relevant Subsampling
title Facial Expression Recognition Robust to Occlusion and to Intra-Similarity Problem Using Relevant Subsampling
title_full Facial Expression Recognition Robust to Occlusion and to Intra-Similarity Problem Using Relevant Subsampling
title_fullStr Facial Expression Recognition Robust to Occlusion and to Intra-Similarity Problem Using Relevant Subsampling
title_full_unstemmed Facial Expression Recognition Robust to Occlusion and to Intra-Similarity Problem Using Relevant Subsampling
title_short Facial Expression Recognition Robust to Occlusion and to Intra-Similarity Problem Using Relevant Subsampling
title_sort facial expression recognition robust to occlusion and to intra-similarity problem using relevant subsampling
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10007059/
https://www.ncbi.nlm.nih.gov/pubmed/36904823
http://dx.doi.org/10.3390/s23052619
work_keys_str_mv AT kimjieun facialexpressionrecognitionrobusttoocclusionandtointrasimilarityproblemusingrelevantsubsampling
AT leedeokwoo facialexpressionrecognitionrobusttoocclusionandtointrasimilarityproblemusingrelevantsubsampling