Cargando…
Distract Your Attention: Multi-Head Cross Attention Network for Facial Expression Recognition
This paper presents a novel facial expression recognition network, called Distract your Attention Network (DAN). Our method is based on two key observations in biological visual perception. Firstly, multiple facial expression classes share inherently similar underlying facial appearance, and their d...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10204414/ https://www.ncbi.nlm.nih.gov/pubmed/37218785 http://dx.doi.org/10.3390/biomimetics8020199 |
_version_ | 1785045828261380096 |
---|---|
author | Wen, Zhengyao Lin, Wenzhong Wang, Tao Xu, Ge |
author_facet | Wen, Zhengyao Lin, Wenzhong Wang, Tao Xu, Ge |
author_sort | Wen, Zhengyao |
collection | PubMed |
description | This paper presents a novel facial expression recognition network, called Distract your Attention Network (DAN). Our method is based on two key observations in biological visual perception. Firstly, multiple facial expression classes share inherently similar underlying facial appearance, and their differences could be subtle. Secondly, facial expressions simultaneously exhibit themselves through multiple facial regions, and for recognition, a holistic approach by encoding high-order interactions among local features is required. To address these issues, this work proposes DAN with three key components: Feature Clustering Network (FCN), Multi-head Attention Network (MAN), and Attention Fusion Network (AFN). Specifically, FCN extracts robust features by adopting a large-margin learning objective to maximize class separability. In addition, MAN instantiates a number of attention heads to simultaneously attend to multiple facial areas and build attention maps on these regions. Further, AFN distracts these attentions to multiple locations before fusing the feature maps to a comprehensive one. Extensive experiments on three public datasets (including AffectNet, RAF-DB, and SFEW 2.0) verified that the proposed method consistently achieves state-of-the-art facial expression recognition performance. The DAN code is publicly available. |
format | Online Article Text |
id | pubmed-10204414 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-102044142023-05-24 Distract Your Attention: Multi-Head Cross Attention Network for Facial Expression Recognition Wen, Zhengyao Lin, Wenzhong Wang, Tao Xu, Ge Biomimetics (Basel) Article This paper presents a novel facial expression recognition network, called Distract your Attention Network (DAN). Our method is based on two key observations in biological visual perception. Firstly, multiple facial expression classes share inherently similar underlying facial appearance, and their differences could be subtle. Secondly, facial expressions simultaneously exhibit themselves through multiple facial regions, and for recognition, a holistic approach by encoding high-order interactions among local features is required. To address these issues, this work proposes DAN with three key components: Feature Clustering Network (FCN), Multi-head Attention Network (MAN), and Attention Fusion Network (AFN). Specifically, FCN extracts robust features by adopting a large-margin learning objective to maximize class separability. In addition, MAN instantiates a number of attention heads to simultaneously attend to multiple facial areas and build attention maps on these regions. Further, AFN distracts these attentions to multiple locations before fusing the feature maps to a comprehensive one. Extensive experiments on three public datasets (including AffectNet, RAF-DB, and SFEW 2.0) verified that the proposed method consistently achieves state-of-the-art facial expression recognition performance. The DAN code is publicly available. MDPI 2023-05-11 /pmc/articles/PMC10204414/ /pubmed/37218785 http://dx.doi.org/10.3390/biomimetics8020199 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Wen, Zhengyao Lin, Wenzhong Wang, Tao Xu, Ge Distract Your Attention: Multi-Head Cross Attention Network for Facial Expression Recognition |
title | Distract Your Attention: Multi-Head Cross Attention Network for Facial Expression Recognition |
title_full | Distract Your Attention: Multi-Head Cross Attention Network for Facial Expression Recognition |
title_fullStr | Distract Your Attention: Multi-Head Cross Attention Network for Facial Expression Recognition |
title_full_unstemmed | Distract Your Attention: Multi-Head Cross Attention Network for Facial Expression Recognition |
title_short | Distract Your Attention: Multi-Head Cross Attention Network for Facial Expression Recognition |
title_sort | distract your attention: multi-head cross attention network for facial expression recognition |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10204414/ https://www.ncbi.nlm.nih.gov/pubmed/37218785 http://dx.doi.org/10.3390/biomimetics8020199 |
work_keys_str_mv | AT wenzhengyao distractyourattentionmultiheadcrossattentionnetworkforfacialexpressionrecognition AT linwenzhong distractyourattentionmultiheadcrossattentionnetworkforfacialexpressionrecognition AT wangtao distractyourattentionmultiheadcrossattentionnetworkforfacialexpressionrecognition AT xuge distractyourattentionmultiheadcrossattentionnetworkforfacialexpressionrecognition |