Cargando…
Region Dual Attention-Based Video Emotion Recognition
To solve the emotional differences between different regions of the video frame and make use of the interrelationship between different regions, a region dual attention-based video emotion recognition method (RDAM) is proposed. RDAM takes as input video frame sequences and learns a discriminatory vi...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Hindawi
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9217593/ https://www.ncbi.nlm.nih.gov/pubmed/35755752 http://dx.doi.org/10.1155/2022/6096325 |
_version_ | 1784731682925969408 |
---|---|
author | Liu, Xiaodong Xu, Huating wang, Miao |
author_facet | Liu, Xiaodong Xu, Huating wang, Miao |
author_sort | Liu, Xiaodong |
collection | PubMed |
description | To solve the emotional differences between different regions of the video frame and make use of the interrelationship between different regions, a region dual attention-based video emotion recognition method (RDAM) is proposed. RDAM takes as input video frame sequences and learns a discriminatory video emotion representation that can make full use of the emotional differences of different regions and the interrelationship between regions. Specifically, we construct two parallel attention modules: one is the regional location attention module, which generates a weight value for each feature region to identify the relative importance of different regions. Based on the weight, the emotion feature that can perceive the emotional sensitive region is generated. The other is the regional relationship attention module, which generates a region relation matrix that represents the interrelationship of different regions of a video frame. Based on the region relation matrix, the emotion feature that can perceive interrelationship between different regions is generated. The outputs of these two attention modules are fused to produce the emotional features of video frames. Then, the features of video frame sequences are fused by attention-based fusion network, and the final emotion feature of the video is produced. The experimental results on the video emotion recognition data sets show that the proposed method outperforms the other related works. |
format | Online Article Text |
id | pubmed-9217593 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Hindawi |
record_format | MEDLINE/PubMed |
spelling | pubmed-92175932022-06-23 Region Dual Attention-Based Video Emotion Recognition Liu, Xiaodong Xu, Huating wang, Miao Comput Intell Neurosci Research Article To solve the emotional differences between different regions of the video frame and make use of the interrelationship between different regions, a region dual attention-based video emotion recognition method (RDAM) is proposed. RDAM takes as input video frame sequences and learns a discriminatory video emotion representation that can make full use of the emotional differences of different regions and the interrelationship between regions. Specifically, we construct two parallel attention modules: one is the regional location attention module, which generates a weight value for each feature region to identify the relative importance of different regions. Based on the weight, the emotion feature that can perceive the emotional sensitive region is generated. The other is the regional relationship attention module, which generates a region relation matrix that represents the interrelationship of different regions of a video frame. Based on the region relation matrix, the emotion feature that can perceive interrelationship between different regions is generated. The outputs of these two attention modules are fused to produce the emotional features of video frames. Then, the features of video frame sequences are fused by attention-based fusion network, and the final emotion feature of the video is produced. The experimental results on the video emotion recognition data sets show that the proposed method outperforms the other related works. Hindawi 2022-06-15 /pmc/articles/PMC9217593/ /pubmed/35755752 http://dx.doi.org/10.1155/2022/6096325 Text en Copyright © 2022 Xiaodong Liu et al. https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. |
spellingShingle | Research Article Liu, Xiaodong Xu, Huating wang, Miao Region Dual Attention-Based Video Emotion Recognition |
title | Region Dual Attention-Based Video Emotion Recognition |
title_full | Region Dual Attention-Based Video Emotion Recognition |
title_fullStr | Region Dual Attention-Based Video Emotion Recognition |
title_full_unstemmed | Region Dual Attention-Based Video Emotion Recognition |
title_short | Region Dual Attention-Based Video Emotion Recognition |
title_sort | region dual attention-based video emotion recognition |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9217593/ https://www.ncbi.nlm.nih.gov/pubmed/35755752 http://dx.doi.org/10.1155/2022/6096325 |
work_keys_str_mv | AT liuxiaodong regiondualattentionbasedvideoemotionrecognition AT xuhuating regiondualattentionbasedvideoemotionrecognition AT wangmiao regiondualattentionbasedvideoemotionrecognition |