Cargando…

Region Dual Attention-Based Video Emotion Recognition

To solve the emotional differences between different regions of the video frame and make use of the interrelationship between different regions, a region dual attention-based video emotion recognition method (RDAM) is proposed. RDAM takes as input video frame sequences and learns a discriminatory vi...

Descripción completa

Detalles Bibliográficos
Autores principales: Liu, Xiaodong, Xu, Huating, wang, Miao
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Hindawi 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9217593/
https://www.ncbi.nlm.nih.gov/pubmed/35755752
http://dx.doi.org/10.1155/2022/6096325
_version_ 1784731682925969408
author Liu, Xiaodong
Xu, Huating
wang, Miao
author_facet Liu, Xiaodong
Xu, Huating
wang, Miao
author_sort Liu, Xiaodong
collection PubMed
description To solve the emotional differences between different regions of the video frame and make use of the interrelationship between different regions, a region dual attention-based video emotion recognition method (RDAM) is proposed. RDAM takes as input video frame sequences and learns a discriminatory video emotion representation that can make full use of the emotional differences of different regions and the interrelationship between regions. Specifically, we construct two parallel attention modules: one is the regional location attention module, which generates a weight value for each feature region to identify the relative importance of different regions. Based on the weight, the emotion feature that can perceive the emotional sensitive region is generated. The other is the regional relationship attention module, which generates a region relation matrix that represents the interrelationship of different regions of a video frame. Based on the region relation matrix, the emotion feature that can perceive interrelationship between different regions is generated. The outputs of these two attention modules are fused to produce the emotional features of video frames. Then, the features of video frame sequences are fused by attention-based fusion network, and the final emotion feature of the video is produced. The experimental results on the video emotion recognition data sets show that the proposed method outperforms the other related works.
format Online
Article
Text
id pubmed-9217593
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Hindawi
record_format MEDLINE/PubMed
spelling pubmed-92175932022-06-23 Region Dual Attention-Based Video Emotion Recognition Liu, Xiaodong Xu, Huating wang, Miao Comput Intell Neurosci Research Article To solve the emotional differences between different regions of the video frame and make use of the interrelationship between different regions, a region dual attention-based video emotion recognition method (RDAM) is proposed. RDAM takes as input video frame sequences and learns a discriminatory video emotion representation that can make full use of the emotional differences of different regions and the interrelationship between regions. Specifically, we construct two parallel attention modules: one is the regional location attention module, which generates a weight value for each feature region to identify the relative importance of different regions. Based on the weight, the emotion feature that can perceive the emotional sensitive region is generated. The other is the regional relationship attention module, which generates a region relation matrix that represents the interrelationship of different regions of a video frame. Based on the region relation matrix, the emotion feature that can perceive interrelationship between different regions is generated. The outputs of these two attention modules are fused to produce the emotional features of video frames. Then, the features of video frame sequences are fused by attention-based fusion network, and the final emotion feature of the video is produced. The experimental results on the video emotion recognition data sets show that the proposed method outperforms the other related works. Hindawi 2022-06-15 /pmc/articles/PMC9217593/ /pubmed/35755752 http://dx.doi.org/10.1155/2022/6096325 Text en Copyright © 2022 Xiaodong Liu et al. https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
spellingShingle Research Article
Liu, Xiaodong
Xu, Huating
wang, Miao
Region Dual Attention-Based Video Emotion Recognition
title Region Dual Attention-Based Video Emotion Recognition
title_full Region Dual Attention-Based Video Emotion Recognition
title_fullStr Region Dual Attention-Based Video Emotion Recognition
title_full_unstemmed Region Dual Attention-Based Video Emotion Recognition
title_short Region Dual Attention-Based Video Emotion Recognition
title_sort region dual attention-based video emotion recognition
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9217593/
https://www.ncbi.nlm.nih.gov/pubmed/35755752
http://dx.doi.org/10.1155/2022/6096325
work_keys_str_mv AT liuxiaodong regiondualattentionbasedvideoemotionrecognition
AT xuhuating regiondualattentionbasedvideoemotionrecognition
AT wangmiao regiondualattentionbasedvideoemotionrecognition