Cargando…

Multi-Modality Emotion Recognition Model with GAT-Based Multi-Head Inter-Modality Attention

Emotion recognition has been gaining attention in recent years due to its applications on artificial agents. To achieve a good performance with this task, much research has been conducted on the multi-modality emotion recognition model for leveraging the different strengths of each modality. However...

Descripción completa

Detalles Bibliográficos
Autores principales: Fu, Changzeng, Liu, Chaoran, Ishi, Carlos Toshinori, Ishiguro, Hiroshi
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7506856/
https://www.ncbi.nlm.nih.gov/pubmed/32872511
http://dx.doi.org/10.3390/s20174894
_version_ 1783585108815511552
author Fu, Changzeng
Liu, Chaoran
Ishi, Carlos Toshinori
Ishiguro, Hiroshi
author_facet Fu, Changzeng
Liu, Chaoran
Ishi, Carlos Toshinori
Ishiguro, Hiroshi
author_sort Fu, Changzeng
collection PubMed
description Emotion recognition has been gaining attention in recent years due to its applications on artificial agents. To achieve a good performance with this task, much research has been conducted on the multi-modality emotion recognition model for leveraging the different strengths of each modality. However, a research question remains: what exactly is the most appropriate way to fuse the information from different modalities? In this paper, we proposed audio sample augmentation and an emotion-oriented encoder-decoder to improve the performance of emotion recognition and discussed an inter-modality, decision-level fusion method based on a graph attention network (GAT). Compared to the baseline, our model improved the weighted average F1-scores from 64.18 to 68.31% and the weighted average accuracy from 65.25 to 69.88%.
format Online
Article
Text
id pubmed-7506856
institution National Center for Biotechnology Information
language English
publishDate 2020
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-75068562020-09-26 Multi-Modality Emotion Recognition Model with GAT-Based Multi-Head Inter-Modality Attention Fu, Changzeng Liu, Chaoran Ishi, Carlos Toshinori Ishiguro, Hiroshi Sensors (Basel) Article Emotion recognition has been gaining attention in recent years due to its applications on artificial agents. To achieve a good performance with this task, much research has been conducted on the multi-modality emotion recognition model for leveraging the different strengths of each modality. However, a research question remains: what exactly is the most appropriate way to fuse the information from different modalities? In this paper, we proposed audio sample augmentation and an emotion-oriented encoder-decoder to improve the performance of emotion recognition and discussed an inter-modality, decision-level fusion method based on a graph attention network (GAT). Compared to the baseline, our model improved the weighted average F1-scores from 64.18 to 68.31% and the weighted average accuracy from 65.25 to 69.88%. MDPI 2020-08-29 /pmc/articles/PMC7506856/ /pubmed/32872511 http://dx.doi.org/10.3390/s20174894 Text en © 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Fu, Changzeng
Liu, Chaoran
Ishi, Carlos Toshinori
Ishiguro, Hiroshi
Multi-Modality Emotion Recognition Model with GAT-Based Multi-Head Inter-Modality Attention
title Multi-Modality Emotion Recognition Model with GAT-Based Multi-Head Inter-Modality Attention
title_full Multi-Modality Emotion Recognition Model with GAT-Based Multi-Head Inter-Modality Attention
title_fullStr Multi-Modality Emotion Recognition Model with GAT-Based Multi-Head Inter-Modality Attention
title_full_unstemmed Multi-Modality Emotion Recognition Model with GAT-Based Multi-Head Inter-Modality Attention
title_short Multi-Modality Emotion Recognition Model with GAT-Based Multi-Head Inter-Modality Attention
title_sort multi-modality emotion recognition model with gat-based multi-head inter-modality attention
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7506856/
https://www.ncbi.nlm.nih.gov/pubmed/32872511
http://dx.doi.org/10.3390/s20174894
work_keys_str_mv AT fuchangzeng multimodalityemotionrecognitionmodelwithgatbasedmultiheadintermodalityattention
AT liuchaoran multimodalityemotionrecognitionmodelwithgatbasedmultiheadintermodalityattention
AT ishicarlostoshinori multimodalityemotionrecognitionmodelwithgatbasedmultiheadintermodalityattention
AT ishigurohiroshi multimodalityemotionrecognitionmodelwithgatbasedmultiheadintermodalityattention