Cargando…
Multi-Modality Emotion Recognition Model with GAT-Based Multi-Head Inter-Modality Attention
Emotion recognition has been gaining attention in recent years due to its applications on artificial agents. To achieve a good performance with this task, much research has been conducted on the multi-modality emotion recognition model for leveraging the different strengths of each modality. However...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2020
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7506856/ https://www.ncbi.nlm.nih.gov/pubmed/32872511 http://dx.doi.org/10.3390/s20174894 |
_version_ | 1783585108815511552 |
---|---|
author | Fu, Changzeng Liu, Chaoran Ishi, Carlos Toshinori Ishiguro, Hiroshi |
author_facet | Fu, Changzeng Liu, Chaoran Ishi, Carlos Toshinori Ishiguro, Hiroshi |
author_sort | Fu, Changzeng |
collection | PubMed |
description | Emotion recognition has been gaining attention in recent years due to its applications on artificial agents. To achieve a good performance with this task, much research has been conducted on the multi-modality emotion recognition model for leveraging the different strengths of each modality. However, a research question remains: what exactly is the most appropriate way to fuse the information from different modalities? In this paper, we proposed audio sample augmentation and an emotion-oriented encoder-decoder to improve the performance of emotion recognition and discussed an inter-modality, decision-level fusion method based on a graph attention network (GAT). Compared to the baseline, our model improved the weighted average F1-scores from 64.18 to 68.31% and the weighted average accuracy from 65.25 to 69.88%. |
format | Online Article Text |
id | pubmed-7506856 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2020 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-75068562020-09-26 Multi-Modality Emotion Recognition Model with GAT-Based Multi-Head Inter-Modality Attention Fu, Changzeng Liu, Chaoran Ishi, Carlos Toshinori Ishiguro, Hiroshi Sensors (Basel) Article Emotion recognition has been gaining attention in recent years due to its applications on artificial agents. To achieve a good performance with this task, much research has been conducted on the multi-modality emotion recognition model for leveraging the different strengths of each modality. However, a research question remains: what exactly is the most appropriate way to fuse the information from different modalities? In this paper, we proposed audio sample augmentation and an emotion-oriented encoder-decoder to improve the performance of emotion recognition and discussed an inter-modality, decision-level fusion method based on a graph attention network (GAT). Compared to the baseline, our model improved the weighted average F1-scores from 64.18 to 68.31% and the weighted average accuracy from 65.25 to 69.88%. MDPI 2020-08-29 /pmc/articles/PMC7506856/ /pubmed/32872511 http://dx.doi.org/10.3390/s20174894 Text en © 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Fu, Changzeng Liu, Chaoran Ishi, Carlos Toshinori Ishiguro, Hiroshi Multi-Modality Emotion Recognition Model with GAT-Based Multi-Head Inter-Modality Attention |
title | Multi-Modality Emotion Recognition Model with GAT-Based Multi-Head Inter-Modality Attention |
title_full | Multi-Modality Emotion Recognition Model with GAT-Based Multi-Head Inter-Modality Attention |
title_fullStr | Multi-Modality Emotion Recognition Model with GAT-Based Multi-Head Inter-Modality Attention |
title_full_unstemmed | Multi-Modality Emotion Recognition Model with GAT-Based Multi-Head Inter-Modality Attention |
title_short | Multi-Modality Emotion Recognition Model with GAT-Based Multi-Head Inter-Modality Attention |
title_sort | multi-modality emotion recognition model with gat-based multi-head inter-modality attention |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7506856/ https://www.ncbi.nlm.nih.gov/pubmed/32872511 http://dx.doi.org/10.3390/s20174894 |
work_keys_str_mv | AT fuchangzeng multimodalityemotionrecognitionmodelwithgatbasedmultiheadintermodalityattention AT liuchaoran multimodalityemotionrecognitionmodelwithgatbasedmultiheadintermodalityattention AT ishicarlostoshinori multimodalityemotionrecognitionmodelwithgatbasedmultiheadintermodalityattention AT ishigurohiroshi multimodalityemotionrecognitionmodelwithgatbasedmultiheadintermodalityattention |