Cargando…

Multimodal Sentiment Analysis Based on Cross-Modal Attention and Gated Cyclic Hierarchical Fusion Networks

Multimodal sentiment analysis has been an active subfield in natural language processing. This makes multimodal sentiment tasks challenging due to the use of different sources for predicting a speaker's sentiment. Previous research has focused on extracting single contextual information within...

Descripción completa

Detalles Bibliográficos
Autores principales: Quan, Zhibang, Sun, Tao, Su, Mengli, Wei, Jishu
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Hindawi 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9381258/
https://www.ncbi.nlm.nih.gov/pubmed/35983132
http://dx.doi.org/10.1155/2022/4767437
_version_ 1784769039608840192
author Quan, Zhibang
Sun, Tao
Su, Mengli
Wei, Jishu
author_facet Quan, Zhibang
Sun, Tao
Su, Mengli
Wei, Jishu
author_sort Quan, Zhibang
collection PubMed
description Multimodal sentiment analysis has been an active subfield in natural language processing. This makes multimodal sentiment tasks challenging due to the use of different sources for predicting a speaker's sentiment. Previous research has focused on extracting single contextual information within a modality and trying different modality fusion stages to improve prediction accuracy. However, a factor that may lead to poor model performance is that this does not consider the variability between modalities. Furthermore, existing fusion methods tend to extract the representational information of individual modalities before fusion. This ignores the critical role of intermodal interaction information for model prediction. This paper proposes a multimodal sentiment analysis method based on cross-modal attention and gated cyclic hierarchical fusion network MGHF. MGHF is based on the idea of distribution matching, which enables modalities to obtain representational information with a synergistic effect on the overall sentiment orientation in the temporal interaction phase. After that, we designed a gated cyclic hierarchical fusion network that takes text-based acoustic representation, text-based visual representation, and text representation as inputs and eliminates redundant information through a gating mechanism to achieve effective multimodal representation interaction fusion. Our extensive experiments on two publicly available and popular multimodal datasets show that MGHF has significant advantages over previous complex and robust baselines.
format Online
Article
Text
id pubmed-9381258
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Hindawi
record_format MEDLINE/PubMed
spelling pubmed-93812582022-08-17 Multimodal Sentiment Analysis Based on Cross-Modal Attention and Gated Cyclic Hierarchical Fusion Networks Quan, Zhibang Sun, Tao Su, Mengli Wei, Jishu Comput Intell Neurosci Research Article Multimodal sentiment analysis has been an active subfield in natural language processing. This makes multimodal sentiment tasks challenging due to the use of different sources for predicting a speaker's sentiment. Previous research has focused on extracting single contextual information within a modality and trying different modality fusion stages to improve prediction accuracy. However, a factor that may lead to poor model performance is that this does not consider the variability between modalities. Furthermore, existing fusion methods tend to extract the representational information of individual modalities before fusion. This ignores the critical role of intermodal interaction information for model prediction. This paper proposes a multimodal sentiment analysis method based on cross-modal attention and gated cyclic hierarchical fusion network MGHF. MGHF is based on the idea of distribution matching, which enables modalities to obtain representational information with a synergistic effect on the overall sentiment orientation in the temporal interaction phase. After that, we designed a gated cyclic hierarchical fusion network that takes text-based acoustic representation, text-based visual representation, and text representation as inputs and eliminates redundant information through a gating mechanism to achieve effective multimodal representation interaction fusion. Our extensive experiments on two publicly available and popular multimodal datasets show that MGHF has significant advantages over previous complex and robust baselines. Hindawi 2022-08-09 /pmc/articles/PMC9381258/ /pubmed/35983132 http://dx.doi.org/10.1155/2022/4767437 Text en Copyright © 2022 Zhibang Quan et al. https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
spellingShingle Research Article
Quan, Zhibang
Sun, Tao
Su, Mengli
Wei, Jishu
Multimodal Sentiment Analysis Based on Cross-Modal Attention and Gated Cyclic Hierarchical Fusion Networks
title Multimodal Sentiment Analysis Based on Cross-Modal Attention and Gated Cyclic Hierarchical Fusion Networks
title_full Multimodal Sentiment Analysis Based on Cross-Modal Attention and Gated Cyclic Hierarchical Fusion Networks
title_fullStr Multimodal Sentiment Analysis Based on Cross-Modal Attention and Gated Cyclic Hierarchical Fusion Networks
title_full_unstemmed Multimodal Sentiment Analysis Based on Cross-Modal Attention and Gated Cyclic Hierarchical Fusion Networks
title_short Multimodal Sentiment Analysis Based on Cross-Modal Attention and Gated Cyclic Hierarchical Fusion Networks
title_sort multimodal sentiment analysis based on cross-modal attention and gated cyclic hierarchical fusion networks
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9381258/
https://www.ncbi.nlm.nih.gov/pubmed/35983132
http://dx.doi.org/10.1155/2022/4767437
work_keys_str_mv AT quanzhibang multimodalsentimentanalysisbasedoncrossmodalattentionandgatedcyclichierarchicalfusionnetworks
AT suntao multimodalsentimentanalysisbasedoncrossmodalattentionandgatedcyclichierarchicalfusionnetworks
AT sumengli multimodalsentimentanalysisbasedoncrossmodalattentionandgatedcyclichierarchicalfusionnetworks
AT weijishu multimodalsentimentanalysisbasedoncrossmodalattentionandgatedcyclichierarchicalfusionnetworks