Cargando…
Multimodal Sentiment Analysis Based on Cross-Modal Attention and Gated Cyclic Hierarchical Fusion Networks
Multimodal sentiment analysis has been an active subfield in natural language processing. This makes multimodal sentiment tasks challenging due to the use of different sources for predicting a speaker's sentiment. Previous research has focused on extracting single contextual information within...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Hindawi
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9381258/ https://www.ncbi.nlm.nih.gov/pubmed/35983132 http://dx.doi.org/10.1155/2022/4767437 |
_version_ | 1784769039608840192 |
---|---|
author | Quan, Zhibang Sun, Tao Su, Mengli Wei, Jishu |
author_facet | Quan, Zhibang Sun, Tao Su, Mengli Wei, Jishu |
author_sort | Quan, Zhibang |
collection | PubMed |
description | Multimodal sentiment analysis has been an active subfield in natural language processing. This makes multimodal sentiment tasks challenging due to the use of different sources for predicting a speaker's sentiment. Previous research has focused on extracting single contextual information within a modality and trying different modality fusion stages to improve prediction accuracy. However, a factor that may lead to poor model performance is that this does not consider the variability between modalities. Furthermore, existing fusion methods tend to extract the representational information of individual modalities before fusion. This ignores the critical role of intermodal interaction information for model prediction. This paper proposes a multimodal sentiment analysis method based on cross-modal attention and gated cyclic hierarchical fusion network MGHF. MGHF is based on the idea of distribution matching, which enables modalities to obtain representational information with a synergistic effect on the overall sentiment orientation in the temporal interaction phase. After that, we designed a gated cyclic hierarchical fusion network that takes text-based acoustic representation, text-based visual representation, and text representation as inputs and eliminates redundant information through a gating mechanism to achieve effective multimodal representation interaction fusion. Our extensive experiments on two publicly available and popular multimodal datasets show that MGHF has significant advantages over previous complex and robust baselines. |
format | Online Article Text |
id | pubmed-9381258 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Hindawi |
record_format | MEDLINE/PubMed |
spelling | pubmed-93812582022-08-17 Multimodal Sentiment Analysis Based on Cross-Modal Attention and Gated Cyclic Hierarchical Fusion Networks Quan, Zhibang Sun, Tao Su, Mengli Wei, Jishu Comput Intell Neurosci Research Article Multimodal sentiment analysis has been an active subfield in natural language processing. This makes multimodal sentiment tasks challenging due to the use of different sources for predicting a speaker's sentiment. Previous research has focused on extracting single contextual information within a modality and trying different modality fusion stages to improve prediction accuracy. However, a factor that may lead to poor model performance is that this does not consider the variability between modalities. Furthermore, existing fusion methods tend to extract the representational information of individual modalities before fusion. This ignores the critical role of intermodal interaction information for model prediction. This paper proposes a multimodal sentiment analysis method based on cross-modal attention and gated cyclic hierarchical fusion network MGHF. MGHF is based on the idea of distribution matching, which enables modalities to obtain representational information with a synergistic effect on the overall sentiment orientation in the temporal interaction phase. After that, we designed a gated cyclic hierarchical fusion network that takes text-based acoustic representation, text-based visual representation, and text representation as inputs and eliminates redundant information through a gating mechanism to achieve effective multimodal representation interaction fusion. Our extensive experiments on two publicly available and popular multimodal datasets show that MGHF has significant advantages over previous complex and robust baselines. Hindawi 2022-08-09 /pmc/articles/PMC9381258/ /pubmed/35983132 http://dx.doi.org/10.1155/2022/4767437 Text en Copyright © 2022 Zhibang Quan et al. https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. |
spellingShingle | Research Article Quan, Zhibang Sun, Tao Su, Mengli Wei, Jishu Multimodal Sentiment Analysis Based on Cross-Modal Attention and Gated Cyclic Hierarchical Fusion Networks |
title | Multimodal Sentiment Analysis Based on Cross-Modal Attention and Gated Cyclic Hierarchical Fusion Networks |
title_full | Multimodal Sentiment Analysis Based on Cross-Modal Attention and Gated Cyclic Hierarchical Fusion Networks |
title_fullStr | Multimodal Sentiment Analysis Based on Cross-Modal Attention and Gated Cyclic Hierarchical Fusion Networks |
title_full_unstemmed | Multimodal Sentiment Analysis Based on Cross-Modal Attention and Gated Cyclic Hierarchical Fusion Networks |
title_short | Multimodal Sentiment Analysis Based on Cross-Modal Attention and Gated Cyclic Hierarchical Fusion Networks |
title_sort | multimodal sentiment analysis based on cross-modal attention and gated cyclic hierarchical fusion networks |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9381258/ https://www.ncbi.nlm.nih.gov/pubmed/35983132 http://dx.doi.org/10.1155/2022/4767437 |
work_keys_str_mv | AT quanzhibang multimodalsentimentanalysisbasedoncrossmodalattentionandgatedcyclichierarchicalfusionnetworks AT suntao multimodalsentimentanalysisbasedoncrossmodalattentionandgatedcyclichierarchicalfusionnetworks AT sumengli multimodalsentimentanalysisbasedoncrossmodalattentionandgatedcyclichierarchicalfusionnetworks AT weijishu multimodalsentimentanalysisbasedoncrossmodalattentionandgatedcyclichierarchicalfusionnetworks |