Cargando…
Dynamic Invariant-Specific Representation Fusion Network for Multimodal Sentiment Analysis
Multimodal sentiment analysis (MSA) aims to infer emotions from linguistic, auditory, and visual sequences. Multimodal information representation method and fusion technology are keys to MSA. However, the problem of difficulty in fully obtaining heterogeneous data interactions in MSA usually exists....
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Hindawi
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8803469/ https://www.ncbi.nlm.nih.gov/pubmed/35111209 http://dx.doi.org/10.1155/2022/2105593 |
_version_ | 1784642877302767616 |
---|---|
author | He, Jing Yanga, Haonan Zhang, Changfan Chen, Hongrun Xua, Yifu |
author_facet | He, Jing Yanga, Haonan Zhang, Changfan Chen, Hongrun Xua, Yifu |
author_sort | He, Jing |
collection | PubMed |
description | Multimodal sentiment analysis (MSA) aims to infer emotions from linguistic, auditory, and visual sequences. Multimodal information representation method and fusion technology are keys to MSA. However, the problem of difficulty in fully obtaining heterogeneous data interactions in MSA usually exists. To solve these problems, a new framework, namely, dynamic invariant-specific representation fusion network (DISRFN), is put forward in this study. Firstly, in order to effectively utilize redundant information, the joint domain separation representations of all modes are obtained through the improved joint domain separation network. Then, the hierarchical graph fusion net (HGFN) is used for dynamically fusing each representation to obtain the interaction of multimodal data for guidance in the sentiment analysis. Moreover, comparative experiments are performed on popular MSA data sets MOSI and MOSEI, and the research on fusion strategy, loss function ablation, and similarity loss function analysis experiments is designed. The experimental results verify the effectiveness of the DISRFN framework and loss function. |
format | Online Article Text |
id | pubmed-8803469 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Hindawi |
record_format | MEDLINE/PubMed |
spelling | pubmed-88034692022-02-01 Dynamic Invariant-Specific Representation Fusion Network for Multimodal Sentiment Analysis He, Jing Yanga, Haonan Zhang, Changfan Chen, Hongrun Xua, Yifu Comput Intell Neurosci Research Article Multimodal sentiment analysis (MSA) aims to infer emotions from linguistic, auditory, and visual sequences. Multimodal information representation method and fusion technology are keys to MSA. However, the problem of difficulty in fully obtaining heterogeneous data interactions in MSA usually exists. To solve these problems, a new framework, namely, dynamic invariant-specific representation fusion network (DISRFN), is put forward in this study. Firstly, in order to effectively utilize redundant information, the joint domain separation representations of all modes are obtained through the improved joint domain separation network. Then, the hierarchical graph fusion net (HGFN) is used for dynamically fusing each representation to obtain the interaction of multimodal data for guidance in the sentiment analysis. Moreover, comparative experiments are performed on popular MSA data sets MOSI and MOSEI, and the research on fusion strategy, loss function ablation, and similarity loss function analysis experiments is designed. The experimental results verify the effectiveness of the DISRFN framework and loss function. Hindawi 2022-01-24 /pmc/articles/PMC8803469/ /pubmed/35111209 http://dx.doi.org/10.1155/2022/2105593 Text en Copyright © 2022 Jing He et al. https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. |
spellingShingle | Research Article He, Jing Yanga, Haonan Zhang, Changfan Chen, Hongrun Xua, Yifu Dynamic Invariant-Specific Representation Fusion Network for Multimodal Sentiment Analysis |
title | Dynamic Invariant-Specific Representation Fusion Network for Multimodal Sentiment Analysis |
title_full | Dynamic Invariant-Specific Representation Fusion Network for Multimodal Sentiment Analysis |
title_fullStr | Dynamic Invariant-Specific Representation Fusion Network for Multimodal Sentiment Analysis |
title_full_unstemmed | Dynamic Invariant-Specific Representation Fusion Network for Multimodal Sentiment Analysis |
title_short | Dynamic Invariant-Specific Representation Fusion Network for Multimodal Sentiment Analysis |
title_sort | dynamic invariant-specific representation fusion network for multimodal sentiment analysis |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8803469/ https://www.ncbi.nlm.nih.gov/pubmed/35111209 http://dx.doi.org/10.1155/2022/2105593 |
work_keys_str_mv | AT hejing dynamicinvariantspecificrepresentationfusionnetworkformultimodalsentimentanalysis AT yangahaonan dynamicinvariantspecificrepresentationfusionnetworkformultimodalsentimentanalysis AT zhangchangfan dynamicinvariantspecificrepresentationfusionnetworkformultimodalsentimentanalysis AT chenhongrun dynamicinvariantspecificrepresentationfusionnetworkformultimodalsentimentanalysis AT xuayifu dynamicinvariantspecificrepresentationfusionnetworkformultimodalsentimentanalysis |