Cargando…

Dynamic Invariant-Specific Representation Fusion Network for Multimodal Sentiment Analysis

Multimodal sentiment analysis (MSA) aims to infer emotions from linguistic, auditory, and visual sequences. Multimodal information representation method and fusion technology are keys to MSA. However, the problem of difficulty in fully obtaining heterogeneous data interactions in MSA usually exists....

Descripción completa

Detalles Bibliográficos
Autores principales: He, Jing, Yanga, Haonan, Zhang, Changfan, Chen, Hongrun, Xua, Yifu
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Hindawi 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8803469/
https://www.ncbi.nlm.nih.gov/pubmed/35111209
http://dx.doi.org/10.1155/2022/2105593
_version_ 1784642877302767616
author He, Jing
Yanga, Haonan
Zhang, Changfan
Chen, Hongrun
Xua, Yifu
author_facet He, Jing
Yanga, Haonan
Zhang, Changfan
Chen, Hongrun
Xua, Yifu
author_sort He, Jing
collection PubMed
description Multimodal sentiment analysis (MSA) aims to infer emotions from linguistic, auditory, and visual sequences. Multimodal information representation method and fusion technology are keys to MSA. However, the problem of difficulty in fully obtaining heterogeneous data interactions in MSA usually exists. To solve these problems, a new framework, namely, dynamic invariant-specific representation fusion network (DISRFN), is put forward in this study. Firstly, in order to effectively utilize redundant information, the joint domain separation representations of all modes are obtained through the improved joint domain separation network. Then, the hierarchical graph fusion net (HGFN) is used for dynamically fusing each representation to obtain the interaction of multimodal data for guidance in the sentiment analysis. Moreover, comparative experiments are performed on popular MSA data sets MOSI and MOSEI, and the research on fusion strategy, loss function ablation, and similarity loss function analysis experiments is designed. The experimental results verify the effectiveness of the DISRFN framework and loss function.
format Online
Article
Text
id pubmed-8803469
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Hindawi
record_format MEDLINE/PubMed
spelling pubmed-88034692022-02-01 Dynamic Invariant-Specific Representation Fusion Network for Multimodal Sentiment Analysis He, Jing Yanga, Haonan Zhang, Changfan Chen, Hongrun Xua, Yifu Comput Intell Neurosci Research Article Multimodal sentiment analysis (MSA) aims to infer emotions from linguistic, auditory, and visual sequences. Multimodal information representation method and fusion technology are keys to MSA. However, the problem of difficulty in fully obtaining heterogeneous data interactions in MSA usually exists. To solve these problems, a new framework, namely, dynamic invariant-specific representation fusion network (DISRFN), is put forward in this study. Firstly, in order to effectively utilize redundant information, the joint domain separation representations of all modes are obtained through the improved joint domain separation network. Then, the hierarchical graph fusion net (HGFN) is used for dynamically fusing each representation to obtain the interaction of multimodal data for guidance in the sentiment analysis. Moreover, comparative experiments are performed on popular MSA data sets MOSI and MOSEI, and the research on fusion strategy, loss function ablation, and similarity loss function analysis experiments is designed. The experimental results verify the effectiveness of the DISRFN framework and loss function. Hindawi 2022-01-24 /pmc/articles/PMC8803469/ /pubmed/35111209 http://dx.doi.org/10.1155/2022/2105593 Text en Copyright © 2022 Jing He et al. https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
spellingShingle Research Article
He, Jing
Yanga, Haonan
Zhang, Changfan
Chen, Hongrun
Xua, Yifu
Dynamic Invariant-Specific Representation Fusion Network for Multimodal Sentiment Analysis
title Dynamic Invariant-Specific Representation Fusion Network for Multimodal Sentiment Analysis
title_full Dynamic Invariant-Specific Representation Fusion Network for Multimodal Sentiment Analysis
title_fullStr Dynamic Invariant-Specific Representation Fusion Network for Multimodal Sentiment Analysis
title_full_unstemmed Dynamic Invariant-Specific Representation Fusion Network for Multimodal Sentiment Analysis
title_short Dynamic Invariant-Specific Representation Fusion Network for Multimodal Sentiment Analysis
title_sort dynamic invariant-specific representation fusion network for multimodal sentiment analysis
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8803469/
https://www.ncbi.nlm.nih.gov/pubmed/35111209
http://dx.doi.org/10.1155/2022/2105593
work_keys_str_mv AT hejing dynamicinvariantspecificrepresentationfusionnetworkformultimodalsentimentanalysis
AT yangahaonan dynamicinvariantspecificrepresentationfusionnetworkformultimodalsentimentanalysis
AT zhangchangfan dynamicinvariantspecificrepresentationfusionnetworkformultimodalsentimentanalysis
AT chenhongrun dynamicinvariantspecificrepresentationfusionnetworkformultimodalsentimentanalysis
AT xuayifu dynamicinvariantspecificrepresentationfusionnetworkformultimodalsentimentanalysis