Cargando…

Multimodal Sentiment Analysis Representations Learning via Contrastive Learning with Condense Attention Fusion

Multimodal sentiment analysis has gained popularity as a research field for its ability to predict users’ emotional tendencies more comprehensively. The data fusion module is a critical component of multimodal sentiment analysis, as it allows for integrating information from multiple modalities. How...

Descripción completa

Detalles Bibliográficos
Autores principales: Wang, Huiru, Li, Xiuhong, Ren, Zenyu, Wang, Min, Ma, Chunming
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10007095/
https://www.ncbi.nlm.nih.gov/pubmed/36904883
http://dx.doi.org/10.3390/s23052679
_version_ 1784905433765380096
author Wang, Huiru
Li, Xiuhong
Ren, Zenyu
Wang, Min
Ma, Chunming
author_facet Wang, Huiru
Li, Xiuhong
Ren, Zenyu
Wang, Min
Ma, Chunming
author_sort Wang, Huiru
collection PubMed
description Multimodal sentiment analysis has gained popularity as a research field for its ability to predict users’ emotional tendencies more comprehensively. The data fusion module is a critical component of multimodal sentiment analysis, as it allows for integrating information from multiple modalities. However, it is challenging to combine modalities and remove redundant information effectively. In our research, we address these challenges by proposing a multimodal sentiment analysis model based on supervised contrastive learning, which leads to more effective data representation and richer multimodal features. Specifically, we introduce the MLFC module, which utilizes a convolutional neural network (CNN) and Transformer to solve the redundancy problem of each modal feature and reduce irrelevant information. Moreover, our model employs supervised contrastive learning to enhance its ability to learn standard sentiment features from data. We evaluate our model on three widely-used datasets, namely MVSA-single, MVSA-multiple, and HFM, demonstrating that our model outperforms the state-of-the-art model. Finally, we conduct ablation experiments to validate the efficacy of our proposed method.
format Online
Article
Text
id pubmed-10007095
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-100070952023-03-12 Multimodal Sentiment Analysis Representations Learning via Contrastive Learning with Condense Attention Fusion Wang, Huiru Li, Xiuhong Ren, Zenyu Wang, Min Ma, Chunming Sensors (Basel) Article Multimodal sentiment analysis has gained popularity as a research field for its ability to predict users’ emotional tendencies more comprehensively. The data fusion module is a critical component of multimodal sentiment analysis, as it allows for integrating information from multiple modalities. However, it is challenging to combine modalities and remove redundant information effectively. In our research, we address these challenges by proposing a multimodal sentiment analysis model based on supervised contrastive learning, which leads to more effective data representation and richer multimodal features. Specifically, we introduce the MLFC module, which utilizes a convolutional neural network (CNN) and Transformer to solve the redundancy problem of each modal feature and reduce irrelevant information. Moreover, our model employs supervised contrastive learning to enhance its ability to learn standard sentiment features from data. We evaluate our model on three widely-used datasets, namely MVSA-single, MVSA-multiple, and HFM, demonstrating that our model outperforms the state-of-the-art model. Finally, we conduct ablation experiments to validate the efficacy of our proposed method. MDPI 2023-03-01 /pmc/articles/PMC10007095/ /pubmed/36904883 http://dx.doi.org/10.3390/s23052679 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Wang, Huiru
Li, Xiuhong
Ren, Zenyu
Wang, Min
Ma, Chunming
Multimodal Sentiment Analysis Representations Learning via Contrastive Learning with Condense Attention Fusion
title Multimodal Sentiment Analysis Representations Learning via Contrastive Learning with Condense Attention Fusion
title_full Multimodal Sentiment Analysis Representations Learning via Contrastive Learning with Condense Attention Fusion
title_fullStr Multimodal Sentiment Analysis Representations Learning via Contrastive Learning with Condense Attention Fusion
title_full_unstemmed Multimodal Sentiment Analysis Representations Learning via Contrastive Learning with Condense Attention Fusion
title_short Multimodal Sentiment Analysis Representations Learning via Contrastive Learning with Condense Attention Fusion
title_sort multimodal sentiment analysis representations learning via contrastive learning with condense attention fusion
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10007095/
https://www.ncbi.nlm.nih.gov/pubmed/36904883
http://dx.doi.org/10.3390/s23052679
work_keys_str_mv AT wanghuiru multimodalsentimentanalysisrepresentationslearningviacontrastivelearningwithcondenseattentionfusion
AT lixiuhong multimodalsentimentanalysisrepresentationslearningviacontrastivelearningwithcondenseattentionfusion
AT renzenyu multimodalsentimentanalysisrepresentationslearningviacontrastivelearningwithcondenseattentionfusion
AT wangmin multimodalsentimentanalysisrepresentationslearningviacontrastivelearningwithcondenseattentionfusion
AT machunming multimodalsentimentanalysisrepresentationslearningviacontrastivelearningwithcondenseattentionfusion