Cargando…
Multimodal Sentiment Analysis Representations Learning via Contrastive Learning with Condense Attention Fusion
Multimodal sentiment analysis has gained popularity as a research field for its ability to predict users’ emotional tendencies more comprehensively. The data fusion module is a critical component of multimodal sentiment analysis, as it allows for integrating information from multiple modalities. How...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10007095/ https://www.ncbi.nlm.nih.gov/pubmed/36904883 http://dx.doi.org/10.3390/s23052679 |
_version_ | 1784905433765380096 |
---|---|
author | Wang, Huiru Li, Xiuhong Ren, Zenyu Wang, Min Ma, Chunming |
author_facet | Wang, Huiru Li, Xiuhong Ren, Zenyu Wang, Min Ma, Chunming |
author_sort | Wang, Huiru |
collection | PubMed |
description | Multimodal sentiment analysis has gained popularity as a research field for its ability to predict users’ emotional tendencies more comprehensively. The data fusion module is a critical component of multimodal sentiment analysis, as it allows for integrating information from multiple modalities. However, it is challenging to combine modalities and remove redundant information effectively. In our research, we address these challenges by proposing a multimodal sentiment analysis model based on supervised contrastive learning, which leads to more effective data representation and richer multimodal features. Specifically, we introduce the MLFC module, which utilizes a convolutional neural network (CNN) and Transformer to solve the redundancy problem of each modal feature and reduce irrelevant information. Moreover, our model employs supervised contrastive learning to enhance its ability to learn standard sentiment features from data. We evaluate our model on three widely-used datasets, namely MVSA-single, MVSA-multiple, and HFM, demonstrating that our model outperforms the state-of-the-art model. Finally, we conduct ablation experiments to validate the efficacy of our proposed method. |
format | Online Article Text |
id | pubmed-10007095 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-100070952023-03-12 Multimodal Sentiment Analysis Representations Learning via Contrastive Learning with Condense Attention Fusion Wang, Huiru Li, Xiuhong Ren, Zenyu Wang, Min Ma, Chunming Sensors (Basel) Article Multimodal sentiment analysis has gained popularity as a research field for its ability to predict users’ emotional tendencies more comprehensively. The data fusion module is a critical component of multimodal sentiment analysis, as it allows for integrating information from multiple modalities. However, it is challenging to combine modalities and remove redundant information effectively. In our research, we address these challenges by proposing a multimodal sentiment analysis model based on supervised contrastive learning, which leads to more effective data representation and richer multimodal features. Specifically, we introduce the MLFC module, which utilizes a convolutional neural network (CNN) and Transformer to solve the redundancy problem of each modal feature and reduce irrelevant information. Moreover, our model employs supervised contrastive learning to enhance its ability to learn standard sentiment features from data. We evaluate our model on three widely-used datasets, namely MVSA-single, MVSA-multiple, and HFM, demonstrating that our model outperforms the state-of-the-art model. Finally, we conduct ablation experiments to validate the efficacy of our proposed method. MDPI 2023-03-01 /pmc/articles/PMC10007095/ /pubmed/36904883 http://dx.doi.org/10.3390/s23052679 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Wang, Huiru Li, Xiuhong Ren, Zenyu Wang, Min Ma, Chunming Multimodal Sentiment Analysis Representations Learning via Contrastive Learning with Condense Attention Fusion |
title | Multimodal Sentiment Analysis Representations Learning via Contrastive Learning with Condense Attention Fusion |
title_full | Multimodal Sentiment Analysis Representations Learning via Contrastive Learning with Condense Attention Fusion |
title_fullStr | Multimodal Sentiment Analysis Representations Learning via Contrastive Learning with Condense Attention Fusion |
title_full_unstemmed | Multimodal Sentiment Analysis Representations Learning via Contrastive Learning with Condense Attention Fusion |
title_short | Multimodal Sentiment Analysis Representations Learning via Contrastive Learning with Condense Attention Fusion |
title_sort | multimodal sentiment analysis representations learning via contrastive learning with condense attention fusion |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10007095/ https://www.ncbi.nlm.nih.gov/pubmed/36904883 http://dx.doi.org/10.3390/s23052679 |
work_keys_str_mv | AT wanghuiru multimodalsentimentanalysisrepresentationslearningviacontrastivelearningwithcondenseattentionfusion AT lixiuhong multimodalsentimentanalysisrepresentationslearningviacontrastivelearningwithcondenseattentionfusion AT renzenyu multimodalsentimentanalysisrepresentationslearningviacontrastivelearningwithcondenseattentionfusion AT wangmin multimodalsentimentanalysisrepresentationslearningviacontrastivelearningwithcondenseattentionfusion AT machunming multimodalsentimentanalysisrepresentationslearningviacontrastivelearningwithcondenseattentionfusion |