Cargando…
Multimodal Feature Fusion Method for Unbalanced Sample Data in Social Network Public Opinion
With the wide application of social media, public opinion analysis in social networks has been unable to be met through text alone because the existing public opinion information includes data information of various modalities, such as voice, text, and facial expressions. Therefore multi-modal emoti...
Autores principales: | , , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9331324/ https://www.ncbi.nlm.nih.gov/pubmed/35898032 http://dx.doi.org/10.3390/s22155528 |
_version_ | 1784758373953044480 |
---|---|
author | Zhao, Jian Dong, Wenhua Shi, Lijuan Qiang, Wenqian Kuang, Zhejun Xu, Dawei An, Tianbo |
author_facet | Zhao, Jian Dong, Wenhua Shi, Lijuan Qiang, Wenqian Kuang, Zhejun Xu, Dawei An, Tianbo |
author_sort | Zhao, Jian |
collection | PubMed |
description | With the wide application of social media, public opinion analysis in social networks has been unable to be met through text alone because the existing public opinion information includes data information of various modalities, such as voice, text, and facial expressions. Therefore multi-modal emotion analysis is the current focus of public opinion analysis. In addition, multi-modal emotion recognition of speech is an important factor restricting the multi-modal emotion analysis. In this paper, the emotion feature retrieval method for speech is firstly explored and the processing method of sample disequilibrium data is then analyzed. By comparing and studying the different feature fusion methods of text and speech, respectively, the multi-modal feature fusion method for sample disequilibrium data is proposed to realize multi-modal emotion recognition. Experiments are performed using two publicly available datasets (IEMOCAP and MELD), which shows that processing multi-modality data through this method can obtain good fine-grained emotion recognition results, laying a foundation for subsequent social public opinion analysis. |
format | Online Article Text |
id | pubmed-9331324 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-93313242022-07-29 Multimodal Feature Fusion Method for Unbalanced Sample Data in Social Network Public Opinion Zhao, Jian Dong, Wenhua Shi, Lijuan Qiang, Wenqian Kuang, Zhejun Xu, Dawei An, Tianbo Sensors (Basel) Article With the wide application of social media, public opinion analysis in social networks has been unable to be met through text alone because the existing public opinion information includes data information of various modalities, such as voice, text, and facial expressions. Therefore multi-modal emotion analysis is the current focus of public opinion analysis. In addition, multi-modal emotion recognition of speech is an important factor restricting the multi-modal emotion analysis. In this paper, the emotion feature retrieval method for speech is firstly explored and the processing method of sample disequilibrium data is then analyzed. By comparing and studying the different feature fusion methods of text and speech, respectively, the multi-modal feature fusion method for sample disequilibrium data is proposed to realize multi-modal emotion recognition. Experiments are performed using two publicly available datasets (IEMOCAP and MELD), which shows that processing multi-modality data through this method can obtain good fine-grained emotion recognition results, laying a foundation for subsequent social public opinion analysis. MDPI 2022-07-25 /pmc/articles/PMC9331324/ /pubmed/35898032 http://dx.doi.org/10.3390/s22155528 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Zhao, Jian Dong, Wenhua Shi, Lijuan Qiang, Wenqian Kuang, Zhejun Xu, Dawei An, Tianbo Multimodal Feature Fusion Method for Unbalanced Sample Data in Social Network Public Opinion |
title | Multimodal Feature Fusion Method for Unbalanced Sample Data in Social Network Public Opinion |
title_full | Multimodal Feature Fusion Method for Unbalanced Sample Data in Social Network Public Opinion |
title_fullStr | Multimodal Feature Fusion Method for Unbalanced Sample Data in Social Network Public Opinion |
title_full_unstemmed | Multimodal Feature Fusion Method for Unbalanced Sample Data in Social Network Public Opinion |
title_short | Multimodal Feature Fusion Method for Unbalanced Sample Data in Social Network Public Opinion |
title_sort | multimodal feature fusion method for unbalanced sample data in social network public opinion |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9331324/ https://www.ncbi.nlm.nih.gov/pubmed/35898032 http://dx.doi.org/10.3390/s22155528 |
work_keys_str_mv | AT zhaojian multimodalfeaturefusionmethodforunbalancedsampledatainsocialnetworkpublicopinion AT dongwenhua multimodalfeaturefusionmethodforunbalancedsampledatainsocialnetworkpublicopinion AT shilijuan multimodalfeaturefusionmethodforunbalancedsampledatainsocialnetworkpublicopinion AT qiangwenqian multimodalfeaturefusionmethodforunbalancedsampledatainsocialnetworkpublicopinion AT kuangzhejun multimodalfeaturefusionmethodforunbalancedsampledatainsocialnetworkpublicopinion AT xudawei multimodalfeaturefusionmethodforunbalancedsampledatainsocialnetworkpublicopinion AT antianbo multimodalfeaturefusionmethodforunbalancedsampledatainsocialnetworkpublicopinion |