Cargando…

Multimodal Feature Fusion Method for Unbalanced Sample Data in Social Network Public Opinion

With the wide application of social media, public opinion analysis in social networks has been unable to be met through text alone because the existing public opinion information includes data information of various modalities, such as voice, text, and facial expressions. Therefore multi-modal emoti...

Descripción completa

Detalles Bibliográficos
Autores principales: Zhao, Jian, Dong, Wenhua, Shi, Lijuan, Qiang, Wenqian, Kuang, Zhejun, Xu, Dawei, An, Tianbo
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9331324/
https://www.ncbi.nlm.nih.gov/pubmed/35898032
http://dx.doi.org/10.3390/s22155528
_version_ 1784758373953044480
author Zhao, Jian
Dong, Wenhua
Shi, Lijuan
Qiang, Wenqian
Kuang, Zhejun
Xu, Dawei
An, Tianbo
author_facet Zhao, Jian
Dong, Wenhua
Shi, Lijuan
Qiang, Wenqian
Kuang, Zhejun
Xu, Dawei
An, Tianbo
author_sort Zhao, Jian
collection PubMed
description With the wide application of social media, public opinion analysis in social networks has been unable to be met through text alone because the existing public opinion information includes data information of various modalities, such as voice, text, and facial expressions. Therefore multi-modal emotion analysis is the current focus of public opinion analysis. In addition, multi-modal emotion recognition of speech is an important factor restricting the multi-modal emotion analysis. In this paper, the emotion feature retrieval method for speech is firstly explored and the processing method of sample disequilibrium data is then analyzed. By comparing and studying the different feature fusion methods of text and speech, respectively, the multi-modal feature fusion method for sample disequilibrium data is proposed to realize multi-modal emotion recognition. Experiments are performed using two publicly available datasets (IEMOCAP and MELD), which shows that processing multi-modality data through this method can obtain good fine-grained emotion recognition results, laying a foundation for subsequent social public opinion analysis.
format Online
Article
Text
id pubmed-9331324
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-93313242022-07-29 Multimodal Feature Fusion Method for Unbalanced Sample Data in Social Network Public Opinion Zhao, Jian Dong, Wenhua Shi, Lijuan Qiang, Wenqian Kuang, Zhejun Xu, Dawei An, Tianbo Sensors (Basel) Article With the wide application of social media, public opinion analysis in social networks has been unable to be met through text alone because the existing public opinion information includes data information of various modalities, such as voice, text, and facial expressions. Therefore multi-modal emotion analysis is the current focus of public opinion analysis. In addition, multi-modal emotion recognition of speech is an important factor restricting the multi-modal emotion analysis. In this paper, the emotion feature retrieval method for speech is firstly explored and the processing method of sample disequilibrium data is then analyzed. By comparing and studying the different feature fusion methods of text and speech, respectively, the multi-modal feature fusion method for sample disequilibrium data is proposed to realize multi-modal emotion recognition. Experiments are performed using two publicly available datasets (IEMOCAP and MELD), which shows that processing multi-modality data through this method can obtain good fine-grained emotion recognition results, laying a foundation for subsequent social public opinion analysis. MDPI 2022-07-25 /pmc/articles/PMC9331324/ /pubmed/35898032 http://dx.doi.org/10.3390/s22155528 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Zhao, Jian
Dong, Wenhua
Shi, Lijuan
Qiang, Wenqian
Kuang, Zhejun
Xu, Dawei
An, Tianbo
Multimodal Feature Fusion Method for Unbalanced Sample Data in Social Network Public Opinion
title Multimodal Feature Fusion Method for Unbalanced Sample Data in Social Network Public Opinion
title_full Multimodal Feature Fusion Method for Unbalanced Sample Data in Social Network Public Opinion
title_fullStr Multimodal Feature Fusion Method for Unbalanced Sample Data in Social Network Public Opinion
title_full_unstemmed Multimodal Feature Fusion Method for Unbalanced Sample Data in Social Network Public Opinion
title_short Multimodal Feature Fusion Method for Unbalanced Sample Data in Social Network Public Opinion
title_sort multimodal feature fusion method for unbalanced sample data in social network public opinion
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9331324/
https://www.ncbi.nlm.nih.gov/pubmed/35898032
http://dx.doi.org/10.3390/s22155528
work_keys_str_mv AT zhaojian multimodalfeaturefusionmethodforunbalancedsampledatainsocialnetworkpublicopinion
AT dongwenhua multimodalfeaturefusionmethodforunbalancedsampledatainsocialnetworkpublicopinion
AT shilijuan multimodalfeaturefusionmethodforunbalancedsampledatainsocialnetworkpublicopinion
AT qiangwenqian multimodalfeaturefusionmethodforunbalancedsampledatainsocialnetworkpublicopinion
AT kuangzhejun multimodalfeaturefusionmethodforunbalancedsampledatainsocialnetworkpublicopinion
AT xudawei multimodalfeaturefusionmethodforunbalancedsampledatainsocialnetworkpublicopinion
AT antianbo multimodalfeaturefusionmethodforunbalancedsampledatainsocialnetworkpublicopinion