Cargando…

Fusion-Extraction Network for Multimodal Sentiment Analysis

Multiple modality data bring new challenges for sentiment analysis, as combining varieties of information in an effective manner is a rigorous task. Previous works do not effectively utilize the relationship and influence between texts and images. This paper proposes a fusion-extraction network mode...

Descripción completa

Detalles Bibliográficos
Autores principales: Jiang, Tao, Wang, Jiahai, Liu, Zhiyue, Ling, Yingbiao
Formato: Online Artículo Texto
Lenguaje:English
Publicado: 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7206291/
http://dx.doi.org/10.1007/978-3-030-47436-2_59
_version_ 1783530387378536448
author Jiang, Tao
Wang, Jiahai
Liu, Zhiyue
Ling, Yingbiao
author_facet Jiang, Tao
Wang, Jiahai
Liu, Zhiyue
Ling, Yingbiao
author_sort Jiang, Tao
collection PubMed
description Multiple modality data bring new challenges for sentiment analysis, as combining varieties of information in an effective manner is a rigorous task. Previous works do not effectively utilize the relationship and influence between texts and images. This paper proposes a fusion-extraction network model for multimodal sentiment analysis. First, our model uses an interactive information fusion mechanism to interactively learn the visual-specific textual representations and the textual-specific visual representations. Then, we propose an information extraction mechanism to extract valid information and filter redundant parts for the specific textual and visual representations. The experimental results on two public multimodal sentiment datasets show that our model outperforms existing state-of-the-art methods.
format Online
Article
Text
id pubmed-7206291
institution National Center for Biotechnology Information
language English
publishDate 2020
record_format MEDLINE/PubMed
spelling pubmed-72062912020-05-08 Fusion-Extraction Network for Multimodal Sentiment Analysis Jiang, Tao Wang, Jiahai Liu, Zhiyue Ling, Yingbiao Advances in Knowledge Discovery and Data Mining Article Multiple modality data bring new challenges for sentiment analysis, as combining varieties of information in an effective manner is a rigorous task. Previous works do not effectively utilize the relationship and influence between texts and images. This paper proposes a fusion-extraction network model for multimodal sentiment analysis. First, our model uses an interactive information fusion mechanism to interactively learn the visual-specific textual representations and the textual-specific visual representations. Then, we propose an information extraction mechanism to extract valid information and filter redundant parts for the specific textual and visual representations. The experimental results on two public multimodal sentiment datasets show that our model outperforms existing state-of-the-art methods. 2020-04-17 /pmc/articles/PMC7206291/ http://dx.doi.org/10.1007/978-3-030-47436-2_59 Text en © Springer Nature Switzerland AG 2020 This article is made available via the PMC Open Access Subset for unrestricted research re-use and secondary analysis in any form or by any means with acknowledgement of the original source. These permissions are granted for the duration of the World Health Organization (WHO) declaration of COVID-19 as a global pandemic.
spellingShingle Article
Jiang, Tao
Wang, Jiahai
Liu, Zhiyue
Ling, Yingbiao
Fusion-Extraction Network for Multimodal Sentiment Analysis
title Fusion-Extraction Network for Multimodal Sentiment Analysis
title_full Fusion-Extraction Network for Multimodal Sentiment Analysis
title_fullStr Fusion-Extraction Network for Multimodal Sentiment Analysis
title_full_unstemmed Fusion-Extraction Network for Multimodal Sentiment Analysis
title_short Fusion-Extraction Network for Multimodal Sentiment Analysis
title_sort fusion-extraction network for multimodal sentiment analysis
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7206291/
http://dx.doi.org/10.1007/978-3-030-47436-2_59
work_keys_str_mv AT jiangtao fusionextractionnetworkformultimodalsentimentanalysis
AT wangjiahai fusionextractionnetworkformultimodalsentimentanalysis
AT liuzhiyue fusionextractionnetworkformultimodalsentimentanalysis
AT lingyingbiao fusionextractionnetworkformultimodalsentimentanalysis