Cargando…

Inter- and Intra-Modal Contrastive Hybrid Learning Framework for Multimodal Abstractive Summarization

Internet users are benefiting from technologies of abstractive summarization enabling them to view articles on the internet by reading article summaries only instead of an entire article. However, there are disadvantages to technologies for analyzing articles with texts and images due to the semanti...

Descripción completa

Detalles Bibliográficos
Autores principales: Li, Jiangfeng, Zhang, Zijian, Wang, Bowen, Zhao, Qinpei, Zhang, Chenxi
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9222507/
https://www.ncbi.nlm.nih.gov/pubmed/35741485
http://dx.doi.org/10.3390/e24060764
_version_ 1784732882704531456
author Li, Jiangfeng
Zhang, Zijian
Wang, Bowen
Zhao, Qinpei
Zhang, Chenxi
author_facet Li, Jiangfeng
Zhang, Zijian
Wang, Bowen
Zhao, Qinpei
Zhang, Chenxi
author_sort Li, Jiangfeng
collection PubMed
description Internet users are benefiting from technologies of abstractive summarization enabling them to view articles on the internet by reading article summaries only instead of an entire article. However, there are disadvantages to technologies for analyzing articles with texts and images due to the semantic gap between vision and language. These technologies focus more on aggregating features and neglect the heterogeneity of each modality. At the same time, the lack of consideration of intrinsic data properties within each modality and semantic information from cross-modal correlations result in the poor quality of learned representations. Therefore, we propose a novel Inter- and Intra-modal Contrastive Hybrid learning framework which learns to automatically align the multimodal information and maintains the semantic consistency of input/output flows. Moreover, ITCH can be taken as a component to make the model suitable for both supervised and unsupervised learning approaches. Experiments on two public datasets, MMS and MSMO, show that the ITCH performances are better than the current baselines.
format Online
Article
Text
id pubmed-9222507
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-92225072022-06-24 Inter- and Intra-Modal Contrastive Hybrid Learning Framework for Multimodal Abstractive Summarization Li, Jiangfeng Zhang, Zijian Wang, Bowen Zhao, Qinpei Zhang, Chenxi Entropy (Basel) Article Internet users are benefiting from technologies of abstractive summarization enabling them to view articles on the internet by reading article summaries only instead of an entire article. However, there are disadvantages to technologies for analyzing articles with texts and images due to the semantic gap between vision and language. These technologies focus more on aggregating features and neglect the heterogeneity of each modality. At the same time, the lack of consideration of intrinsic data properties within each modality and semantic information from cross-modal correlations result in the poor quality of learned representations. Therefore, we propose a novel Inter- and Intra-modal Contrastive Hybrid learning framework which learns to automatically align the multimodal information and maintains the semantic consistency of input/output flows. Moreover, ITCH can be taken as a component to make the model suitable for both supervised and unsupervised learning approaches. Experiments on two public datasets, MMS and MSMO, show that the ITCH performances are better than the current baselines. MDPI 2022-05-29 /pmc/articles/PMC9222507/ /pubmed/35741485 http://dx.doi.org/10.3390/e24060764 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Li, Jiangfeng
Zhang, Zijian
Wang, Bowen
Zhao, Qinpei
Zhang, Chenxi
Inter- and Intra-Modal Contrastive Hybrid Learning Framework for Multimodal Abstractive Summarization
title Inter- and Intra-Modal Contrastive Hybrid Learning Framework for Multimodal Abstractive Summarization
title_full Inter- and Intra-Modal Contrastive Hybrid Learning Framework for Multimodal Abstractive Summarization
title_fullStr Inter- and Intra-Modal Contrastive Hybrid Learning Framework for Multimodal Abstractive Summarization
title_full_unstemmed Inter- and Intra-Modal Contrastive Hybrid Learning Framework for Multimodal Abstractive Summarization
title_short Inter- and Intra-Modal Contrastive Hybrid Learning Framework for Multimodal Abstractive Summarization
title_sort inter- and intra-modal contrastive hybrid learning framework for multimodal abstractive summarization
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9222507/
https://www.ncbi.nlm.nih.gov/pubmed/35741485
http://dx.doi.org/10.3390/e24060764
work_keys_str_mv AT lijiangfeng interandintramodalcontrastivehybridlearningframeworkformultimodalabstractivesummarization
AT zhangzijian interandintramodalcontrastivehybridlearningframeworkformultimodalabstractivesummarization
AT wangbowen interandintramodalcontrastivehybridlearningframeworkformultimodalabstractivesummarization
AT zhaoqinpei interandintramodalcontrastivehybridlearningframeworkformultimodalabstractivesummarization
AT zhangchenxi interandintramodalcontrastivehybridlearningframeworkformultimodalabstractivesummarization