Cargando…
Inter- and Intra-Modal Contrastive Hybrid Learning Framework for Multimodal Abstractive Summarization
Internet users are benefiting from technologies of abstractive summarization enabling them to view articles on the internet by reading article summaries only instead of an entire article. However, there are disadvantages to technologies for analyzing articles with texts and images due to the semanti...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9222507/ https://www.ncbi.nlm.nih.gov/pubmed/35741485 http://dx.doi.org/10.3390/e24060764 |
_version_ | 1784732882704531456 |
---|---|
author | Li, Jiangfeng Zhang, Zijian Wang, Bowen Zhao, Qinpei Zhang, Chenxi |
author_facet | Li, Jiangfeng Zhang, Zijian Wang, Bowen Zhao, Qinpei Zhang, Chenxi |
author_sort | Li, Jiangfeng |
collection | PubMed |
description | Internet users are benefiting from technologies of abstractive summarization enabling them to view articles on the internet by reading article summaries only instead of an entire article. However, there are disadvantages to technologies for analyzing articles with texts and images due to the semantic gap between vision and language. These technologies focus more on aggregating features and neglect the heterogeneity of each modality. At the same time, the lack of consideration of intrinsic data properties within each modality and semantic information from cross-modal correlations result in the poor quality of learned representations. Therefore, we propose a novel Inter- and Intra-modal Contrastive Hybrid learning framework which learns to automatically align the multimodal information and maintains the semantic consistency of input/output flows. Moreover, ITCH can be taken as a component to make the model suitable for both supervised and unsupervised learning approaches. Experiments on two public datasets, MMS and MSMO, show that the ITCH performances are better than the current baselines. |
format | Online Article Text |
id | pubmed-9222507 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-92225072022-06-24 Inter- and Intra-Modal Contrastive Hybrid Learning Framework for Multimodal Abstractive Summarization Li, Jiangfeng Zhang, Zijian Wang, Bowen Zhao, Qinpei Zhang, Chenxi Entropy (Basel) Article Internet users are benefiting from technologies of abstractive summarization enabling them to view articles on the internet by reading article summaries only instead of an entire article. However, there are disadvantages to technologies for analyzing articles with texts and images due to the semantic gap between vision and language. These technologies focus more on aggregating features and neglect the heterogeneity of each modality. At the same time, the lack of consideration of intrinsic data properties within each modality and semantic information from cross-modal correlations result in the poor quality of learned representations. Therefore, we propose a novel Inter- and Intra-modal Contrastive Hybrid learning framework which learns to automatically align the multimodal information and maintains the semantic consistency of input/output flows. Moreover, ITCH can be taken as a component to make the model suitable for both supervised and unsupervised learning approaches. Experiments on two public datasets, MMS and MSMO, show that the ITCH performances are better than the current baselines. MDPI 2022-05-29 /pmc/articles/PMC9222507/ /pubmed/35741485 http://dx.doi.org/10.3390/e24060764 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Li, Jiangfeng Zhang, Zijian Wang, Bowen Zhao, Qinpei Zhang, Chenxi Inter- and Intra-Modal Contrastive Hybrid Learning Framework for Multimodal Abstractive Summarization |
title | Inter- and Intra-Modal Contrastive Hybrid Learning Framework for Multimodal Abstractive Summarization |
title_full | Inter- and Intra-Modal Contrastive Hybrid Learning Framework for Multimodal Abstractive Summarization |
title_fullStr | Inter- and Intra-Modal Contrastive Hybrid Learning Framework for Multimodal Abstractive Summarization |
title_full_unstemmed | Inter- and Intra-Modal Contrastive Hybrid Learning Framework for Multimodal Abstractive Summarization |
title_short | Inter- and Intra-Modal Contrastive Hybrid Learning Framework for Multimodal Abstractive Summarization |
title_sort | inter- and intra-modal contrastive hybrid learning framework for multimodal abstractive summarization |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9222507/ https://www.ncbi.nlm.nih.gov/pubmed/35741485 http://dx.doi.org/10.3390/e24060764 |
work_keys_str_mv | AT lijiangfeng interandintramodalcontrastivehybridlearningframeworkformultimodalabstractivesummarization AT zhangzijian interandintramodalcontrastivehybridlearningframeworkformultimodalabstractivesummarization AT wangbowen interandintramodalcontrastivehybridlearningframeworkformultimodalabstractivesummarization AT zhaoqinpei interandintramodalcontrastivehybridlearningframeworkformultimodalabstractivesummarization AT zhangchenxi interandintramodalcontrastivehybridlearningframeworkformultimodalabstractivesummarization |