Cargando…

CEG: A joint model for causal commonsense events enhanced story ending generation

With the success of pre-trained language models, the performance of story ending generation has been dramatically improved while remaining challenging due to the lack of commonsense reasoning ability. Most previous works mainly focus on using commonsense knowledge to enhance the implicit correlation...

Descripción completa

Detalles Bibliográficos
Autores principales: Zhang, Yushi, Yang, Yan, Gu, Ming, Gao, Feng, Chen, Chengcai, He, Liang
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Public Library of Science 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10204949/
https://www.ncbi.nlm.nih.gov/pubmed/37220112
http://dx.doi.org/10.1371/journal.pone.0286049
_version_ 1785045936389488640
author Zhang, Yushi
Yang, Yan
Gu, Ming
Gao, Feng
Chen, Chengcai
He, Liang
author_facet Zhang, Yushi
Yang, Yan
Gu, Ming
Gao, Feng
Chen, Chengcai
He, Liang
author_sort Zhang, Yushi
collection PubMed
description With the success of pre-trained language models, the performance of story ending generation has been dramatically improved while remaining challenging due to the lack of commonsense reasoning ability. Most previous works mainly focus on using commonsense knowledge to enhance the implicit correlations between words but ignore the hidden causality of sentences or events. In this paper, we propose Causal commonsense Enhanced joint model for story ending Generation (CEG), which incorporates causal commonsense events knowledge to generate a reasonable story ending. Specifically, we first develop a commonsense events inference model trained on GLUCOSE, which converts static knowledge into a dynamic generation model to discover unseen knowledge. It uses prompts to produce various commonsense events behind the stories as pseudo-labels of the dataset. Then, we propose a joint model for the causal events inference task and the story ending generation task to inject inference knowledge into the generation, which consists of a shared encoder, an inference decoder, and a generation decoder. In the causal events inference task, we use the shared encoder and the inference decoder to reason the causal events behind each sentence of the story context to help the model better understand the story and provide long-distance dependencies for the story ending generation. In story ending generation, we combine the hidden states of the causal events with the story context to generate the story ending by the shared encoder and the generation decoder. We jointly train the model on two tasks so that the generation decoder produces the story endings that better match the clues. Experimental results on the ROCStories dataset show that our model outperforms the previous works, demonstrating the effectiveness of the joint model and the generated causal events.
format Online
Article
Text
id pubmed-10204949
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Public Library of Science
record_format MEDLINE/PubMed
spelling pubmed-102049492023-05-24 CEG: A joint model for causal commonsense events enhanced story ending generation Zhang, Yushi Yang, Yan Gu, Ming Gao, Feng Chen, Chengcai He, Liang PLoS One Research Article With the success of pre-trained language models, the performance of story ending generation has been dramatically improved while remaining challenging due to the lack of commonsense reasoning ability. Most previous works mainly focus on using commonsense knowledge to enhance the implicit correlations between words but ignore the hidden causality of sentences or events. In this paper, we propose Causal commonsense Enhanced joint model for story ending Generation (CEG), which incorporates causal commonsense events knowledge to generate a reasonable story ending. Specifically, we first develop a commonsense events inference model trained on GLUCOSE, which converts static knowledge into a dynamic generation model to discover unseen knowledge. It uses prompts to produce various commonsense events behind the stories as pseudo-labels of the dataset. Then, we propose a joint model for the causal events inference task and the story ending generation task to inject inference knowledge into the generation, which consists of a shared encoder, an inference decoder, and a generation decoder. In the causal events inference task, we use the shared encoder and the inference decoder to reason the causal events behind each sentence of the story context to help the model better understand the story and provide long-distance dependencies for the story ending generation. In story ending generation, we combine the hidden states of the causal events with the story context to generate the story ending by the shared encoder and the generation decoder. We jointly train the model on two tasks so that the generation decoder produces the story endings that better match the clues. Experimental results on the ROCStories dataset show that our model outperforms the previous works, demonstrating the effectiveness of the joint model and the generated causal events. Public Library of Science 2023-05-23 /pmc/articles/PMC10204949/ /pubmed/37220112 http://dx.doi.org/10.1371/journal.pone.0286049 Text en © 2023 Zhang et al https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
spellingShingle Research Article
Zhang, Yushi
Yang, Yan
Gu, Ming
Gao, Feng
Chen, Chengcai
He, Liang
CEG: A joint model for causal commonsense events enhanced story ending generation
title CEG: A joint model for causal commonsense events enhanced story ending generation
title_full CEG: A joint model for causal commonsense events enhanced story ending generation
title_fullStr CEG: A joint model for causal commonsense events enhanced story ending generation
title_full_unstemmed CEG: A joint model for causal commonsense events enhanced story ending generation
title_short CEG: A joint model for causal commonsense events enhanced story ending generation
title_sort ceg: a joint model for causal commonsense events enhanced story ending generation
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10204949/
https://www.ncbi.nlm.nih.gov/pubmed/37220112
http://dx.doi.org/10.1371/journal.pone.0286049
work_keys_str_mv AT zhangyushi cegajointmodelforcausalcommonsenseeventsenhancedstoryendinggeneration
AT yangyan cegajointmodelforcausalcommonsenseeventsenhancedstoryendinggeneration
AT guming cegajointmodelforcausalcommonsenseeventsenhancedstoryendinggeneration
AT gaofeng cegajointmodelforcausalcommonsenseeventsenhancedstoryendinggeneration
AT chenchengcai cegajointmodelforcausalcommonsenseeventsenhancedstoryendinggeneration
AT heliang cegajointmodelforcausalcommonsenseeventsenhancedstoryendinggeneration