Cargando…
CEG: A joint model for causal commonsense events enhanced story ending generation
With the success of pre-trained language models, the performance of story ending generation has been dramatically improved while remaining challenging due to the lack of commonsense reasoning ability. Most previous works mainly focus on using commonsense knowledge to enhance the implicit correlation...
Autores principales: | , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Public Library of Science
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10204949/ https://www.ncbi.nlm.nih.gov/pubmed/37220112 http://dx.doi.org/10.1371/journal.pone.0286049 |
_version_ | 1785045936389488640 |
---|---|
author | Zhang, Yushi Yang, Yan Gu, Ming Gao, Feng Chen, Chengcai He, Liang |
author_facet | Zhang, Yushi Yang, Yan Gu, Ming Gao, Feng Chen, Chengcai He, Liang |
author_sort | Zhang, Yushi |
collection | PubMed |
description | With the success of pre-trained language models, the performance of story ending generation has been dramatically improved while remaining challenging due to the lack of commonsense reasoning ability. Most previous works mainly focus on using commonsense knowledge to enhance the implicit correlations between words but ignore the hidden causality of sentences or events. In this paper, we propose Causal commonsense Enhanced joint model for story ending Generation (CEG), which incorporates causal commonsense events knowledge to generate a reasonable story ending. Specifically, we first develop a commonsense events inference model trained on GLUCOSE, which converts static knowledge into a dynamic generation model to discover unseen knowledge. It uses prompts to produce various commonsense events behind the stories as pseudo-labels of the dataset. Then, we propose a joint model for the causal events inference task and the story ending generation task to inject inference knowledge into the generation, which consists of a shared encoder, an inference decoder, and a generation decoder. In the causal events inference task, we use the shared encoder and the inference decoder to reason the causal events behind each sentence of the story context to help the model better understand the story and provide long-distance dependencies for the story ending generation. In story ending generation, we combine the hidden states of the causal events with the story context to generate the story ending by the shared encoder and the generation decoder. We jointly train the model on two tasks so that the generation decoder produces the story endings that better match the clues. Experimental results on the ROCStories dataset show that our model outperforms the previous works, demonstrating the effectiveness of the joint model and the generated causal events. |
format | Online Article Text |
id | pubmed-10204949 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | Public Library of Science |
record_format | MEDLINE/PubMed |
spelling | pubmed-102049492023-05-24 CEG: A joint model for causal commonsense events enhanced story ending generation Zhang, Yushi Yang, Yan Gu, Ming Gao, Feng Chen, Chengcai He, Liang PLoS One Research Article With the success of pre-trained language models, the performance of story ending generation has been dramatically improved while remaining challenging due to the lack of commonsense reasoning ability. Most previous works mainly focus on using commonsense knowledge to enhance the implicit correlations between words but ignore the hidden causality of sentences or events. In this paper, we propose Causal commonsense Enhanced joint model for story ending Generation (CEG), which incorporates causal commonsense events knowledge to generate a reasonable story ending. Specifically, we first develop a commonsense events inference model trained on GLUCOSE, which converts static knowledge into a dynamic generation model to discover unseen knowledge. It uses prompts to produce various commonsense events behind the stories as pseudo-labels of the dataset. Then, we propose a joint model for the causal events inference task and the story ending generation task to inject inference knowledge into the generation, which consists of a shared encoder, an inference decoder, and a generation decoder. In the causal events inference task, we use the shared encoder and the inference decoder to reason the causal events behind each sentence of the story context to help the model better understand the story and provide long-distance dependencies for the story ending generation. In story ending generation, we combine the hidden states of the causal events with the story context to generate the story ending by the shared encoder and the generation decoder. We jointly train the model on two tasks so that the generation decoder produces the story endings that better match the clues. Experimental results on the ROCStories dataset show that our model outperforms the previous works, demonstrating the effectiveness of the joint model and the generated causal events. Public Library of Science 2023-05-23 /pmc/articles/PMC10204949/ /pubmed/37220112 http://dx.doi.org/10.1371/journal.pone.0286049 Text en © 2023 Zhang et al https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. |
spellingShingle | Research Article Zhang, Yushi Yang, Yan Gu, Ming Gao, Feng Chen, Chengcai He, Liang CEG: A joint model for causal commonsense events enhanced story ending generation |
title | CEG: A joint model for causal commonsense events enhanced story ending generation |
title_full | CEG: A joint model for causal commonsense events enhanced story ending generation |
title_fullStr | CEG: A joint model for causal commonsense events enhanced story ending generation |
title_full_unstemmed | CEG: A joint model for causal commonsense events enhanced story ending generation |
title_short | CEG: A joint model for causal commonsense events enhanced story ending generation |
title_sort | ceg: a joint model for causal commonsense events enhanced story ending generation |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10204949/ https://www.ncbi.nlm.nih.gov/pubmed/37220112 http://dx.doi.org/10.1371/journal.pone.0286049 |
work_keys_str_mv | AT zhangyushi cegajointmodelforcausalcommonsenseeventsenhancedstoryendinggeneration AT yangyan cegajointmodelforcausalcommonsenseeventsenhancedstoryendinggeneration AT guming cegajointmodelforcausalcommonsenseeventsenhancedstoryendinggeneration AT gaofeng cegajointmodelforcausalcommonsenseeventsenhancedstoryendinggeneration AT chenchengcai cegajointmodelforcausalcommonsenseeventsenhancedstoryendinggeneration AT heliang cegajointmodelforcausalcommonsenseeventsenhancedstoryendinggeneration |