Cargando…

Multi-Hop Question Generation Using Hierarchical Encoding-Decoding and Context Switch Mechanism

Neural auto-regressive sequence-to-sequence models have been dominant in text generation tasks, especially the question generation task. However, neural generation models suffer from the global and local semantic semantic drift problems. Hence, we propose the hierarchical encoding–decoding mechanism...

Descripción completa

Detalles Bibliográficos
Autores principales: Ji, Tianbo, Lyu, Chenyang, Cao, Zhichao, Cheng, Peng
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8618393/
https://www.ncbi.nlm.nih.gov/pubmed/34828147
http://dx.doi.org/10.3390/e23111449
_version_ 1784604736684556288
author Ji, Tianbo
Lyu, Chenyang
Cao, Zhichao
Cheng, Peng
author_facet Ji, Tianbo
Lyu, Chenyang
Cao, Zhichao
Cheng, Peng
author_sort Ji, Tianbo
collection PubMed
description Neural auto-regressive sequence-to-sequence models have been dominant in text generation tasks, especially the question generation task. However, neural generation models suffer from the global and local semantic semantic drift problems. Hence, we propose the hierarchical encoding–decoding mechanism that aims at encoding rich structure information of the input passages and reducing the variance in the decoding phase. In the encoder, we hierarchically encode the input passages according to its structure at four granularity-levels: [word, chunk, sentence, document]-level. Second, we progressively select the context vector from the document-level representations to the word-level representations at each decoding time step. At each time-step in the decoding phase, we progressively select the context vector from the document-level representations to word-level. We also propose the context switch mechanism that enables the decoder to use the context vector from the last step when generating the current word at each time-step.It provides a means of improving the stability of the text generation process during the decoding phase when generating a set of consecutive words. Additionally, we inject syntactic parsing knowledge to enrich the word representations. Experimental results show that our proposed model substantially improves the performance and outperforms previous baselines according to both automatic and human evaluation. Besides, we implement a deep and comprehensive analysis of generated questions based on their types.
format Online
Article
Text
id pubmed-8618393
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-86183932021-11-27 Multi-Hop Question Generation Using Hierarchical Encoding-Decoding and Context Switch Mechanism Ji, Tianbo Lyu, Chenyang Cao, Zhichao Cheng, Peng Entropy (Basel) Article Neural auto-regressive sequence-to-sequence models have been dominant in text generation tasks, especially the question generation task. However, neural generation models suffer from the global and local semantic semantic drift problems. Hence, we propose the hierarchical encoding–decoding mechanism that aims at encoding rich structure information of the input passages and reducing the variance in the decoding phase. In the encoder, we hierarchically encode the input passages according to its structure at four granularity-levels: [word, chunk, sentence, document]-level. Second, we progressively select the context vector from the document-level representations to the word-level representations at each decoding time step. At each time-step in the decoding phase, we progressively select the context vector from the document-level representations to word-level. We also propose the context switch mechanism that enables the decoder to use the context vector from the last step when generating the current word at each time-step.It provides a means of improving the stability of the text generation process during the decoding phase when generating a set of consecutive words. Additionally, we inject syntactic parsing knowledge to enrich the word representations. Experimental results show that our proposed model substantially improves the performance and outperforms previous baselines according to both automatic and human evaluation. Besides, we implement a deep and comprehensive analysis of generated questions based on their types. MDPI 2021-10-31 /pmc/articles/PMC8618393/ /pubmed/34828147 http://dx.doi.org/10.3390/e23111449 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Ji, Tianbo
Lyu, Chenyang
Cao, Zhichao
Cheng, Peng
Multi-Hop Question Generation Using Hierarchical Encoding-Decoding and Context Switch Mechanism
title Multi-Hop Question Generation Using Hierarchical Encoding-Decoding and Context Switch Mechanism
title_full Multi-Hop Question Generation Using Hierarchical Encoding-Decoding and Context Switch Mechanism
title_fullStr Multi-Hop Question Generation Using Hierarchical Encoding-Decoding and Context Switch Mechanism
title_full_unstemmed Multi-Hop Question Generation Using Hierarchical Encoding-Decoding and Context Switch Mechanism
title_short Multi-Hop Question Generation Using Hierarchical Encoding-Decoding and Context Switch Mechanism
title_sort multi-hop question generation using hierarchical encoding-decoding and context switch mechanism
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8618393/
https://www.ncbi.nlm.nih.gov/pubmed/34828147
http://dx.doi.org/10.3390/e23111449
work_keys_str_mv AT jitianbo multihopquestiongenerationusinghierarchicalencodingdecodingandcontextswitchmechanism
AT lyuchenyang multihopquestiongenerationusinghierarchicalencodingdecodingandcontextswitchmechanism
AT caozhichao multihopquestiongenerationusinghierarchicalencodingdecodingandcontextswitchmechanism
AT chengpeng multihopquestiongenerationusinghierarchicalencodingdecodingandcontextswitchmechanism