Cargando…
Backward-Forward Sequence Generative Network for Multiple Lexical Constraints
Advancements in Long Short Term Memory (LSTM) Networks have shown remarkable success in various Natural Language Generation (NLG) tasks. However, generating sequence from pre-specified lexical constraints is a new, challenging and less researched area in NLG. Lexical constraints take the form of wor...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
2020
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7256622/ http://dx.doi.org/10.1007/978-3-030-49186-4_4 |
_version_ | 1783539952196255744 |
---|---|
author | Latif, Seemab Bashir, Sarmad Agha, Mir Muntasar Ali Latif, Rabia |
author_facet | Latif, Seemab Bashir, Sarmad Agha, Mir Muntasar Ali Latif, Rabia |
author_sort | Latif, Seemab |
collection | PubMed |
description | Advancements in Long Short Term Memory (LSTM) Networks have shown remarkable success in various Natural Language Generation (NLG) tasks. However, generating sequence from pre-specified lexical constraints is a new, challenging and less researched area in NLG. Lexical constraints take the form of words in the language model’s output to create fluent and meaningful sequences. Furthermore, most of the previous approaches cater this problem by allowing the inclusion of pre-specified lexical constraints during the decoding process, which increases the decoding complexity exponentially or linearly with the number of constraints. Moreover, some of the previous approaches can only deal with single constraint. Additionally, most of the previous approaches only deal with single constraints. In this paper, we propose a novel neural probabilistic architecture based on backward-forward language model and word embedding substitution method that can cater multiple lexical constraints for generating quality sequences. Experiments shows that our proposed architecture outperforms previous methods in terms of intrinsic evaluation. |
format | Online Article Text |
id | pubmed-7256622 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2020 |
record_format | MEDLINE/PubMed |
spelling | pubmed-72566222020-05-29 Backward-Forward Sequence Generative Network for Multiple Lexical Constraints Latif, Seemab Bashir, Sarmad Agha, Mir Muntasar Ali Latif, Rabia Artificial Intelligence Applications and Innovations Article Advancements in Long Short Term Memory (LSTM) Networks have shown remarkable success in various Natural Language Generation (NLG) tasks. However, generating sequence from pre-specified lexical constraints is a new, challenging and less researched area in NLG. Lexical constraints take the form of words in the language model’s output to create fluent and meaningful sequences. Furthermore, most of the previous approaches cater this problem by allowing the inclusion of pre-specified lexical constraints during the decoding process, which increases the decoding complexity exponentially or linearly with the number of constraints. Moreover, some of the previous approaches can only deal with single constraint. Additionally, most of the previous approaches only deal with single constraints. In this paper, we propose a novel neural probabilistic architecture based on backward-forward language model and word embedding substitution method that can cater multiple lexical constraints for generating quality sequences. Experiments shows that our proposed architecture outperforms previous methods in terms of intrinsic evaluation. 2020-05-06 /pmc/articles/PMC7256622/ http://dx.doi.org/10.1007/978-3-030-49186-4_4 Text en © IFIP International Federation for Information Processing 2020 This article is made available via the PMC Open Access Subset for unrestricted research re-use and secondary analysis in any form or by any means with acknowledgement of the original source. These permissions are granted for the duration of the World Health Organization (WHO) declaration of COVID-19 as a global pandemic. |
spellingShingle | Article Latif, Seemab Bashir, Sarmad Agha, Mir Muntasar Ali Latif, Rabia Backward-Forward Sequence Generative Network for Multiple Lexical Constraints |
title | Backward-Forward Sequence Generative Network for Multiple Lexical Constraints |
title_full | Backward-Forward Sequence Generative Network for Multiple Lexical Constraints |
title_fullStr | Backward-Forward Sequence Generative Network for Multiple Lexical Constraints |
title_full_unstemmed | Backward-Forward Sequence Generative Network for Multiple Lexical Constraints |
title_short | Backward-Forward Sequence Generative Network for Multiple Lexical Constraints |
title_sort | backward-forward sequence generative network for multiple lexical constraints |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7256622/ http://dx.doi.org/10.1007/978-3-030-49186-4_4 |
work_keys_str_mv | AT latifseemab backwardforwardsequencegenerativenetworkformultiplelexicalconstraints AT bashirsarmad backwardforwardsequencegenerativenetworkformultiplelexicalconstraints AT aghamirmuntasarali backwardforwardsequencegenerativenetworkformultiplelexicalconstraints AT latifrabia backwardforwardsequencegenerativenetworkformultiplelexicalconstraints |