Cargando…

A Transformer-Based Hierarchical Variational AutoEncoder Combined Hidden Markov Model for Long Text Generation

The Variational AutoEncoder (VAE) has made significant progress in text generation, but it focused on short text (always a sentence). Long texts consist of multiple sentences. There is a particular relationship between each sentence, especially between the latent variables that control the generatio...

Descripción completa

Detalles Bibliográficos
Autores principales: Zhao, Kun, Ding, Hongwei, Ye, Kai, Cui, Xiaohui
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8534582/
https://www.ncbi.nlm.nih.gov/pubmed/34682001
http://dx.doi.org/10.3390/e23101277
_version_ 1784587587542843392
author Zhao, Kun
Ding, Hongwei
Ye, Kai
Cui, Xiaohui
author_facet Zhao, Kun
Ding, Hongwei
Ye, Kai
Cui, Xiaohui
author_sort Zhao, Kun
collection PubMed
description The Variational AutoEncoder (VAE) has made significant progress in text generation, but it focused on short text (always a sentence). Long texts consist of multiple sentences. There is a particular relationship between each sentence, especially between the latent variables that control the generation of the sentences. The relationships between these latent variables help in generating continuous and logically connected long texts. There exist very few studies on the relationships between these latent variables. We proposed a method for combining the Transformer-Based Hierarchical Variational AutoEncoder and Hidden Markov Model (HT-HVAE) to learn multiple hierarchical latent variables and their relationships. This application improves long text generation. We use a hierarchical Transformer encoder to encode the long texts in order to obtain better hierarchical information of the long text. HT-HVAE’s generation network uses HMM to learn the relationship between latent variables. We also proposed a method for calculating the perplexity for the multiple hierarchical latent variable structure. The experimental results show that our model is more effective in the dataset with strong logic, alleviates the notorious posterior collapse problem, and generates more continuous and logically connected long text.
format Online
Article
Text
id pubmed-8534582
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-85345822021-10-23 A Transformer-Based Hierarchical Variational AutoEncoder Combined Hidden Markov Model for Long Text Generation Zhao, Kun Ding, Hongwei Ye, Kai Cui, Xiaohui Entropy (Basel) Article The Variational AutoEncoder (VAE) has made significant progress in text generation, but it focused on short text (always a sentence). Long texts consist of multiple sentences. There is a particular relationship between each sentence, especially between the latent variables that control the generation of the sentences. The relationships between these latent variables help in generating continuous and logically connected long texts. There exist very few studies on the relationships between these latent variables. We proposed a method for combining the Transformer-Based Hierarchical Variational AutoEncoder and Hidden Markov Model (HT-HVAE) to learn multiple hierarchical latent variables and their relationships. This application improves long text generation. We use a hierarchical Transformer encoder to encode the long texts in order to obtain better hierarchical information of the long text. HT-HVAE’s generation network uses HMM to learn the relationship between latent variables. We also proposed a method for calculating the perplexity for the multiple hierarchical latent variable structure. The experimental results show that our model is more effective in the dataset with strong logic, alleviates the notorious posterior collapse problem, and generates more continuous and logically connected long text. MDPI 2021-09-29 /pmc/articles/PMC8534582/ /pubmed/34682001 http://dx.doi.org/10.3390/e23101277 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Zhao, Kun
Ding, Hongwei
Ye, Kai
Cui, Xiaohui
A Transformer-Based Hierarchical Variational AutoEncoder Combined Hidden Markov Model for Long Text Generation
title A Transformer-Based Hierarchical Variational AutoEncoder Combined Hidden Markov Model for Long Text Generation
title_full A Transformer-Based Hierarchical Variational AutoEncoder Combined Hidden Markov Model for Long Text Generation
title_fullStr A Transformer-Based Hierarchical Variational AutoEncoder Combined Hidden Markov Model for Long Text Generation
title_full_unstemmed A Transformer-Based Hierarchical Variational AutoEncoder Combined Hidden Markov Model for Long Text Generation
title_short A Transformer-Based Hierarchical Variational AutoEncoder Combined Hidden Markov Model for Long Text Generation
title_sort transformer-based hierarchical variational autoencoder combined hidden markov model for long text generation
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8534582/
https://www.ncbi.nlm.nih.gov/pubmed/34682001
http://dx.doi.org/10.3390/e23101277
work_keys_str_mv AT zhaokun atransformerbasedhierarchicalvariationalautoencodercombinedhiddenmarkovmodelforlongtextgeneration
AT dinghongwei atransformerbasedhierarchicalvariationalautoencodercombinedhiddenmarkovmodelforlongtextgeneration
AT yekai atransformerbasedhierarchicalvariationalautoencodercombinedhiddenmarkovmodelforlongtextgeneration
AT cuixiaohui atransformerbasedhierarchicalvariationalautoencodercombinedhiddenmarkovmodelforlongtextgeneration
AT zhaokun transformerbasedhierarchicalvariationalautoencodercombinedhiddenmarkovmodelforlongtextgeneration
AT dinghongwei transformerbasedhierarchicalvariationalautoencodercombinedhiddenmarkovmodelforlongtextgeneration
AT yekai transformerbasedhierarchicalvariationalautoencodercombinedhiddenmarkovmodelforlongtextgeneration
AT cuixiaohui transformerbasedhierarchicalvariationalautoencodercombinedhiddenmarkovmodelforlongtextgeneration