Cargando…

Cross Entropy of Neural Language Models at Infinity—A New Bound of the Entropy Rate

Neural language models have drawn a lot of attention for their strong ability to predict natural language text. In this paper, we estimate the entropy rate of natural language with state-of-the-art neural language models. To obtain the estimate, we consider the cross entropy, a measure of the predic...

Descripción completa

Detalles Bibliográficos
Autores principales: Takahashi, Shuntaro, Tanaka-Ishii, Kumiko
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2018
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7512401/
https://www.ncbi.nlm.nih.gov/pubmed/33266563
http://dx.doi.org/10.3390/e20110839
_version_ 1783586149369905152
author Takahashi, Shuntaro
Tanaka-Ishii, Kumiko
author_facet Takahashi, Shuntaro
Tanaka-Ishii, Kumiko
author_sort Takahashi, Shuntaro
collection PubMed
description Neural language models have drawn a lot of attention for their strong ability to predict natural language text. In this paper, we estimate the entropy rate of natural language with state-of-the-art neural language models. To obtain the estimate, we consider the cross entropy, a measure of the prediction accuracy of neural language models, under the theoretically ideal conditions that they are trained with an infinitely large dataset and receive an infinitely long context for prediction. We empirically verify that the effects of the two parameters, the training data size and context length, on the cross entropy consistently obey a power-law decay with a positive constant for two different state-of-the-art neural language models with different language datasets. Based on the verification, we obtained 1.12 bits per character for English by extrapolating the two parameters to infinity. This result suggests that the upper bound of the entropy rate of natural language is potentially smaller than the previously reported values.
format Online
Article
Text
id pubmed-7512401
institution National Center for Biotechnology Information
language English
publishDate 2018
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-75124012020-11-09 Cross Entropy of Neural Language Models at Infinity—A New Bound of the Entropy Rate Takahashi, Shuntaro Tanaka-Ishii, Kumiko Entropy (Basel) Article Neural language models have drawn a lot of attention for their strong ability to predict natural language text. In this paper, we estimate the entropy rate of natural language with state-of-the-art neural language models. To obtain the estimate, we consider the cross entropy, a measure of the prediction accuracy of neural language models, under the theoretically ideal conditions that they are trained with an infinitely large dataset and receive an infinitely long context for prediction. We empirically verify that the effects of the two parameters, the training data size and context length, on the cross entropy consistently obey a power-law decay with a positive constant for two different state-of-the-art neural language models with different language datasets. Based on the verification, we obtained 1.12 bits per character for English by extrapolating the two parameters to infinity. This result suggests that the upper bound of the entropy rate of natural language is potentially smaller than the previously reported values. MDPI 2018-11-02 /pmc/articles/PMC7512401/ /pubmed/33266563 http://dx.doi.org/10.3390/e20110839 Text en © 2018 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Takahashi, Shuntaro
Tanaka-Ishii, Kumiko
Cross Entropy of Neural Language Models at Infinity—A New Bound of the Entropy Rate
title Cross Entropy of Neural Language Models at Infinity—A New Bound of the Entropy Rate
title_full Cross Entropy of Neural Language Models at Infinity—A New Bound of the Entropy Rate
title_fullStr Cross Entropy of Neural Language Models at Infinity—A New Bound of the Entropy Rate
title_full_unstemmed Cross Entropy of Neural Language Models at Infinity—A New Bound of the Entropy Rate
title_short Cross Entropy of Neural Language Models at Infinity—A New Bound of the Entropy Rate
title_sort cross entropy of neural language models at infinity—a new bound of the entropy rate
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7512401/
https://www.ncbi.nlm.nih.gov/pubmed/33266563
http://dx.doi.org/10.3390/e20110839
work_keys_str_mv AT takahashishuntaro crossentropyofneurallanguagemodelsatinfinityanewboundoftheentropyrate
AT tanakaishiikumiko crossentropyofneurallanguagemodelsatinfinityanewboundoftheentropyrate