Cargando…
Probabilistic coherence, logical consistency, and Bayesian learning: Neural language models as epistemic agents
It is argued that suitably trained neural language models exhibit key properties of epistemic agency: they hold probabilistically coherent and logically consistent degrees of belief, which they can rationally revise in the face of novel evidence. To this purpose, we conduct computational experiments...
Autores principales: | , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Public Library of Science
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9910757/ https://www.ncbi.nlm.nih.gov/pubmed/36758000 http://dx.doi.org/10.1371/journal.pone.0281372 |
_version_ | 1784884853887467520 |
---|---|
author | Betz, Gregor Richardson, Kyle |
author_facet | Betz, Gregor Richardson, Kyle |
author_sort | Betz, Gregor |
collection | PubMed |
description | It is argued that suitably trained neural language models exhibit key properties of epistemic agency: they hold probabilistically coherent and logically consistent degrees of belief, which they can rationally revise in the face of novel evidence. To this purpose, we conduct computational experiments with rankers: T5 models [Raffel et al. 2020] that are pretrained on carefully designed synthetic corpora. Moreover, we introduce a procedure for eliciting a model’s degrees of belief, and define numerical metrics that measure the extent to which given degrees of belief violate (probabilistic, logical, and Bayesian) rationality constraints. While pretrained rankers are found to suffer from global inconsistency (in agreement with, e.g., [Jang et al. 2021]), we observe that subsequent self-training on auto-generated texts allows rankers to gradually obtain a probabilistically coherent belief system that is aligned with logical constraints. In addition, such self-training is found to have a pivotal role in rational evidential learning, too, for it seems to enable rankers to propagate a novel evidence item through their belief systems, successively re-adjusting individual degrees of belief. All this, we conclude, confirms the Rationality Hypothesis, i.e., the claim that suitable trained NLMs may exhibit advanced rational skills. We suggest that this hypothesis has empirical, yet also normative and conceptual ramifications far beyond the practical linguistic problems NLMs have originally been designed to solve. |
format | Online Article Text |
id | pubmed-9910757 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | Public Library of Science |
record_format | MEDLINE/PubMed |
spelling | pubmed-99107572023-02-10 Probabilistic coherence, logical consistency, and Bayesian learning: Neural language models as epistemic agents Betz, Gregor Richardson, Kyle PLoS One Research Article It is argued that suitably trained neural language models exhibit key properties of epistemic agency: they hold probabilistically coherent and logically consistent degrees of belief, which they can rationally revise in the face of novel evidence. To this purpose, we conduct computational experiments with rankers: T5 models [Raffel et al. 2020] that are pretrained on carefully designed synthetic corpora. Moreover, we introduce a procedure for eliciting a model’s degrees of belief, and define numerical metrics that measure the extent to which given degrees of belief violate (probabilistic, logical, and Bayesian) rationality constraints. While pretrained rankers are found to suffer from global inconsistency (in agreement with, e.g., [Jang et al. 2021]), we observe that subsequent self-training on auto-generated texts allows rankers to gradually obtain a probabilistically coherent belief system that is aligned with logical constraints. In addition, such self-training is found to have a pivotal role in rational evidential learning, too, for it seems to enable rankers to propagate a novel evidence item through their belief systems, successively re-adjusting individual degrees of belief. All this, we conclude, confirms the Rationality Hypothesis, i.e., the claim that suitable trained NLMs may exhibit advanced rational skills. We suggest that this hypothesis has empirical, yet also normative and conceptual ramifications far beyond the practical linguistic problems NLMs have originally been designed to solve. Public Library of Science 2023-02-09 /pmc/articles/PMC9910757/ /pubmed/36758000 http://dx.doi.org/10.1371/journal.pone.0281372 Text en © 2023 Betz, Richardson https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. |
spellingShingle | Research Article Betz, Gregor Richardson, Kyle Probabilistic coherence, logical consistency, and Bayesian learning: Neural language models as epistemic agents |
title | Probabilistic coherence, logical consistency, and Bayesian learning: Neural language models as epistemic agents |
title_full | Probabilistic coherence, logical consistency, and Bayesian learning: Neural language models as epistemic agents |
title_fullStr | Probabilistic coherence, logical consistency, and Bayesian learning: Neural language models as epistemic agents |
title_full_unstemmed | Probabilistic coherence, logical consistency, and Bayesian learning: Neural language models as epistemic agents |
title_short | Probabilistic coherence, logical consistency, and Bayesian learning: Neural language models as epistemic agents |
title_sort | probabilistic coherence, logical consistency, and bayesian learning: neural language models as epistemic agents |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9910757/ https://www.ncbi.nlm.nih.gov/pubmed/36758000 http://dx.doi.org/10.1371/journal.pone.0281372 |
work_keys_str_mv | AT betzgregor probabilisticcoherencelogicalconsistencyandbayesianlearningneurallanguagemodelsasepistemicagents AT richardsonkyle probabilisticcoherencelogicalconsistencyandbayesianlearningneurallanguagemodelsasepistemicagents |