Cargando…

Probabilistic coherence, logical consistency, and Bayesian learning: Neural language models as epistemic agents

It is argued that suitably trained neural language models exhibit key properties of epistemic agency: they hold probabilistically coherent and logically consistent degrees of belief, which they can rationally revise in the face of novel evidence. To this purpose, we conduct computational experiments...

Descripción completa

Detalles Bibliográficos
Autores principales: Betz, Gregor, Richardson, Kyle
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Public Library of Science 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9910757/
https://www.ncbi.nlm.nih.gov/pubmed/36758000
http://dx.doi.org/10.1371/journal.pone.0281372
_version_ 1784884853887467520
author Betz, Gregor
Richardson, Kyle
author_facet Betz, Gregor
Richardson, Kyle
author_sort Betz, Gregor
collection PubMed
description It is argued that suitably trained neural language models exhibit key properties of epistemic agency: they hold probabilistically coherent and logically consistent degrees of belief, which they can rationally revise in the face of novel evidence. To this purpose, we conduct computational experiments with rankers: T5 models [Raffel et al. 2020] that are pretrained on carefully designed synthetic corpora. Moreover, we introduce a procedure for eliciting a model’s degrees of belief, and define numerical metrics that measure the extent to which given degrees of belief violate (probabilistic, logical, and Bayesian) rationality constraints. While pretrained rankers are found to suffer from global inconsistency (in agreement with, e.g., [Jang et al. 2021]), we observe that subsequent self-training on auto-generated texts allows rankers to gradually obtain a probabilistically coherent belief system that is aligned with logical constraints. In addition, such self-training is found to have a pivotal role in rational evidential learning, too, for it seems to enable rankers to propagate a novel evidence item through their belief systems, successively re-adjusting individual degrees of belief. All this, we conclude, confirms the Rationality Hypothesis, i.e., the claim that suitable trained NLMs may exhibit advanced rational skills. We suggest that this hypothesis has empirical, yet also normative and conceptual ramifications far beyond the practical linguistic problems NLMs have originally been designed to solve.
format Online
Article
Text
id pubmed-9910757
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Public Library of Science
record_format MEDLINE/PubMed
spelling pubmed-99107572023-02-10 Probabilistic coherence, logical consistency, and Bayesian learning: Neural language models as epistemic agents Betz, Gregor Richardson, Kyle PLoS One Research Article It is argued that suitably trained neural language models exhibit key properties of epistemic agency: they hold probabilistically coherent and logically consistent degrees of belief, which they can rationally revise in the face of novel evidence. To this purpose, we conduct computational experiments with rankers: T5 models [Raffel et al. 2020] that are pretrained on carefully designed synthetic corpora. Moreover, we introduce a procedure for eliciting a model’s degrees of belief, and define numerical metrics that measure the extent to which given degrees of belief violate (probabilistic, logical, and Bayesian) rationality constraints. While pretrained rankers are found to suffer from global inconsistency (in agreement with, e.g., [Jang et al. 2021]), we observe that subsequent self-training on auto-generated texts allows rankers to gradually obtain a probabilistically coherent belief system that is aligned with logical constraints. In addition, such self-training is found to have a pivotal role in rational evidential learning, too, for it seems to enable rankers to propagate a novel evidence item through their belief systems, successively re-adjusting individual degrees of belief. All this, we conclude, confirms the Rationality Hypothesis, i.e., the claim that suitable trained NLMs may exhibit advanced rational skills. We suggest that this hypothesis has empirical, yet also normative and conceptual ramifications far beyond the practical linguistic problems NLMs have originally been designed to solve. Public Library of Science 2023-02-09 /pmc/articles/PMC9910757/ /pubmed/36758000 http://dx.doi.org/10.1371/journal.pone.0281372 Text en © 2023 Betz, Richardson https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
spellingShingle Research Article
Betz, Gregor
Richardson, Kyle
Probabilistic coherence, logical consistency, and Bayesian learning: Neural language models as epistemic agents
title Probabilistic coherence, logical consistency, and Bayesian learning: Neural language models as epistemic agents
title_full Probabilistic coherence, logical consistency, and Bayesian learning: Neural language models as epistemic agents
title_fullStr Probabilistic coherence, logical consistency, and Bayesian learning: Neural language models as epistemic agents
title_full_unstemmed Probabilistic coherence, logical consistency, and Bayesian learning: Neural language models as epistemic agents
title_short Probabilistic coherence, logical consistency, and Bayesian learning: Neural language models as epistemic agents
title_sort probabilistic coherence, logical consistency, and bayesian learning: neural language models as epistemic agents
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9910757/
https://www.ncbi.nlm.nih.gov/pubmed/36758000
http://dx.doi.org/10.1371/journal.pone.0281372
work_keys_str_mv AT betzgregor probabilisticcoherencelogicalconsistencyandbayesianlearningneurallanguagemodelsasepistemicagents
AT richardsonkyle probabilisticcoherencelogicalconsistencyandbayesianlearningneurallanguagemodelsasepistemicagents