Cargando…

Judgment aggregation, discursive dilemma and reflective equilibrium: Neural language models as self-improving doxastic agents

Neural language models (NLMs) are susceptible to producing inconsistent output. This paper proposes a new diagnosis as well as a novel remedy for NLMs' incoherence. We train NLMs on synthetic text corpora that are created by simulating text production in a society. For diagnostic purposes, we e...

Descripción completa

Detalles Bibliográficos
Autores principales: Betz, Gregor, Richardson, Kyle
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9623417/
https://www.ncbi.nlm.nih.gov/pubmed/36329681
http://dx.doi.org/10.3389/frai.2022.900943
_version_ 1784821993060696064
author Betz, Gregor
Richardson, Kyle
author_facet Betz, Gregor
Richardson, Kyle
author_sort Betz, Gregor
collection PubMed
description Neural language models (NLMs) are susceptible to producing inconsistent output. This paper proposes a new diagnosis as well as a novel remedy for NLMs' incoherence. We train NLMs on synthetic text corpora that are created by simulating text production in a society. For diagnostic purposes, we explicitly model the individual belief systems of artificial agents (authors) who produce corpus texts. NLMs, trained on those texts, can be shown to aggregate the judgments of individual authors during pre-training according to sentence-wise vote ratios (roughly, reporting frequencies), which inevitably leads to so-called discursive dilemmas: aggregate judgments are inconsistent even though all individual belief states are consistent. As a remedy for such inconsistencies, we develop a self-training procedure—inspired by the concept of reflective equilibrium—that effectively reduces the extent of logical incoherence in a model's belief system, corrects global mis-confidence, and eventually allows the model to settle on a new, epistemically superior belief state. Thus, social choice theory helps to understand why NLMs are prone to produce inconsistencies; epistemology suggests how to get rid of them.
format Online
Article
Text
id pubmed-9623417
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-96234172022-11-02 Judgment aggregation, discursive dilemma and reflective equilibrium: Neural language models as self-improving doxastic agents Betz, Gregor Richardson, Kyle Front Artif Intell Artificial Intelligence Neural language models (NLMs) are susceptible to producing inconsistent output. This paper proposes a new diagnosis as well as a novel remedy for NLMs' incoherence. We train NLMs on synthetic text corpora that are created by simulating text production in a society. For diagnostic purposes, we explicitly model the individual belief systems of artificial agents (authors) who produce corpus texts. NLMs, trained on those texts, can be shown to aggregate the judgments of individual authors during pre-training according to sentence-wise vote ratios (roughly, reporting frequencies), which inevitably leads to so-called discursive dilemmas: aggregate judgments are inconsistent even though all individual belief states are consistent. As a remedy for such inconsistencies, we develop a self-training procedure—inspired by the concept of reflective equilibrium—that effectively reduces the extent of logical incoherence in a model's belief system, corrects global mis-confidence, and eventually allows the model to settle on a new, epistemically superior belief state. Thus, social choice theory helps to understand why NLMs are prone to produce inconsistencies; epistemology suggests how to get rid of them. Frontiers Media S.A. 2022-10-18 /pmc/articles/PMC9623417/ /pubmed/36329681 http://dx.doi.org/10.3389/frai.2022.900943 Text en Copyright © 2022 Betz and Richardson. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Artificial Intelligence
Betz, Gregor
Richardson, Kyle
Judgment aggregation, discursive dilemma and reflective equilibrium: Neural language models as self-improving doxastic agents
title Judgment aggregation, discursive dilemma and reflective equilibrium: Neural language models as self-improving doxastic agents
title_full Judgment aggregation, discursive dilemma and reflective equilibrium: Neural language models as self-improving doxastic agents
title_fullStr Judgment aggregation, discursive dilemma and reflective equilibrium: Neural language models as self-improving doxastic agents
title_full_unstemmed Judgment aggregation, discursive dilemma and reflective equilibrium: Neural language models as self-improving doxastic agents
title_short Judgment aggregation, discursive dilemma and reflective equilibrium: Neural language models as self-improving doxastic agents
title_sort judgment aggregation, discursive dilemma and reflective equilibrium: neural language models as self-improving doxastic agents
topic Artificial Intelligence
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9623417/
https://www.ncbi.nlm.nih.gov/pubmed/36329681
http://dx.doi.org/10.3389/frai.2022.900943
work_keys_str_mv AT betzgregor judgmentaggregationdiscursivedilemmaandreflectiveequilibriumneurallanguagemodelsasselfimprovingdoxasticagents
AT richardsonkyle judgmentaggregationdiscursivedilemmaandreflectiveequilibriumneurallanguagemodelsasselfimprovingdoxasticagents