Cargando…

Learning action-oriented models through active inference

Converging theories suggest that organisms learn and exploit probabilistic models of their environment. However, it remains unclear how such models can be learned in practice. The open-ended complexity of natural environments means that it is generally infeasible for organisms to model their environ...

Descripción completa

Detalles Bibliográficos
Autores principales: Tschantz, Alexander, Seth, Anil K., Buckley, Christopher L.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Public Library of Science 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7200021/
https://www.ncbi.nlm.nih.gov/pubmed/32324758
http://dx.doi.org/10.1371/journal.pcbi.1007805
_version_ 1783529258313842688
author Tschantz, Alexander
Seth, Anil K.
Buckley, Christopher L.
author_facet Tschantz, Alexander
Seth, Anil K.
Buckley, Christopher L.
author_sort Tschantz, Alexander
collection PubMed
description Converging theories suggest that organisms learn and exploit probabilistic models of their environment. However, it remains unclear how such models can be learned in practice. The open-ended complexity of natural environments means that it is generally infeasible for organisms to model their environment comprehensively. Alternatively, action-oriented models attempt to encode a parsimonious representation of adaptive agent-environment interactions. One approach to learning action-oriented models is to learn online in the presence of goal-directed behaviours. This constrains an agent to behaviourally relevant trajectories, reducing the diversity of the data a model need account for. Unfortunately, this approach can cause models to prematurely converge to sub-optimal solutions, through a process we refer to as a bad-bootstrap. Here, we exploit the normative framework of active inference to show that efficient action-oriented models can be learned by balancing goal-oriented and epistemic (information-seeking) behaviours in a principled manner. We illustrate our approach using a simple agent-based model of bacterial chemotaxis. We first demonstrate that learning via goal-directed behaviour indeed constrains models to behaviorally relevant aspects of the environment, but that this approach is prone to sub-optimal convergence. We then demonstrate that epistemic behaviours facilitate the construction of accurate and comprehensive models, but that these models are not tailored to any specific behavioural niche and are therefore less efficient in their use of data. Finally, we show that active inference agents learn models that are parsimonious, tailored to action, and which avoid bad bootstraps and sub-optimal convergence. Critically, our results indicate that models learned through active inference can support adaptive behaviour in spite of, and indeed because of, their departure from veridical representations of the environment. Our approach provides a principled method for learning adaptive models from limited interactions with an environment, highlighting a route to sample efficient learning algorithms.
format Online
Article
Text
id pubmed-7200021
institution National Center for Biotechnology Information
language English
publishDate 2020
publisher Public Library of Science
record_format MEDLINE/PubMed
spelling pubmed-72000212020-05-12 Learning action-oriented models through active inference Tschantz, Alexander Seth, Anil K. Buckley, Christopher L. PLoS Comput Biol Research Article Converging theories suggest that organisms learn and exploit probabilistic models of their environment. However, it remains unclear how such models can be learned in practice. The open-ended complexity of natural environments means that it is generally infeasible for organisms to model their environment comprehensively. Alternatively, action-oriented models attempt to encode a parsimonious representation of adaptive agent-environment interactions. One approach to learning action-oriented models is to learn online in the presence of goal-directed behaviours. This constrains an agent to behaviourally relevant trajectories, reducing the diversity of the data a model need account for. Unfortunately, this approach can cause models to prematurely converge to sub-optimal solutions, through a process we refer to as a bad-bootstrap. Here, we exploit the normative framework of active inference to show that efficient action-oriented models can be learned by balancing goal-oriented and epistemic (information-seeking) behaviours in a principled manner. We illustrate our approach using a simple agent-based model of bacterial chemotaxis. We first demonstrate that learning via goal-directed behaviour indeed constrains models to behaviorally relevant aspects of the environment, but that this approach is prone to sub-optimal convergence. We then demonstrate that epistemic behaviours facilitate the construction of accurate and comprehensive models, but that these models are not tailored to any specific behavioural niche and are therefore less efficient in their use of data. Finally, we show that active inference agents learn models that are parsimonious, tailored to action, and which avoid bad bootstraps and sub-optimal convergence. Critically, our results indicate that models learned through active inference can support adaptive behaviour in spite of, and indeed because of, their departure from veridical representations of the environment. Our approach provides a principled method for learning adaptive models from limited interactions with an environment, highlighting a route to sample efficient learning algorithms. Public Library of Science 2020-04-23 /pmc/articles/PMC7200021/ /pubmed/32324758 http://dx.doi.org/10.1371/journal.pcbi.1007805 Text en © 2020 Tschantz et al http://creativecommons.org/licenses/by/4.0/ This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
spellingShingle Research Article
Tschantz, Alexander
Seth, Anil K.
Buckley, Christopher L.
Learning action-oriented models through active inference
title Learning action-oriented models through active inference
title_full Learning action-oriented models through active inference
title_fullStr Learning action-oriented models through active inference
title_full_unstemmed Learning action-oriented models through active inference
title_short Learning action-oriented models through active inference
title_sort learning action-oriented models through active inference
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7200021/
https://www.ncbi.nlm.nih.gov/pubmed/32324758
http://dx.doi.org/10.1371/journal.pcbi.1007805
work_keys_str_mv AT tschantzalexander learningactionorientedmodelsthroughactiveinference
AT sethanilk learningactionorientedmodelsthroughactiveinference
AT buckleychristopherl learningactionorientedmodelsthroughactiveinference