Cargando…

Learning Generative State Space Models for Active Inference

In this paper we investigate the active inference framework as a means to enable autonomous behavior in artificial agents. Active inference is a theoretical framework underpinning the way organisms act and observe in the real world. In active inference, agents act in order to minimize their so calle...

Descripción completa

Detalles Bibliográficos
Autores principales: Çatal, Ozan, Wauthier, Samuel, De Boom, Cedric, Verbelen, Tim, Dhoedt, Bart
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7701292/
https://www.ncbi.nlm.nih.gov/pubmed/33304260
http://dx.doi.org/10.3389/fncom.2020.574372
_version_ 1783616463359180800
author Çatal, Ozan
Wauthier, Samuel
De Boom, Cedric
Verbelen, Tim
Dhoedt, Bart
author_facet Çatal, Ozan
Wauthier, Samuel
De Boom, Cedric
Verbelen, Tim
Dhoedt, Bart
author_sort Çatal, Ozan
collection PubMed
description In this paper we investigate the active inference framework as a means to enable autonomous behavior in artificial agents. Active inference is a theoretical framework underpinning the way organisms act and observe in the real world. In active inference, agents act in order to minimize their so called free energy, or prediction error. Besides being biologically plausible, active inference has been shown to solve hard exploration problems in various simulated environments. However, these simulations typically require handcrafting a generative model for the agent. Therefore we propose to use recent advances in deep artificial neural networks to learn generative state space models from scratch, using only observation-action sequences. This way we are able to scale active inference to new and challenging problem domains, whilst still building on the theoretical backing of the free energy principle. We validate our approach on the mountain car problem to illustrate that our learnt models can indeed trade-off instrumental value and ambiguity. Furthermore, we show that generative models can also be learnt using high-dimensional pixel observations, both in the OpenAI Gym car racing environment and a real-world robotic navigation task. Finally we show that active inference based policies are an order of magnitude more sample efficient than Deep Q Networks on RL tasks.
format Online
Article
Text
id pubmed-7701292
institution National Center for Biotechnology Information
language English
publishDate 2020
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-77012922020-12-09 Learning Generative State Space Models for Active Inference Çatal, Ozan Wauthier, Samuel De Boom, Cedric Verbelen, Tim Dhoedt, Bart Front Comput Neurosci Neuroscience In this paper we investigate the active inference framework as a means to enable autonomous behavior in artificial agents. Active inference is a theoretical framework underpinning the way organisms act and observe in the real world. In active inference, agents act in order to minimize their so called free energy, or prediction error. Besides being biologically plausible, active inference has been shown to solve hard exploration problems in various simulated environments. However, these simulations typically require handcrafting a generative model for the agent. Therefore we propose to use recent advances in deep artificial neural networks to learn generative state space models from scratch, using only observation-action sequences. This way we are able to scale active inference to new and challenging problem domains, whilst still building on the theoretical backing of the free energy principle. We validate our approach on the mountain car problem to illustrate that our learnt models can indeed trade-off instrumental value and ambiguity. Furthermore, we show that generative models can also be learnt using high-dimensional pixel observations, both in the OpenAI Gym car racing environment and a real-world robotic navigation task. Finally we show that active inference based policies are an order of magnitude more sample efficient than Deep Q Networks on RL tasks. Frontiers Media S.A. 2020-11-16 /pmc/articles/PMC7701292/ /pubmed/33304260 http://dx.doi.org/10.3389/fncom.2020.574372 Text en Copyright © 2020 Çatal, Wauthier, De Boom, Verbelen and Dhoedt. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Neuroscience
Çatal, Ozan
Wauthier, Samuel
De Boom, Cedric
Verbelen, Tim
Dhoedt, Bart
Learning Generative State Space Models for Active Inference
title Learning Generative State Space Models for Active Inference
title_full Learning Generative State Space Models for Active Inference
title_fullStr Learning Generative State Space Models for Active Inference
title_full_unstemmed Learning Generative State Space Models for Active Inference
title_short Learning Generative State Space Models for Active Inference
title_sort learning generative state space models for active inference
topic Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7701292/
https://www.ncbi.nlm.nih.gov/pubmed/33304260
http://dx.doi.org/10.3389/fncom.2020.574372
work_keys_str_mv AT catalozan learninggenerativestatespacemodelsforactiveinference
AT wauthiersamuel learninggenerativestatespacemodelsforactiveinference
AT deboomcedric learninggenerativestatespacemodelsforactiveinference
AT verbelentim learninggenerativestatespacemodelsforactiveinference
AT dhoedtbart learninggenerativestatespacemodelsforactiveinference