Cargando…
Goal-Directed Planning for Habituated Agents by Active Inference Using a Variational Recurrent Neural Network
It is crucial to ask how agents can achieve goals by generating action plans using only partial models of the world acquired through habituated sensory-motor experiences. Although many existing robotics studies use a forward model framework, there are generalization issues with high degrees of freed...
Autores principales: | , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2020
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7517093/ https://www.ncbi.nlm.nih.gov/pubmed/33286336 http://dx.doi.org/10.3390/e22050564 |
_version_ | 1783587150754742272 |
---|---|
author | Matsumoto, Takazumi Tani, Jun |
author_facet | Matsumoto, Takazumi Tani, Jun |
author_sort | Matsumoto, Takazumi |
collection | PubMed |
description | It is crucial to ask how agents can achieve goals by generating action plans using only partial models of the world acquired through habituated sensory-motor experiences. Although many existing robotics studies use a forward model framework, there are generalization issues with high degrees of freedom. The current study shows that the predictive coding (PC) and active inference (AIF) frameworks, which employ a generative model, can develop better generalization by learning a prior distribution in a low dimensional latent state space representing probabilistic structures extracted from well habituated sensory-motor trajectories. In our proposed model, learning is carried out by inferring optimal latent variables as well as synaptic weights for maximizing the evidence lower bound, while goal-directed planning is accomplished by inferring latent variables for maximizing the estimated lower bound. Our proposed model was evaluated with both simple and complex robotic tasks in simulation, which demonstrated sufficient generalization in learning with limited training data by setting an intermediate value for a regularization coefficient. Furthermore, comparative simulation results show that the proposed model outperforms a conventional forward model in goal-directed planning, due to the learned prior confining the search of motor plans within the range of habituated trajectories. |
format | Online Article Text |
id | pubmed-7517093 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2020 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-75170932020-11-09 Goal-Directed Planning for Habituated Agents by Active Inference Using a Variational Recurrent Neural Network Matsumoto, Takazumi Tani, Jun Entropy (Basel) Article It is crucial to ask how agents can achieve goals by generating action plans using only partial models of the world acquired through habituated sensory-motor experiences. Although many existing robotics studies use a forward model framework, there are generalization issues with high degrees of freedom. The current study shows that the predictive coding (PC) and active inference (AIF) frameworks, which employ a generative model, can develop better generalization by learning a prior distribution in a low dimensional latent state space representing probabilistic structures extracted from well habituated sensory-motor trajectories. In our proposed model, learning is carried out by inferring optimal latent variables as well as synaptic weights for maximizing the evidence lower bound, while goal-directed planning is accomplished by inferring latent variables for maximizing the estimated lower bound. Our proposed model was evaluated with both simple and complex robotic tasks in simulation, which demonstrated sufficient generalization in learning with limited training data by setting an intermediate value for a regularization coefficient. Furthermore, comparative simulation results show that the proposed model outperforms a conventional forward model in goal-directed planning, due to the learned prior confining the search of motor plans within the range of habituated trajectories. MDPI 2020-05-18 /pmc/articles/PMC7517093/ /pubmed/33286336 http://dx.doi.org/10.3390/e22050564 Text en © 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Matsumoto, Takazumi Tani, Jun Goal-Directed Planning for Habituated Agents by Active Inference Using a Variational Recurrent Neural Network |
title | Goal-Directed Planning for Habituated Agents by Active Inference Using a Variational Recurrent Neural Network |
title_full | Goal-Directed Planning for Habituated Agents by Active Inference Using a Variational Recurrent Neural Network |
title_fullStr | Goal-Directed Planning for Habituated Agents by Active Inference Using a Variational Recurrent Neural Network |
title_full_unstemmed | Goal-Directed Planning for Habituated Agents by Active Inference Using a Variational Recurrent Neural Network |
title_short | Goal-Directed Planning for Habituated Agents by Active Inference Using a Variational Recurrent Neural Network |
title_sort | goal-directed planning for habituated agents by active inference using a variational recurrent neural network |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7517093/ https://www.ncbi.nlm.nih.gov/pubmed/33286336 http://dx.doi.org/10.3390/e22050564 |
work_keys_str_mv | AT matsumototakazumi goaldirectedplanningforhabituatedagentsbyactiveinferenceusingavariationalrecurrentneuralnetwork AT tanijun goaldirectedplanningforhabituatedagentsbyactiveinferenceusingavariationalrecurrentneuralnetwork |