Cargando…
Scaled free-energy based reinforcement learning for robust and efficient learning in high-dimensional state spaces
Free-energy based reinforcement learning (FERL) was proposed for learning in high-dimensional state- and action spaces, which cannot be handled by standard function approximation methods. In this study, we propose a scaled version of free-energy based reinforcement learning to achieve more robust an...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Frontiers Media S.A.
2013
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3584292/ https://www.ncbi.nlm.nih.gov/pubmed/23450126 http://dx.doi.org/10.3389/fnbot.2013.00003 |
_version_ | 1782261010475778048 |
---|---|
author | Elfwing, Stefan Uchibe, Eiji Doya, Kenji |
author_facet | Elfwing, Stefan Uchibe, Eiji Doya, Kenji |
author_sort | Elfwing, Stefan |
collection | PubMed |
description | Free-energy based reinforcement learning (FERL) was proposed for learning in high-dimensional state- and action spaces, which cannot be handled by standard function approximation methods. In this study, we propose a scaled version of free-energy based reinforcement learning to achieve more robust and more efficient learning performance. The action-value function is approximated by the negative free-energy of a restricted Boltzmann machine, divided by a constant scaling factor that is related to the size of the Boltzmann machine (the square root of the number of state nodes in this study). Our first task is a digit floor gridworld task, where the states are represented by images of handwritten digits from the MNIST data set. The purpose of the task is to investigate the proposed method's ability, through the extraction of task-relevant features in the hidden layer, to cluster images of the same digit and to cluster images of different digits that corresponds to states with the same optimal action. We also test the method's robustness with respect to different exploration schedules, i.e., different settings of the initial temperature and the temperature discount rate in softmax action selection. Our second task is a robot visual navigation task, where the robot can learn its position by the different colors of the lower part of four landmarks and it can infer the correct corner goal area by the color of the upper part of the landmarks. The state space consists of binarized camera images with, at most, nine different colors, which is equal to 6642 binary states. For both tasks, the learning performance is compared with standard FERL and with function approximation where the action-value function is approximated by a two-layered feedforward neural network. |
format | Online Article Text |
id | pubmed-3584292 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2013 |
publisher | Frontiers Media S.A. |
record_format | MEDLINE/PubMed |
spelling | pubmed-35842922013-02-28 Scaled free-energy based reinforcement learning for robust and efficient learning in high-dimensional state spaces Elfwing, Stefan Uchibe, Eiji Doya, Kenji Front Neurorobot Neuroscience Free-energy based reinforcement learning (FERL) was proposed for learning in high-dimensional state- and action spaces, which cannot be handled by standard function approximation methods. In this study, we propose a scaled version of free-energy based reinforcement learning to achieve more robust and more efficient learning performance. The action-value function is approximated by the negative free-energy of a restricted Boltzmann machine, divided by a constant scaling factor that is related to the size of the Boltzmann machine (the square root of the number of state nodes in this study). Our first task is a digit floor gridworld task, where the states are represented by images of handwritten digits from the MNIST data set. The purpose of the task is to investigate the proposed method's ability, through the extraction of task-relevant features in the hidden layer, to cluster images of the same digit and to cluster images of different digits that corresponds to states with the same optimal action. We also test the method's robustness with respect to different exploration schedules, i.e., different settings of the initial temperature and the temperature discount rate in softmax action selection. Our second task is a robot visual navigation task, where the robot can learn its position by the different colors of the lower part of four landmarks and it can infer the correct corner goal area by the color of the upper part of the landmarks. The state space consists of binarized camera images with, at most, nine different colors, which is equal to 6642 binary states. For both tasks, the learning performance is compared with standard FERL and with function approximation where the action-value function is approximated by a two-layered feedforward neural network. Frontiers Media S.A. 2013-02-28 /pmc/articles/PMC3584292/ /pubmed/23450126 http://dx.doi.org/10.3389/fnbot.2013.00003 Text en Copyright © 2013 Elfwing, Uchibe and Doya. http://creativecommons.org/licenses/by/3.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License, which permits use, distribution and reproduction in other forums, provided the original authors and source are credited and subject to any copyright notices concerning any third-party graphics etc. |
spellingShingle | Neuroscience Elfwing, Stefan Uchibe, Eiji Doya, Kenji Scaled free-energy based reinforcement learning for robust and efficient learning in high-dimensional state spaces |
title | Scaled free-energy based reinforcement learning for robust and efficient learning in high-dimensional state spaces |
title_full | Scaled free-energy based reinforcement learning for robust and efficient learning in high-dimensional state spaces |
title_fullStr | Scaled free-energy based reinforcement learning for robust and efficient learning in high-dimensional state spaces |
title_full_unstemmed | Scaled free-energy based reinforcement learning for robust and efficient learning in high-dimensional state spaces |
title_short | Scaled free-energy based reinforcement learning for robust and efficient learning in high-dimensional state spaces |
title_sort | scaled free-energy based reinforcement learning for robust and efficient learning in high-dimensional state spaces |
topic | Neuroscience |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3584292/ https://www.ncbi.nlm.nih.gov/pubmed/23450126 http://dx.doi.org/10.3389/fnbot.2013.00003 |
work_keys_str_mv | AT elfwingstefan scaledfreeenergybasedreinforcementlearningforrobustandefficientlearninginhighdimensionalstatespaces AT uchibeeiji scaledfreeenergybasedreinforcementlearningforrobustandefficientlearninginhighdimensionalstatespaces AT doyakenji scaledfreeenergybasedreinforcementlearningforrobustandefficientlearninginhighdimensionalstatespaces |