Cargando…
Learning to reach by reinforcement learning using a receptive field based function approximation approach with continuous actions
Reinforcement learning methods can be used in robotics applications especially for specific target-oriented problems, for example the reward-based recalibration of goal directed actions. To this end still relatively large and continuous state-action spaces need to be efficiently handled. The goal of...
Autores principales: | , , |
---|---|
Formato: | Texto |
Lenguaje: | English |
Publicado: |
Springer-Verlag
2009
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC2798030/ https://www.ncbi.nlm.nih.gov/pubmed/19229556 http://dx.doi.org/10.1007/s00422-009-0295-8 |
_version_ | 1782175709522821120 |
---|---|
author | Tamosiunaite, Minija Asfour, Tamim Wörgötter, Florentin |
author_facet | Tamosiunaite, Minija Asfour, Tamim Wörgötter, Florentin |
author_sort | Tamosiunaite, Minija |
collection | PubMed |
description | Reinforcement learning methods can be used in robotics applications especially for specific target-oriented problems, for example the reward-based recalibration of goal directed actions. To this end still relatively large and continuous state-action spaces need to be efficiently handled. The goal of this paper is, thus, to develop a novel, rather simple method which uses reinforcement learning with function approximation in conjunction with different reward-strategies for solving such problems. For the testing of our method, we use a four degree-of-freedom reaching problem in 3D-space simulated by a two-joint robot arm system with two DOF each. Function approximation is based on 4D, overlapping kernels (receptive fields) and the state-action space contains about 10,000 of these. Different types of reward structures are being compared, for example, reward-on- touching-only against reward-on-approach. Furthermore, forbidden joint configurations are punished. A continuous action space is used. In spite of a rather large number of states and the continuous action space these reward/punishment strategies allow the system to find a good solution usually within about 20 trials. The efficiency of our method demonstrated in this test scenario suggests that it might be possible to use it on a real robot for problems where mixed rewards can be defined in situations where other types of learning might be difficult. |
format | Text |
id | pubmed-2798030 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2009 |
publisher | Springer-Verlag |
record_format | MEDLINE/PubMed |
spelling | pubmed-27980302010-01-13 Learning to reach by reinforcement learning using a receptive field based function approximation approach with continuous actions Tamosiunaite, Minija Asfour, Tamim Wörgötter, Florentin Biol Cybern Original Paper Reinforcement learning methods can be used in robotics applications especially for specific target-oriented problems, for example the reward-based recalibration of goal directed actions. To this end still relatively large and continuous state-action spaces need to be efficiently handled. The goal of this paper is, thus, to develop a novel, rather simple method which uses reinforcement learning with function approximation in conjunction with different reward-strategies for solving such problems. For the testing of our method, we use a four degree-of-freedom reaching problem in 3D-space simulated by a two-joint robot arm system with two DOF each. Function approximation is based on 4D, overlapping kernels (receptive fields) and the state-action space contains about 10,000 of these. Different types of reward structures are being compared, for example, reward-on- touching-only against reward-on-approach. Furthermore, forbidden joint configurations are punished. A continuous action space is used. In spite of a rather large number of states and the continuous action space these reward/punishment strategies allow the system to find a good solution usually within about 20 trials. The efficiency of our method demonstrated in this test scenario suggests that it might be possible to use it on a real robot for problems where mixed rewards can be defined in situations where other types of learning might be difficult. Springer-Verlag 2009-02-20 2009 /pmc/articles/PMC2798030/ /pubmed/19229556 http://dx.doi.org/10.1007/s00422-009-0295-8 Text en © The Author(s) 2009 https://creativecommons.org/licenses/by-nc/4.0/This article is distributed under the terms of the Creative Commons Attribution Noncommercial License which permits any noncommercial use, distribution, and reproduction in any medium, provided the original author(s) and source are credited. |
spellingShingle | Original Paper Tamosiunaite, Minija Asfour, Tamim Wörgötter, Florentin Learning to reach by reinforcement learning using a receptive field based function approximation approach with continuous actions |
title | Learning to reach by reinforcement learning using a receptive field based function approximation approach with continuous actions |
title_full | Learning to reach by reinforcement learning using a receptive field based function approximation approach with continuous actions |
title_fullStr | Learning to reach by reinforcement learning using a receptive field based function approximation approach with continuous actions |
title_full_unstemmed | Learning to reach by reinforcement learning using a receptive field based function approximation approach with continuous actions |
title_short | Learning to reach by reinforcement learning using a receptive field based function approximation approach with continuous actions |
title_sort | learning to reach by reinforcement learning using a receptive field based function approximation approach with continuous actions |
topic | Original Paper |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC2798030/ https://www.ncbi.nlm.nih.gov/pubmed/19229556 http://dx.doi.org/10.1007/s00422-009-0295-8 |
work_keys_str_mv | AT tamosiunaiteminija learningtoreachbyreinforcementlearningusingareceptivefieldbasedfunctionapproximationapproachwithcontinuousactions AT asfourtamim learningtoreachbyreinforcementlearningusingareceptivefieldbasedfunctionapproximationapproachwithcontinuousactions AT worgotterflorentin learningtoreachbyreinforcementlearningusingareceptivefieldbasedfunctionapproximationapproachwithcontinuousactions |