Cargando…

Learning to reach by reinforcement learning using a receptive field based function approximation approach with continuous actions

Reinforcement learning methods can be used in robotics applications especially for specific target-oriented problems, for example the reward-based recalibration of goal directed actions. To this end still relatively large and continuous state-action spaces need to be efficiently handled. The goal of...

Descripción completa

Detalles Bibliográficos
Autores principales: Tamosiunaite, Minija, Asfour, Tamim, Wörgötter, Florentin
Formato: Texto
Lenguaje:English
Publicado: Springer-Verlag 2009
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC2798030/
https://www.ncbi.nlm.nih.gov/pubmed/19229556
http://dx.doi.org/10.1007/s00422-009-0295-8
_version_ 1782175709522821120
author Tamosiunaite, Minija
Asfour, Tamim
Wörgötter, Florentin
author_facet Tamosiunaite, Minija
Asfour, Tamim
Wörgötter, Florentin
author_sort Tamosiunaite, Minija
collection PubMed
description Reinforcement learning methods can be used in robotics applications especially for specific target-oriented problems, for example the reward-based recalibration of goal directed actions. To this end still relatively large and continuous state-action spaces need to be efficiently handled. The goal of this paper is, thus, to develop a novel, rather simple method which uses reinforcement learning with function approximation in conjunction with different reward-strategies for solving such problems. For the testing of our method, we use a four degree-of-freedom reaching problem in 3D-space simulated by a two-joint robot arm system with two DOF each. Function approximation is based on 4D, overlapping kernels (receptive fields) and the state-action space contains about 10,000 of these. Different types of reward structures are being compared, for example, reward-on- touching-only against reward-on-approach. Furthermore, forbidden joint configurations are punished. A continuous action space is used. In spite of a rather large number of states and the continuous action space these reward/punishment strategies allow the system to find a good solution usually within about 20 trials. The efficiency of our method demonstrated in this test scenario suggests that it might be possible to use it on a real robot for problems where mixed rewards can be defined in situations where other types of learning might be difficult.
format Text
id pubmed-2798030
institution National Center for Biotechnology Information
language English
publishDate 2009
publisher Springer-Verlag
record_format MEDLINE/PubMed
spelling pubmed-27980302010-01-13 Learning to reach by reinforcement learning using a receptive field based function approximation approach with continuous actions Tamosiunaite, Minija Asfour, Tamim Wörgötter, Florentin Biol Cybern Original Paper Reinforcement learning methods can be used in robotics applications especially for specific target-oriented problems, for example the reward-based recalibration of goal directed actions. To this end still relatively large and continuous state-action spaces need to be efficiently handled. The goal of this paper is, thus, to develop a novel, rather simple method which uses reinforcement learning with function approximation in conjunction with different reward-strategies for solving such problems. For the testing of our method, we use a four degree-of-freedom reaching problem in 3D-space simulated by a two-joint robot arm system with two DOF each. Function approximation is based on 4D, overlapping kernels (receptive fields) and the state-action space contains about 10,000 of these. Different types of reward structures are being compared, for example, reward-on- touching-only against reward-on-approach. Furthermore, forbidden joint configurations are punished. A continuous action space is used. In spite of a rather large number of states and the continuous action space these reward/punishment strategies allow the system to find a good solution usually within about 20 trials. The efficiency of our method demonstrated in this test scenario suggests that it might be possible to use it on a real robot for problems where mixed rewards can be defined in situations where other types of learning might be difficult. Springer-Verlag 2009-02-20 2009 /pmc/articles/PMC2798030/ /pubmed/19229556 http://dx.doi.org/10.1007/s00422-009-0295-8 Text en © The Author(s) 2009 https://creativecommons.org/licenses/by-nc/4.0/This article is distributed under the terms of the Creative Commons Attribution Noncommercial License which permits any noncommercial use, distribution, and reproduction in any medium, provided the original author(s) and source are credited.
spellingShingle Original Paper
Tamosiunaite, Minija
Asfour, Tamim
Wörgötter, Florentin
Learning to reach by reinforcement learning using a receptive field based function approximation approach with continuous actions
title Learning to reach by reinforcement learning using a receptive field based function approximation approach with continuous actions
title_full Learning to reach by reinforcement learning using a receptive field based function approximation approach with continuous actions
title_fullStr Learning to reach by reinforcement learning using a receptive field based function approximation approach with continuous actions
title_full_unstemmed Learning to reach by reinforcement learning using a receptive field based function approximation approach with continuous actions
title_short Learning to reach by reinforcement learning using a receptive field based function approximation approach with continuous actions
title_sort learning to reach by reinforcement learning using a receptive field based function approximation approach with continuous actions
topic Original Paper
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC2798030/
https://www.ncbi.nlm.nih.gov/pubmed/19229556
http://dx.doi.org/10.1007/s00422-009-0295-8
work_keys_str_mv AT tamosiunaiteminija learningtoreachbyreinforcementlearningusingareceptivefieldbasedfunctionapproximationapproachwithcontinuousactions
AT asfourtamim learningtoreachbyreinforcementlearningusingareceptivefieldbasedfunctionapproximationapproachwithcontinuousactions
AT worgotterflorentin learningtoreachbyreinforcementlearningusingareceptivefieldbasedfunctionapproximationapproachwithcontinuousactions