Cargando…

Curiosity driven reinforcement learning for motion planning on humanoids

Most previous work on artificial curiosity (AC) and intrinsic motivation focuses on basic concepts and theory. Experimental results are generally limited to toy scenarios, such as navigation in a simulated maze, or control of a simple mechanical system with one or two degrees of freedom. To study AC...

Descripción completa

Detalles Bibliográficos
Autores principales: Frank, Mikhail, Leitner, Jürgen, Stollenga, Marijn, Förster, Alexander, Schmidhuber, Jürgen
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2014
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3881010/
https://www.ncbi.nlm.nih.gov/pubmed/24432001
http://dx.doi.org/10.3389/fnbot.2013.00025
_version_ 1782298142118510592
author Frank, Mikhail
Leitner, Jürgen
Stollenga, Marijn
Förster, Alexander
Schmidhuber, Jürgen
author_facet Frank, Mikhail
Leitner, Jürgen
Stollenga, Marijn
Förster, Alexander
Schmidhuber, Jürgen
author_sort Frank, Mikhail
collection PubMed
description Most previous work on artificial curiosity (AC) and intrinsic motivation focuses on basic concepts and theory. Experimental results are generally limited to toy scenarios, such as navigation in a simulated maze, or control of a simple mechanical system with one or two degrees of freedom. To study AC in a more realistic setting, we embody a curious agent in the complex iCub humanoid robot. Our novel reinforcement learning (RL) framework consists of a state-of-the-art, low-level, reactive control layer, which controls the iCub while respecting constraints, and a high-level curious agent, which explores the iCub's state-action space through information gain maximization, learning a world model from experience, controlling the actual iCub hardware in real-time. To the best of our knowledge, this is the first ever embodied, curious agent for real-time motion planning on a humanoid. We demonstrate that it can learn compact Markov models to represent large regions of the iCub's configuration space, and that the iCub explores intelligently, showing interest in its physical constraints as well as in objects it finds in its environment.
format Online
Article
Text
id pubmed-3881010
institution National Center for Biotechnology Information
language English
publishDate 2014
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-38810102014-01-15 Curiosity driven reinforcement learning for motion planning on humanoids Frank, Mikhail Leitner, Jürgen Stollenga, Marijn Förster, Alexander Schmidhuber, Jürgen Front Neurorobot Neuroscience Most previous work on artificial curiosity (AC) and intrinsic motivation focuses on basic concepts and theory. Experimental results are generally limited to toy scenarios, such as navigation in a simulated maze, or control of a simple mechanical system with one or two degrees of freedom. To study AC in a more realistic setting, we embody a curious agent in the complex iCub humanoid robot. Our novel reinforcement learning (RL) framework consists of a state-of-the-art, low-level, reactive control layer, which controls the iCub while respecting constraints, and a high-level curious agent, which explores the iCub's state-action space through information gain maximization, learning a world model from experience, controlling the actual iCub hardware in real-time. To the best of our knowledge, this is the first ever embodied, curious agent for real-time motion planning on a humanoid. We demonstrate that it can learn compact Markov models to represent large regions of the iCub's configuration space, and that the iCub explores intelligently, showing interest in its physical constraints as well as in objects it finds in its environment. Frontiers Media S.A. 2014-01-06 /pmc/articles/PMC3881010/ /pubmed/24432001 http://dx.doi.org/10.3389/fnbot.2013.00025 Text en Copyright © 2014 Frank, Leitner, Stollenga, Förster and Schmidhuber. http://creativecommons.org/licenses/by/3.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Neuroscience
Frank, Mikhail
Leitner, Jürgen
Stollenga, Marijn
Förster, Alexander
Schmidhuber, Jürgen
Curiosity driven reinforcement learning for motion planning on humanoids
title Curiosity driven reinforcement learning for motion planning on humanoids
title_full Curiosity driven reinforcement learning for motion planning on humanoids
title_fullStr Curiosity driven reinforcement learning for motion planning on humanoids
title_full_unstemmed Curiosity driven reinforcement learning for motion planning on humanoids
title_short Curiosity driven reinforcement learning for motion planning on humanoids
title_sort curiosity driven reinforcement learning for motion planning on humanoids
topic Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3881010/
https://www.ncbi.nlm.nih.gov/pubmed/24432001
http://dx.doi.org/10.3389/fnbot.2013.00025
work_keys_str_mv AT frankmikhail curiositydrivenreinforcementlearningformotionplanningonhumanoids
AT leitnerjurgen curiositydrivenreinforcementlearningformotionplanningonhumanoids
AT stollengamarijn curiositydrivenreinforcementlearningformotionplanningonhumanoids
AT forsteralexander curiositydrivenreinforcementlearningformotionplanningonhumanoids
AT schmidhuberjurgen curiositydrivenreinforcementlearningformotionplanningonhumanoids