Cargando…

Learning Trajectory Distributions for Assisted Teleoperation and Path Planning

Several approaches have been proposed to assist humans in co-manipulation and teleoperation tasks given demonstrated trajectories. However, these approaches are not applicable when the demonstrations are suboptimal or when the generalization capabilities of the learned models cannot cope with the ch...

Descripción completa

Detalles Bibliográficos
Autores principales: Ewerton, Marco, Arenz, Oleg, Maeda, Guilherme, Koert, Dorothea, Kolev, Zlatko, Takahashi, Masaki, Peters, Jan
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2019
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7806068/
https://www.ncbi.nlm.nih.gov/pubmed/33501104
http://dx.doi.org/10.3389/frobt.2019.00089
_version_ 1783636448967131136
author Ewerton, Marco
Arenz, Oleg
Maeda, Guilherme
Koert, Dorothea
Kolev, Zlatko
Takahashi, Masaki
Peters, Jan
author_facet Ewerton, Marco
Arenz, Oleg
Maeda, Guilherme
Koert, Dorothea
Kolev, Zlatko
Takahashi, Masaki
Peters, Jan
author_sort Ewerton, Marco
collection PubMed
description Several approaches have been proposed to assist humans in co-manipulation and teleoperation tasks given demonstrated trajectories. However, these approaches are not applicable when the demonstrations are suboptimal or when the generalization capabilities of the learned models cannot cope with the changes in the environment. Nevertheless, in real co-manipulation and teleoperation tasks, the original demonstrations will often be suboptimal and a learning system must be able to cope with new situations. This paper presents a reinforcement learning algorithm that can be applied to such problems. The proposed algorithm is initialized with a probability distribution of demonstrated trajectories and is based on the concept of relevance functions. We show in this paper how the relevance of trajectory parameters to optimization objectives is connected with the concept of Pearson correlation. First, we demonstrate the efficacy of our algorithm by addressing the assisted teleoperation of an object in a static virtual environment. Afterward, we extend this algorithm to deal with dynamic environments by utilizing Gaussian Process regression. The full framework is applied to make a point particle and a 7-DoF robot arm autonomously adapt their movements to changes in the environment as well as to assist the teleoperation of a 7-DoF robot arm in a dynamic environment.
format Online
Article
Text
id pubmed-7806068
institution National Center for Biotechnology Information
language English
publishDate 2019
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-78060682021-01-25 Learning Trajectory Distributions for Assisted Teleoperation and Path Planning Ewerton, Marco Arenz, Oleg Maeda, Guilherme Koert, Dorothea Kolev, Zlatko Takahashi, Masaki Peters, Jan Front Robot AI Robotics and AI Several approaches have been proposed to assist humans in co-manipulation and teleoperation tasks given demonstrated trajectories. However, these approaches are not applicable when the demonstrations are suboptimal or when the generalization capabilities of the learned models cannot cope with the changes in the environment. Nevertheless, in real co-manipulation and teleoperation tasks, the original demonstrations will often be suboptimal and a learning system must be able to cope with new situations. This paper presents a reinforcement learning algorithm that can be applied to such problems. The proposed algorithm is initialized with a probability distribution of demonstrated trajectories and is based on the concept of relevance functions. We show in this paper how the relevance of trajectory parameters to optimization objectives is connected with the concept of Pearson correlation. First, we demonstrate the efficacy of our algorithm by addressing the assisted teleoperation of an object in a static virtual environment. Afterward, we extend this algorithm to deal with dynamic environments by utilizing Gaussian Process regression. The full framework is applied to make a point particle and a 7-DoF robot arm autonomously adapt their movements to changes in the environment as well as to assist the teleoperation of a 7-DoF robot arm in a dynamic environment. Frontiers Media S.A. 2019-09-24 /pmc/articles/PMC7806068/ /pubmed/33501104 http://dx.doi.org/10.3389/frobt.2019.00089 Text en Copyright © 2019 Ewerton, Arenz, Maeda, Koert, Kolev, Takahashi and Peters. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Robotics and AI
Ewerton, Marco
Arenz, Oleg
Maeda, Guilherme
Koert, Dorothea
Kolev, Zlatko
Takahashi, Masaki
Peters, Jan
Learning Trajectory Distributions for Assisted Teleoperation and Path Planning
title Learning Trajectory Distributions for Assisted Teleoperation and Path Planning
title_full Learning Trajectory Distributions for Assisted Teleoperation and Path Planning
title_fullStr Learning Trajectory Distributions for Assisted Teleoperation and Path Planning
title_full_unstemmed Learning Trajectory Distributions for Assisted Teleoperation and Path Planning
title_short Learning Trajectory Distributions for Assisted Teleoperation and Path Planning
title_sort learning trajectory distributions for assisted teleoperation and path planning
topic Robotics and AI
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7806068/
https://www.ncbi.nlm.nih.gov/pubmed/33501104
http://dx.doi.org/10.3389/frobt.2019.00089
work_keys_str_mv AT ewertonmarco learningtrajectorydistributionsforassistedteleoperationandpathplanning
AT arenzoleg learningtrajectorydistributionsforassistedteleoperationandpathplanning
AT maedaguilherme learningtrajectorydistributionsforassistedteleoperationandpathplanning
AT koertdorothea learningtrajectorydistributionsforassistedteleoperationandpathplanning
AT kolevzlatko learningtrajectorydistributionsforassistedteleoperationandpathplanning
AT takahashimasaki learningtrajectorydistributionsforassistedteleoperationandpathplanning
AT petersjan learningtrajectorydistributionsforassistedteleoperationandpathplanning