Cargando…
Learning Multirobot Hose Transportation and Deployment by Distributed Round-Robin Q-Learning
Multi-Agent Reinforcement Learning (MARL) algorithms face two main difficulties: the curse of dimensionality, and environment non-stationarity due to the independent learning processes carried out by the agents concurrently. In this paper we formalize and prove the convergence of a Distributed Round...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Public Library of Science
2015
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4497621/ https://www.ncbi.nlm.nih.gov/pubmed/26158587 http://dx.doi.org/10.1371/journal.pone.0127129 |
_version_ | 1782380530223808512 |
---|---|
author | Fernandez-Gauna, Borja Etxeberria-Agiriano, Ismael Graña, Manuel |
author_facet | Fernandez-Gauna, Borja Etxeberria-Agiriano, Ismael Graña, Manuel |
author_sort | Fernandez-Gauna, Borja |
collection | PubMed |
description | Multi-Agent Reinforcement Learning (MARL) algorithms face two main difficulties: the curse of dimensionality, and environment non-stationarity due to the independent learning processes carried out by the agents concurrently. In this paper we formalize and prove the convergence of a Distributed Round Robin Q-learning (D-RR-QL) algorithm for cooperative systems. The computational complexity of this algorithm increases linearly with the number of agents. Moreover, it eliminates environment non sta tionarity by carrying a round-robin scheduling of the action selection and execution. That this learning scheme allows the implementation of Modular State-Action Vetoes (MSAV) in cooperative multi-agent systems, which speeds up learning convergence in over-constrained systems by vetoing state-action pairs which lead to undesired termination states (UTS) in the relevant state-action subspace. Each agent’s local state-action value function learning is an independent process, including the MSAV policies. Coordination of locally optimal policies to obtain the global optimal joint policy is achieved by a greedy selection procedure using message passing. We show that D-RR-QL improves over state-of-the-art approaches, such as Distributed Q-Learning, Team Q-Learning and Coordinated Reinforcement Learning in a paradigmatic Linked Multi-Component Robotic System (L-MCRS) control problem: the hose transportation task. L-MCRS are over-constrained systems with many UTS induced by the interaction of the passive linking element and the active mobile robots. |
format | Online Article Text |
id | pubmed-4497621 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2015 |
publisher | Public Library of Science |
record_format | MEDLINE/PubMed |
spelling | pubmed-44976212015-07-14 Learning Multirobot Hose Transportation and Deployment by Distributed Round-Robin Q-Learning Fernandez-Gauna, Borja Etxeberria-Agiriano, Ismael Graña, Manuel PLoS One Research Article Multi-Agent Reinforcement Learning (MARL) algorithms face two main difficulties: the curse of dimensionality, and environment non-stationarity due to the independent learning processes carried out by the agents concurrently. In this paper we formalize and prove the convergence of a Distributed Round Robin Q-learning (D-RR-QL) algorithm for cooperative systems. The computational complexity of this algorithm increases linearly with the number of agents. Moreover, it eliminates environment non sta tionarity by carrying a round-robin scheduling of the action selection and execution. That this learning scheme allows the implementation of Modular State-Action Vetoes (MSAV) in cooperative multi-agent systems, which speeds up learning convergence in over-constrained systems by vetoing state-action pairs which lead to undesired termination states (UTS) in the relevant state-action subspace. Each agent’s local state-action value function learning is an independent process, including the MSAV policies. Coordination of locally optimal policies to obtain the global optimal joint policy is achieved by a greedy selection procedure using message passing. We show that D-RR-QL improves over state-of-the-art approaches, such as Distributed Q-Learning, Team Q-Learning and Coordinated Reinforcement Learning in a paradigmatic Linked Multi-Component Robotic System (L-MCRS) control problem: the hose transportation task. L-MCRS are over-constrained systems with many UTS induced by the interaction of the passive linking element and the active mobile robots. Public Library of Science 2015-07-09 /pmc/articles/PMC4497621/ /pubmed/26158587 http://dx.doi.org/10.1371/journal.pone.0127129 Text en © 2015 Fernandez-Gauna et al http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are properly credited. |
spellingShingle | Research Article Fernandez-Gauna, Borja Etxeberria-Agiriano, Ismael Graña, Manuel Learning Multirobot Hose Transportation and Deployment by Distributed Round-Robin Q-Learning |
title | Learning Multirobot Hose Transportation and Deployment by Distributed Round-Robin Q-Learning |
title_full | Learning Multirobot Hose Transportation and Deployment by Distributed Round-Robin Q-Learning |
title_fullStr | Learning Multirobot Hose Transportation and Deployment by Distributed Round-Robin Q-Learning |
title_full_unstemmed | Learning Multirobot Hose Transportation and Deployment by Distributed Round-Robin Q-Learning |
title_short | Learning Multirobot Hose Transportation and Deployment by Distributed Round-Robin Q-Learning |
title_sort | learning multirobot hose transportation and deployment by distributed round-robin q-learning |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4497621/ https://www.ncbi.nlm.nih.gov/pubmed/26158587 http://dx.doi.org/10.1371/journal.pone.0127129 |
work_keys_str_mv | AT fernandezgaunaborja learningmultirobothosetransportationanddeploymentbydistributedroundrobinqlearning AT etxeberriaagirianoismael learningmultirobothosetransportationanddeploymentbydistributedroundrobinqlearning AT granamanuel learningmultirobothosetransportationanddeploymentbydistributedroundrobinqlearning |