Cargando…

KnowRU: Knowledge Reuse via Knowledge Distillation in Multi-Agent Reinforcement Learning

Recently, deep reinforcement learning (RL) algorithms have achieved significant progress in the multi-agent domain. However, training for increasingly complex tasks would be time-consuming and resource intensive. To alleviate this problem, efficient leveraging of historical experience is essential,...

Descripción completa

Detalles Bibliográficos
Autores principales: Gao, Zijian, Xu, Kele, Ding, Bo, Wang, Huaimin
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8393270/
https://www.ncbi.nlm.nih.gov/pubmed/34441184
http://dx.doi.org/10.3390/e23081043
_version_ 1783743694716796928
author Gao, Zijian
Xu, Kele
Ding, Bo
Wang, Huaimin
author_facet Gao, Zijian
Xu, Kele
Ding, Bo
Wang, Huaimin
author_sort Gao, Zijian
collection PubMed
description Recently, deep reinforcement learning (RL) algorithms have achieved significant progress in the multi-agent domain. However, training for increasingly complex tasks would be time-consuming and resource intensive. To alleviate this problem, efficient leveraging of historical experience is essential, which is under-explored in previous studies because most existing methods fail to achieve this goal in a continuously dynamic system owing to their complicated design. In this paper, we propose a method for knowledge reuse called “KnowRU”, which can be easily deployed in the majority of multi-agent reinforcement learning (MARL) algorithms without requiring complicated hand-coded design. We employ the knowledge distillation paradigm to transfer knowledge among agents to shorten the training phase for new tasks while improving the asymptotic performance of agents. To empirically demonstrate the robustness and effectiveness of KnowRU, we perform extensive experiments on state-of-the-art MARL algorithms in collaborative and competitive scenarios. The results show that KnowRU outperforms recently reported methods and not only successfully accelerates the training phase, but also improves the training performance, emphasizing the importance of the proposed knowledge reuse for MARL.
format Online
Article
Text
id pubmed-8393270
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-83932702021-08-28 KnowRU: Knowledge Reuse via Knowledge Distillation in Multi-Agent Reinforcement Learning Gao, Zijian Xu, Kele Ding, Bo Wang, Huaimin Entropy (Basel) Article Recently, deep reinforcement learning (RL) algorithms have achieved significant progress in the multi-agent domain. However, training for increasingly complex tasks would be time-consuming and resource intensive. To alleviate this problem, efficient leveraging of historical experience is essential, which is under-explored in previous studies because most existing methods fail to achieve this goal in a continuously dynamic system owing to their complicated design. In this paper, we propose a method for knowledge reuse called “KnowRU”, which can be easily deployed in the majority of multi-agent reinforcement learning (MARL) algorithms without requiring complicated hand-coded design. We employ the knowledge distillation paradigm to transfer knowledge among agents to shorten the training phase for new tasks while improving the asymptotic performance of agents. To empirically demonstrate the robustness and effectiveness of KnowRU, we perform extensive experiments on state-of-the-art MARL algorithms in collaborative and competitive scenarios. The results show that KnowRU outperforms recently reported methods and not only successfully accelerates the training phase, but also improves the training performance, emphasizing the importance of the proposed knowledge reuse for MARL. MDPI 2021-08-13 /pmc/articles/PMC8393270/ /pubmed/34441184 http://dx.doi.org/10.3390/e23081043 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Gao, Zijian
Xu, Kele
Ding, Bo
Wang, Huaimin
KnowRU: Knowledge Reuse via Knowledge Distillation in Multi-Agent Reinforcement Learning
title KnowRU: Knowledge Reuse via Knowledge Distillation in Multi-Agent Reinforcement Learning
title_full KnowRU: Knowledge Reuse via Knowledge Distillation in Multi-Agent Reinforcement Learning
title_fullStr KnowRU: Knowledge Reuse via Knowledge Distillation in Multi-Agent Reinforcement Learning
title_full_unstemmed KnowRU: Knowledge Reuse via Knowledge Distillation in Multi-Agent Reinforcement Learning
title_short KnowRU: Knowledge Reuse via Knowledge Distillation in Multi-Agent Reinforcement Learning
title_sort knowru: knowledge reuse via knowledge distillation in multi-agent reinforcement learning
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8393270/
https://www.ncbi.nlm.nih.gov/pubmed/34441184
http://dx.doi.org/10.3390/e23081043
work_keys_str_mv AT gaozijian knowruknowledgereuseviaknowledgedistillationinmultiagentreinforcementlearning
AT xukele knowruknowledgereuseviaknowledgedistillationinmultiagentreinforcementlearning
AT dingbo knowruknowledgereuseviaknowledgedistillationinmultiagentreinforcementlearning
AT wanghuaimin knowruknowledgereuseviaknowledgedistillationinmultiagentreinforcementlearning