Cargando…

Improved Robot Path Planning Method Based on Deep Reinforcement Learning

With the advancement of robotics, the field of path planning is currently experiencing a period of prosperity. Researchers strive to address this nonlinear problem and have achieved remarkable results through the implementation of the Deep Reinforcement Learning (DRL) algorithm DQN (Deep Q-Network)....

Descripción completa

Detalles Bibliográficos
Autores principales: Han, Huiyan, Wang, Jiaqi, Kuang, Liqun, Han, Xie, Xue, Hongxin
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10304367/
https://www.ncbi.nlm.nih.gov/pubmed/37420785
http://dx.doi.org/10.3390/s23125622
_version_ 1785065490049138688
author Han, Huiyan
Wang, Jiaqi
Kuang, Liqun
Han, Xie
Xue, Hongxin
author_facet Han, Huiyan
Wang, Jiaqi
Kuang, Liqun
Han, Xie
Xue, Hongxin
author_sort Han, Huiyan
collection PubMed
description With the advancement of robotics, the field of path planning is currently experiencing a period of prosperity. Researchers strive to address this nonlinear problem and have achieved remarkable results through the implementation of the Deep Reinforcement Learning (DRL) algorithm DQN (Deep Q-Network). However, persistent challenges remain, including the curse of dimensionality, difficulties of model convergence and sparsity in rewards. To tackle these problems, this paper proposes an enhanced DDQN (Double DQN) path planning approach, in which the information after dimensionality reduction is fed into a two-branch network that incorporates expert knowledge and an optimized reward function to guide the training process. The data generated during the training phase are initially discretized into corresponding low-dimensional spaces. An “expert experience” module is introduced to facilitate the model’s early-stage training acceleration in the Epsilon–Greedy algorithm. To tackle navigation and obstacle avoidance separately, a dual-branch network structure is presented. We further optimize the reward function enabling intelligent agents to receive prompt feedback from the environment after performing each action. Experiments conducted in both virtual and real-world environments have demonstrated that the enhanced algorithm can accelerate model convergence, improve training stability and generate a smooth, shorter and collision-free path.
format Online
Article
Text
id pubmed-10304367
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-103043672023-06-29 Improved Robot Path Planning Method Based on Deep Reinforcement Learning Han, Huiyan Wang, Jiaqi Kuang, Liqun Han, Xie Xue, Hongxin Sensors (Basel) Article With the advancement of robotics, the field of path planning is currently experiencing a period of prosperity. Researchers strive to address this nonlinear problem and have achieved remarkable results through the implementation of the Deep Reinforcement Learning (DRL) algorithm DQN (Deep Q-Network). However, persistent challenges remain, including the curse of dimensionality, difficulties of model convergence and sparsity in rewards. To tackle these problems, this paper proposes an enhanced DDQN (Double DQN) path planning approach, in which the information after dimensionality reduction is fed into a two-branch network that incorporates expert knowledge and an optimized reward function to guide the training process. The data generated during the training phase are initially discretized into corresponding low-dimensional spaces. An “expert experience” module is introduced to facilitate the model’s early-stage training acceleration in the Epsilon–Greedy algorithm. To tackle navigation and obstacle avoidance separately, a dual-branch network structure is presented. We further optimize the reward function enabling intelligent agents to receive prompt feedback from the environment after performing each action. Experiments conducted in both virtual and real-world environments have demonstrated that the enhanced algorithm can accelerate model convergence, improve training stability and generate a smooth, shorter and collision-free path. MDPI 2023-06-15 /pmc/articles/PMC10304367/ /pubmed/37420785 http://dx.doi.org/10.3390/s23125622 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Han, Huiyan
Wang, Jiaqi
Kuang, Liqun
Han, Xie
Xue, Hongxin
Improved Robot Path Planning Method Based on Deep Reinforcement Learning
title Improved Robot Path Planning Method Based on Deep Reinforcement Learning
title_full Improved Robot Path Planning Method Based on Deep Reinforcement Learning
title_fullStr Improved Robot Path Planning Method Based on Deep Reinforcement Learning
title_full_unstemmed Improved Robot Path Planning Method Based on Deep Reinforcement Learning
title_short Improved Robot Path Planning Method Based on Deep Reinforcement Learning
title_sort improved robot path planning method based on deep reinforcement learning
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10304367/
https://www.ncbi.nlm.nih.gov/pubmed/37420785
http://dx.doi.org/10.3390/s23125622
work_keys_str_mv AT hanhuiyan improvedrobotpathplanningmethodbasedondeepreinforcementlearning
AT wangjiaqi improvedrobotpathplanningmethodbasedondeepreinforcementlearning
AT kuangliqun improvedrobotpathplanningmethodbasedondeepreinforcementlearning
AT hanxie improvedrobotpathplanningmethodbasedondeepreinforcementlearning
AT xuehongxin improvedrobotpathplanningmethodbasedondeepreinforcementlearning