Cargando…
Efficient Path Planning for Mobile Robot Based on Deep Deterministic Policy Gradient
When a traditional Deep Deterministic Policy Gradient (DDPG) algorithm is used in mobile robot path planning, due to the limited observable environment of mobile robots, the training efficiency of the path planning model is low, and the convergence speed is slow. In this paper, Long Short-Term Memor...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9102217/ https://www.ncbi.nlm.nih.gov/pubmed/35591271 http://dx.doi.org/10.3390/s22093579 |
_version_ | 1784707277905723392 |
---|---|
author | Gong, Hui Wang, Peng Ni, Cui Cheng, Nuo |
author_facet | Gong, Hui Wang, Peng Ni, Cui Cheng, Nuo |
author_sort | Gong, Hui |
collection | PubMed |
description | When a traditional Deep Deterministic Policy Gradient (DDPG) algorithm is used in mobile robot path planning, due to the limited observable environment of mobile robots, the training efficiency of the path planning model is low, and the convergence speed is slow. In this paper, Long Short-Term Memory (LSTM) is introduced into the DDPG network, the former and current states of the mobile robot are combined to determine the actions of the robot, and a Batch Norm layer is added after each layer of the Actor network. At the same time, the reward function is optimized to guide the mobile robot to move faster towards the target point. In order to improve the learning efficiency, different normalization methods are used to normalize the distance and angle between the mobile robot and the target point, which are used as the input of the DDPG network model. When the model outputs the next action of the mobile robot, mixed noise composed of Gaussian noise and Ornstein–Uhlenbeck (OU) noise is added. Finally, the simulation environment built by a ROS system and a Gazebo platform is used for experiments. The results show that the proposed algorithm can accelerate the convergence speed of DDPG, improve the generalization ability of the path planning model and improve the efficiency and success rate of mobile robot path planning. |
format | Online Article Text |
id | pubmed-9102217 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-91022172022-05-14 Efficient Path Planning for Mobile Robot Based on Deep Deterministic Policy Gradient Gong, Hui Wang, Peng Ni, Cui Cheng, Nuo Sensors (Basel) Article When a traditional Deep Deterministic Policy Gradient (DDPG) algorithm is used in mobile robot path planning, due to the limited observable environment of mobile robots, the training efficiency of the path planning model is low, and the convergence speed is slow. In this paper, Long Short-Term Memory (LSTM) is introduced into the DDPG network, the former and current states of the mobile robot are combined to determine the actions of the robot, and a Batch Norm layer is added after each layer of the Actor network. At the same time, the reward function is optimized to guide the mobile robot to move faster towards the target point. In order to improve the learning efficiency, different normalization methods are used to normalize the distance and angle between the mobile robot and the target point, which are used as the input of the DDPG network model. When the model outputs the next action of the mobile robot, mixed noise composed of Gaussian noise and Ornstein–Uhlenbeck (OU) noise is added. Finally, the simulation environment built by a ROS system and a Gazebo platform is used for experiments. The results show that the proposed algorithm can accelerate the convergence speed of DDPG, improve the generalization ability of the path planning model and improve the efficiency and success rate of mobile robot path planning. MDPI 2022-05-08 /pmc/articles/PMC9102217/ /pubmed/35591271 http://dx.doi.org/10.3390/s22093579 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Gong, Hui Wang, Peng Ni, Cui Cheng, Nuo Efficient Path Planning for Mobile Robot Based on Deep Deterministic Policy Gradient |
title | Efficient Path Planning for Mobile Robot Based on Deep Deterministic Policy Gradient |
title_full | Efficient Path Planning for Mobile Robot Based on Deep Deterministic Policy Gradient |
title_fullStr | Efficient Path Planning for Mobile Robot Based on Deep Deterministic Policy Gradient |
title_full_unstemmed | Efficient Path Planning for Mobile Robot Based on Deep Deterministic Policy Gradient |
title_short | Efficient Path Planning for Mobile Robot Based on Deep Deterministic Policy Gradient |
title_sort | efficient path planning for mobile robot based on deep deterministic policy gradient |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9102217/ https://www.ncbi.nlm.nih.gov/pubmed/35591271 http://dx.doi.org/10.3390/s22093579 |
work_keys_str_mv | AT gonghui efficientpathplanningformobilerobotbasedondeepdeterministicpolicygradient AT wangpeng efficientpathplanningformobilerobotbasedondeepdeterministicpolicygradient AT nicui efficientpathplanningformobilerobotbasedondeepdeterministicpolicygradient AT chengnuo efficientpathplanningformobilerobotbasedondeepdeterministicpolicygradient |