Cargando…
A Mapless Local Path Planning Approach Using Deep Reinforcement Learning Framework
The key module for autonomous mobile robots is path planning and obstacle avoidance. Global path planning based on known maps has been effectively achieved. Local path planning in unknown dynamic environments is still very challenging due to the lack of detailed environmental information and unpredi...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9958619/ https://www.ncbi.nlm.nih.gov/pubmed/36850635 http://dx.doi.org/10.3390/s23042036 |
_version_ | 1784895069136879616 |
---|---|
author | Yin, Yan Chen, Zhiyu Liu, Gang Guo, Jianwei |
author_facet | Yin, Yan Chen, Zhiyu Liu, Gang Guo, Jianwei |
author_sort | Yin, Yan |
collection | PubMed |
description | The key module for autonomous mobile robots is path planning and obstacle avoidance. Global path planning based on known maps has been effectively achieved. Local path planning in unknown dynamic environments is still very challenging due to the lack of detailed environmental information and unpredictability. This paper proposes an end-to-end local path planner n-step dueling double DQN with reward-based [Formula: see text]-greedy (RND3QN) based on a deep reinforcement learning framework, which acquires environmental data from LiDAR as input and uses a neural network to fit Q-values to output the corresponding discrete actions. The bias is reduced using n-step bootstrapping based on deep Q-network (DQN). The [Formula: see text]-greedy exploration-exploitation strategy is improved with the reward value as a measure of exploration, and an auxiliary reward function is introduced to increase the reward distribution of the sparse reward environment. Simulation experiments are conducted on the gazebo to test the algorithm’s effectiveness. The experimental data demonstrate that the average total reward value of RND3QN is higher than that of algorithms such as dueling double DQN (D3QN), and the success rates are increased by 174%, 65%, and 61% over D3QN on three stages, respectively. We experimented on the turtlebot3 waffle pi robot, and the strategies learned from the simulation can be effectively transferred to the real robot. |
format | Online Article Text |
id | pubmed-9958619 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-99586192023-02-26 A Mapless Local Path Planning Approach Using Deep Reinforcement Learning Framework Yin, Yan Chen, Zhiyu Liu, Gang Guo, Jianwei Sensors (Basel) Article The key module for autonomous mobile robots is path planning and obstacle avoidance. Global path planning based on known maps has been effectively achieved. Local path planning in unknown dynamic environments is still very challenging due to the lack of detailed environmental information and unpredictability. This paper proposes an end-to-end local path planner n-step dueling double DQN with reward-based [Formula: see text]-greedy (RND3QN) based on a deep reinforcement learning framework, which acquires environmental data from LiDAR as input and uses a neural network to fit Q-values to output the corresponding discrete actions. The bias is reduced using n-step bootstrapping based on deep Q-network (DQN). The [Formula: see text]-greedy exploration-exploitation strategy is improved with the reward value as a measure of exploration, and an auxiliary reward function is introduced to increase the reward distribution of the sparse reward environment. Simulation experiments are conducted on the gazebo to test the algorithm’s effectiveness. The experimental data demonstrate that the average total reward value of RND3QN is higher than that of algorithms such as dueling double DQN (D3QN), and the success rates are increased by 174%, 65%, and 61% over D3QN on three stages, respectively. We experimented on the turtlebot3 waffle pi robot, and the strategies learned from the simulation can be effectively transferred to the real robot. MDPI 2023-02-10 /pmc/articles/PMC9958619/ /pubmed/36850635 http://dx.doi.org/10.3390/s23042036 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Yin, Yan Chen, Zhiyu Liu, Gang Guo, Jianwei A Mapless Local Path Planning Approach Using Deep Reinforcement Learning Framework |
title | A Mapless Local Path Planning Approach Using Deep Reinforcement Learning Framework |
title_full | A Mapless Local Path Planning Approach Using Deep Reinforcement Learning Framework |
title_fullStr | A Mapless Local Path Planning Approach Using Deep Reinforcement Learning Framework |
title_full_unstemmed | A Mapless Local Path Planning Approach Using Deep Reinforcement Learning Framework |
title_short | A Mapless Local Path Planning Approach Using Deep Reinforcement Learning Framework |
title_sort | mapless local path planning approach using deep reinforcement learning framework |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9958619/ https://www.ncbi.nlm.nih.gov/pubmed/36850635 http://dx.doi.org/10.3390/s23042036 |
work_keys_str_mv | AT yinyan amaplesslocalpathplanningapproachusingdeepreinforcementlearningframework AT chenzhiyu amaplesslocalpathplanningapproachusingdeepreinforcementlearningframework AT liugang amaplesslocalpathplanningapproachusingdeepreinforcementlearningframework AT guojianwei amaplesslocalpathplanningapproachusingdeepreinforcementlearningframework AT yinyan maplesslocalpathplanningapproachusingdeepreinforcementlearningframework AT chenzhiyu maplesslocalpathplanningapproachusingdeepreinforcementlearningframework AT liugang maplesslocalpathplanningapproachusingdeepreinforcementlearningframework AT guojianwei maplesslocalpathplanningapproachusingdeepreinforcementlearningframework |