Cargando…

A Mapless Local Path Planning Approach Using Deep Reinforcement Learning Framework

The key module for autonomous mobile robots is path planning and obstacle avoidance. Global path planning based on known maps has been effectively achieved. Local path planning in unknown dynamic environments is still very challenging due to the lack of detailed environmental information and unpredi...

Descripción completa

Detalles Bibliográficos
Autores principales: Yin, Yan, Chen, Zhiyu, Liu, Gang, Guo, Jianwei
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9958619/
https://www.ncbi.nlm.nih.gov/pubmed/36850635
http://dx.doi.org/10.3390/s23042036
_version_ 1784895069136879616
author Yin, Yan
Chen, Zhiyu
Liu, Gang
Guo, Jianwei
author_facet Yin, Yan
Chen, Zhiyu
Liu, Gang
Guo, Jianwei
author_sort Yin, Yan
collection PubMed
description The key module for autonomous mobile robots is path planning and obstacle avoidance. Global path planning based on known maps has been effectively achieved. Local path planning in unknown dynamic environments is still very challenging due to the lack of detailed environmental information and unpredictability. This paper proposes an end-to-end local path planner n-step dueling double DQN with reward-based [Formula: see text]-greedy (RND3QN) based on a deep reinforcement learning framework, which acquires environmental data from LiDAR as input and uses a neural network to fit Q-values to output the corresponding discrete actions. The bias is reduced using n-step bootstrapping based on deep Q-network (DQN). The [Formula: see text]-greedy exploration-exploitation strategy is improved with the reward value as a measure of exploration, and an auxiliary reward function is introduced to increase the reward distribution of the sparse reward environment. Simulation experiments are conducted on the gazebo to test the algorithm’s effectiveness. The experimental data demonstrate that the average total reward value of RND3QN is higher than that of algorithms such as dueling double DQN (D3QN), and the success rates are increased by 174%, 65%, and 61% over D3QN on three stages, respectively. We experimented on the turtlebot3 waffle pi robot, and the strategies learned from the simulation can be effectively transferred to the real robot.
format Online
Article
Text
id pubmed-9958619
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-99586192023-02-26 A Mapless Local Path Planning Approach Using Deep Reinforcement Learning Framework Yin, Yan Chen, Zhiyu Liu, Gang Guo, Jianwei Sensors (Basel) Article The key module for autonomous mobile robots is path planning and obstacle avoidance. Global path planning based on known maps has been effectively achieved. Local path planning in unknown dynamic environments is still very challenging due to the lack of detailed environmental information and unpredictability. This paper proposes an end-to-end local path planner n-step dueling double DQN with reward-based [Formula: see text]-greedy (RND3QN) based on a deep reinforcement learning framework, which acquires environmental data from LiDAR as input and uses a neural network to fit Q-values to output the corresponding discrete actions. The bias is reduced using n-step bootstrapping based on deep Q-network (DQN). The [Formula: see text]-greedy exploration-exploitation strategy is improved with the reward value as a measure of exploration, and an auxiliary reward function is introduced to increase the reward distribution of the sparse reward environment. Simulation experiments are conducted on the gazebo to test the algorithm’s effectiveness. The experimental data demonstrate that the average total reward value of RND3QN is higher than that of algorithms such as dueling double DQN (D3QN), and the success rates are increased by 174%, 65%, and 61% over D3QN on three stages, respectively. We experimented on the turtlebot3 waffle pi robot, and the strategies learned from the simulation can be effectively transferred to the real robot. MDPI 2023-02-10 /pmc/articles/PMC9958619/ /pubmed/36850635 http://dx.doi.org/10.3390/s23042036 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Yin, Yan
Chen, Zhiyu
Liu, Gang
Guo, Jianwei
A Mapless Local Path Planning Approach Using Deep Reinforcement Learning Framework
title A Mapless Local Path Planning Approach Using Deep Reinforcement Learning Framework
title_full A Mapless Local Path Planning Approach Using Deep Reinforcement Learning Framework
title_fullStr A Mapless Local Path Planning Approach Using Deep Reinforcement Learning Framework
title_full_unstemmed A Mapless Local Path Planning Approach Using Deep Reinforcement Learning Framework
title_short A Mapless Local Path Planning Approach Using Deep Reinforcement Learning Framework
title_sort mapless local path planning approach using deep reinforcement learning framework
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9958619/
https://www.ncbi.nlm.nih.gov/pubmed/36850635
http://dx.doi.org/10.3390/s23042036
work_keys_str_mv AT yinyan amaplesslocalpathplanningapproachusingdeepreinforcementlearningframework
AT chenzhiyu amaplesslocalpathplanningapproachusingdeepreinforcementlearningframework
AT liugang amaplesslocalpathplanningapproachusingdeepreinforcementlearningframework
AT guojianwei amaplesslocalpathplanningapproachusingdeepreinforcementlearningframework
AT yinyan maplesslocalpathplanningapproachusingdeepreinforcementlearningframework
AT chenzhiyu maplesslocalpathplanningapproachusingdeepreinforcementlearningframework
AT liugang maplesslocalpathplanningapproachusingdeepreinforcementlearningframework
AT guojianwei maplesslocalpathplanningapproachusingdeepreinforcementlearningframework