Cargando…

Navigation in Unknown Dynamic Environments Based on Deep Reinforcement Learning

In this paper, we propose a novel Deep Reinforcement Learning (DRL) algorithm which can navigate non-holonomic robots with continuous control in an unknown dynamic environment with moving obstacles. We call the approach MK-A3C (Memory and Knowledge-based Asynchronous Advantage Actor-Critic) for shor...

Descripción completa

Detalles Bibliográficos
Autores principales: Zeng, Junjie, Ju, Rusheng, Qin, Long, Hu, Yue, Yin, Quanjun, Hu, Cong
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2019
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6767106/
https://www.ncbi.nlm.nih.gov/pubmed/31491927
http://dx.doi.org/10.3390/s19183837
_version_ 1783454840825839616
author Zeng, Junjie
Ju, Rusheng
Qin, Long
Hu, Yue
Yin, Quanjun
Hu, Cong
author_facet Zeng, Junjie
Ju, Rusheng
Qin, Long
Hu, Yue
Yin, Quanjun
Hu, Cong
author_sort Zeng, Junjie
collection PubMed
description In this paper, we propose a novel Deep Reinforcement Learning (DRL) algorithm which can navigate non-holonomic robots with continuous control in an unknown dynamic environment with moving obstacles. We call the approach MK-A3C (Memory and Knowledge-based Asynchronous Advantage Actor-Critic) for short. As its first component, MK-A3C builds a GRU-based memory neural network to enhance the robot’s capability for temporal reasoning. Robots without it tend to suffer from a lack of rationality in face of incomplete and noisy estimations for complex environments. Additionally, robots with certain memory ability endowed by MK-A3C can avoid local minima traps by estimating the environmental model. Secondly, MK-A3C combines the domain knowledge-based reward function and the transfer learning-based training task architecture, which can solve the non-convergence policies problems caused by sparse reward. These improvements of MK-A3C can efficiently navigate robots in unknown dynamic environments, and satisfy kinetic constraints while handling moving objects. Simulation experiments show that compared with existing methods, MK-A3C can realize successful robotic navigation in unknown and challenging environments by outputting continuous acceleration commands.
format Online
Article
Text
id pubmed-6767106
institution National Center for Biotechnology Information
language English
publishDate 2019
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-67671062019-10-02 Navigation in Unknown Dynamic Environments Based on Deep Reinforcement Learning Zeng, Junjie Ju, Rusheng Qin, Long Hu, Yue Yin, Quanjun Hu, Cong Sensors (Basel) Article In this paper, we propose a novel Deep Reinforcement Learning (DRL) algorithm which can navigate non-holonomic robots with continuous control in an unknown dynamic environment with moving obstacles. We call the approach MK-A3C (Memory and Knowledge-based Asynchronous Advantage Actor-Critic) for short. As its first component, MK-A3C builds a GRU-based memory neural network to enhance the robot’s capability for temporal reasoning. Robots without it tend to suffer from a lack of rationality in face of incomplete and noisy estimations for complex environments. Additionally, robots with certain memory ability endowed by MK-A3C can avoid local minima traps by estimating the environmental model. Secondly, MK-A3C combines the domain knowledge-based reward function and the transfer learning-based training task architecture, which can solve the non-convergence policies problems caused by sparse reward. These improvements of MK-A3C can efficiently navigate robots in unknown dynamic environments, and satisfy kinetic constraints while handling moving objects. Simulation experiments show that compared with existing methods, MK-A3C can realize successful robotic navigation in unknown and challenging environments by outputting continuous acceleration commands. MDPI 2019-09-05 /pmc/articles/PMC6767106/ /pubmed/31491927 http://dx.doi.org/10.3390/s19183837 Text en © 2019 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Zeng, Junjie
Ju, Rusheng
Qin, Long
Hu, Yue
Yin, Quanjun
Hu, Cong
Navigation in Unknown Dynamic Environments Based on Deep Reinforcement Learning
title Navigation in Unknown Dynamic Environments Based on Deep Reinforcement Learning
title_full Navigation in Unknown Dynamic Environments Based on Deep Reinforcement Learning
title_fullStr Navigation in Unknown Dynamic Environments Based on Deep Reinforcement Learning
title_full_unstemmed Navigation in Unknown Dynamic Environments Based on Deep Reinforcement Learning
title_short Navigation in Unknown Dynamic Environments Based on Deep Reinforcement Learning
title_sort navigation in unknown dynamic environments based on deep reinforcement learning
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6767106/
https://www.ncbi.nlm.nih.gov/pubmed/31491927
http://dx.doi.org/10.3390/s19183837
work_keys_str_mv AT zengjunjie navigationinunknowndynamicenvironmentsbasedondeepreinforcementlearning
AT jurusheng navigationinunknowndynamicenvironmentsbasedondeepreinforcementlearning
AT qinlong navigationinunknowndynamicenvironmentsbasedondeepreinforcementlearning
AT huyue navigationinunknowndynamicenvironmentsbasedondeepreinforcementlearning
AT yinquanjun navigationinunknowndynamicenvironmentsbasedondeepreinforcementlearning
AT hucong navigationinunknowndynamicenvironmentsbasedondeepreinforcementlearning