Cargando…
Navigation in Unknown Dynamic Environments Based on Deep Reinforcement Learning
In this paper, we propose a novel Deep Reinforcement Learning (DRL) algorithm which can navigate non-holonomic robots with continuous control in an unknown dynamic environment with moving obstacles. We call the approach MK-A3C (Memory and Knowledge-based Asynchronous Advantage Actor-Critic) for shor...
Autores principales: | , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2019
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6767106/ https://www.ncbi.nlm.nih.gov/pubmed/31491927 http://dx.doi.org/10.3390/s19183837 |
_version_ | 1783454840825839616 |
---|---|
author | Zeng, Junjie Ju, Rusheng Qin, Long Hu, Yue Yin, Quanjun Hu, Cong |
author_facet | Zeng, Junjie Ju, Rusheng Qin, Long Hu, Yue Yin, Quanjun Hu, Cong |
author_sort | Zeng, Junjie |
collection | PubMed |
description | In this paper, we propose a novel Deep Reinforcement Learning (DRL) algorithm which can navigate non-holonomic robots with continuous control in an unknown dynamic environment with moving obstacles. We call the approach MK-A3C (Memory and Knowledge-based Asynchronous Advantage Actor-Critic) for short. As its first component, MK-A3C builds a GRU-based memory neural network to enhance the robot’s capability for temporal reasoning. Robots without it tend to suffer from a lack of rationality in face of incomplete and noisy estimations for complex environments. Additionally, robots with certain memory ability endowed by MK-A3C can avoid local minima traps by estimating the environmental model. Secondly, MK-A3C combines the domain knowledge-based reward function and the transfer learning-based training task architecture, which can solve the non-convergence policies problems caused by sparse reward. These improvements of MK-A3C can efficiently navigate robots in unknown dynamic environments, and satisfy kinetic constraints while handling moving objects. Simulation experiments show that compared with existing methods, MK-A3C can realize successful robotic navigation in unknown and challenging environments by outputting continuous acceleration commands. |
format | Online Article Text |
id | pubmed-6767106 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2019 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-67671062019-10-02 Navigation in Unknown Dynamic Environments Based on Deep Reinforcement Learning Zeng, Junjie Ju, Rusheng Qin, Long Hu, Yue Yin, Quanjun Hu, Cong Sensors (Basel) Article In this paper, we propose a novel Deep Reinforcement Learning (DRL) algorithm which can navigate non-holonomic robots with continuous control in an unknown dynamic environment with moving obstacles. We call the approach MK-A3C (Memory and Knowledge-based Asynchronous Advantage Actor-Critic) for short. As its first component, MK-A3C builds a GRU-based memory neural network to enhance the robot’s capability for temporal reasoning. Robots without it tend to suffer from a lack of rationality in face of incomplete and noisy estimations for complex environments. Additionally, robots with certain memory ability endowed by MK-A3C can avoid local minima traps by estimating the environmental model. Secondly, MK-A3C combines the domain knowledge-based reward function and the transfer learning-based training task architecture, which can solve the non-convergence policies problems caused by sparse reward. These improvements of MK-A3C can efficiently navigate robots in unknown dynamic environments, and satisfy kinetic constraints while handling moving objects. Simulation experiments show that compared with existing methods, MK-A3C can realize successful robotic navigation in unknown and challenging environments by outputting continuous acceleration commands. MDPI 2019-09-05 /pmc/articles/PMC6767106/ /pubmed/31491927 http://dx.doi.org/10.3390/s19183837 Text en © 2019 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Zeng, Junjie Ju, Rusheng Qin, Long Hu, Yue Yin, Quanjun Hu, Cong Navigation in Unknown Dynamic Environments Based on Deep Reinforcement Learning |
title | Navigation in Unknown Dynamic Environments Based on Deep Reinforcement Learning |
title_full | Navigation in Unknown Dynamic Environments Based on Deep Reinforcement Learning |
title_fullStr | Navigation in Unknown Dynamic Environments Based on Deep Reinforcement Learning |
title_full_unstemmed | Navigation in Unknown Dynamic Environments Based on Deep Reinforcement Learning |
title_short | Navigation in Unknown Dynamic Environments Based on Deep Reinforcement Learning |
title_sort | navigation in unknown dynamic environments based on deep reinforcement learning |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6767106/ https://www.ncbi.nlm.nih.gov/pubmed/31491927 http://dx.doi.org/10.3390/s19183837 |
work_keys_str_mv | AT zengjunjie navigationinunknowndynamicenvironmentsbasedondeepreinforcementlearning AT jurusheng navigationinunknowndynamicenvironmentsbasedondeepreinforcementlearning AT qinlong navigationinunknowndynamicenvironmentsbasedondeepreinforcementlearning AT huyue navigationinunknowndynamicenvironmentsbasedondeepreinforcementlearning AT yinquanjun navigationinunknowndynamicenvironmentsbasedondeepreinforcementlearning AT hucong navigationinunknowndynamicenvironmentsbasedondeepreinforcementlearning |