Cargando…

Deep Reinforcement Learning Approach with Multiple Experience Pools for UAV’s Autonomous Motion Planning in Complex Unknown Environments

Autonomous motion planning (AMP) of unmanned aerial vehicles (UAVs) is aimed at enabling a UAV to safely fly to the target without human intervention. Recently, several emerging deep reinforcement learning (DRL) methods have been employed to address the AMP problem in some simplified environments, a...

Descripción completa

Detalles Bibliográficos
Autores principales: Hu, Zijian, Wan, Kaifang, Gao, Xiaoguang, Zhai, Yiwei, Wang, Qianglong
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7180781/
https://www.ncbi.nlm.nih.gov/pubmed/32235308
http://dx.doi.org/10.3390/s20071890
_version_ 1783525898640687104
author Hu, Zijian
Wan, Kaifang
Gao, Xiaoguang
Zhai, Yiwei
Wang, Qianglong
author_facet Hu, Zijian
Wan, Kaifang
Gao, Xiaoguang
Zhai, Yiwei
Wang, Qianglong
author_sort Hu, Zijian
collection PubMed
description Autonomous motion planning (AMP) of unmanned aerial vehicles (UAVs) is aimed at enabling a UAV to safely fly to the target without human intervention. Recently, several emerging deep reinforcement learning (DRL) methods have been employed to address the AMP problem in some simplified environments, and these methods have yielded good results. This paper proposes a multiple experience pools (MEPs) framework leveraging human expert experiences for DRL to speed up the learning process. Based on the deep deterministic policy gradient (DDPG) algorithm, a MEP–DDPG algorithm was designed using model predictive control and simulated annealing to generate expert experiences. On applying this algorithm to a complex unknown simulation environment constructed based on the parameters of the real UAV, the training experiment results showed that the novel DRL algorithm resulted in a performance improvement exceeding 20% as compared with the state-of-the-art DDPG. The results of the experimental testing indicate that UAVs trained using MEP–DDPG can stably complete a variety of tasks in complex, unknown environments.
format Online
Article
Text
id pubmed-7180781
institution National Center for Biotechnology Information
language English
publishDate 2020
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-71807812020-05-01 Deep Reinforcement Learning Approach with Multiple Experience Pools for UAV’s Autonomous Motion Planning in Complex Unknown Environments Hu, Zijian Wan, Kaifang Gao, Xiaoguang Zhai, Yiwei Wang, Qianglong Sensors (Basel) Article Autonomous motion planning (AMP) of unmanned aerial vehicles (UAVs) is aimed at enabling a UAV to safely fly to the target without human intervention. Recently, several emerging deep reinforcement learning (DRL) methods have been employed to address the AMP problem in some simplified environments, and these methods have yielded good results. This paper proposes a multiple experience pools (MEPs) framework leveraging human expert experiences for DRL to speed up the learning process. Based on the deep deterministic policy gradient (DDPG) algorithm, a MEP–DDPG algorithm was designed using model predictive control and simulated annealing to generate expert experiences. On applying this algorithm to a complex unknown simulation environment constructed based on the parameters of the real UAV, the training experiment results showed that the novel DRL algorithm resulted in a performance improvement exceeding 20% as compared with the state-of-the-art DDPG. The results of the experimental testing indicate that UAVs trained using MEP–DDPG can stably complete a variety of tasks in complex, unknown environments. MDPI 2020-03-29 /pmc/articles/PMC7180781/ /pubmed/32235308 http://dx.doi.org/10.3390/s20071890 Text en © 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Hu, Zijian
Wan, Kaifang
Gao, Xiaoguang
Zhai, Yiwei
Wang, Qianglong
Deep Reinforcement Learning Approach with Multiple Experience Pools for UAV’s Autonomous Motion Planning in Complex Unknown Environments
title Deep Reinforcement Learning Approach with Multiple Experience Pools for UAV’s Autonomous Motion Planning in Complex Unknown Environments
title_full Deep Reinforcement Learning Approach with Multiple Experience Pools for UAV’s Autonomous Motion Planning in Complex Unknown Environments
title_fullStr Deep Reinforcement Learning Approach with Multiple Experience Pools for UAV’s Autonomous Motion Planning in Complex Unknown Environments
title_full_unstemmed Deep Reinforcement Learning Approach with Multiple Experience Pools for UAV’s Autonomous Motion Planning in Complex Unknown Environments
title_short Deep Reinforcement Learning Approach with Multiple Experience Pools for UAV’s Autonomous Motion Planning in Complex Unknown Environments
title_sort deep reinforcement learning approach with multiple experience pools for uav’s autonomous motion planning in complex unknown environments
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7180781/
https://www.ncbi.nlm.nih.gov/pubmed/32235308
http://dx.doi.org/10.3390/s20071890
work_keys_str_mv AT huzijian deepreinforcementlearningapproachwithmultipleexperiencepoolsforuavsautonomousmotionplanningincomplexunknownenvironments
AT wankaifang deepreinforcementlearningapproachwithmultipleexperiencepoolsforuavsautonomousmotionplanningincomplexunknownenvironments
AT gaoxiaoguang deepreinforcementlearningapproachwithmultipleexperiencepoolsforuavsautonomousmotionplanningincomplexunknownenvironments
AT zhaiyiwei deepreinforcementlearningapproachwithmultipleexperiencepoolsforuavsautonomousmotionplanningincomplexunknownenvironments
AT wangqianglong deepreinforcementlearningapproachwithmultipleexperiencepoolsforuavsautonomousmotionplanningincomplexunknownenvironments