Cargando…

Deep reinforcement learning-aided autonomous navigation with landmark generators

Mobile robots are playing an increasingly significant role in social life and industrial production, such as searching and rescuing robots, autonomous exploration of sweeping robots, and so on. Improving the accuracy of autonomous navigation of mobile robots is a hot issue to be solved. However, tra...

Descripción completa

Detalles Bibliográficos
Autores principales: Wang, Xuanzhi, Sun, Yankang, Xie, Yuyang, Bin, Jiang, Xiao, Jian
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10477440/
https://www.ncbi.nlm.nih.gov/pubmed/37674856
http://dx.doi.org/10.3389/fnbot.2023.1200214
_version_ 1785101149516333056
author Wang, Xuanzhi
Sun, Yankang
Xie, Yuyang
Bin, Jiang
Xiao, Jian
author_facet Wang, Xuanzhi
Sun, Yankang
Xie, Yuyang
Bin, Jiang
Xiao, Jian
author_sort Wang, Xuanzhi
collection PubMed
description Mobile robots are playing an increasingly significant role in social life and industrial production, such as searching and rescuing robots, autonomous exploration of sweeping robots, and so on. Improving the accuracy of autonomous navigation of mobile robots is a hot issue to be solved. However, traditional navigation methods are unable to realize crash-free navigation in an environment with dynamic obstacles, more and more scholars are gradually using autonomous navigation based on deep reinforcement learning (DRL) to replace overly conservative traditional methods. But on the other hand, DRL's training time is too long, and the lack of long-term memory easily leads the robot to a dead end, which makes its application in the actual scene more difficult. To shorten training time and prevent mobile robots from getting stuck and spinning around, we design a new robot autonomous navigation framework which combines the traditional global planning and the local planning based on DRL. Therefore, the entire navigation process can be transformed into first using traditional navigation algorithms to find the global path, then searching for several high-value landmarks on the global path, and then using the DRL algorithm to move the mobile robot toward the designated landmarks to complete the final navigation, which makes the robot training difficulty greatly reduced. Furthermore, in order to improve the lack of long-term memory in deep reinforcement learning, we design a feature extraction network containing memory modules to preserve the long-term dependence of input features. Through comparing our methods with traditional navigation methods and reinforcement learning based on end-to-end depth navigation methods, it shows that while the number of dynamic obstacles is large and obstacles are rapidly moving, our proposed method is, on average, 20% better than the second ranked method in navigation efficiency (navigation time and navigation paths' length), 34% better than the second ranked method in safety (collision times), 26.6% higher than the second ranked method in success rate, and shows strong robustness.
format Online
Article
Text
id pubmed-10477440
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-104774402023-09-06 Deep reinforcement learning-aided autonomous navigation with landmark generators Wang, Xuanzhi Sun, Yankang Xie, Yuyang Bin, Jiang Xiao, Jian Front Neurorobot Neuroscience Mobile robots are playing an increasingly significant role in social life and industrial production, such as searching and rescuing robots, autonomous exploration of sweeping robots, and so on. Improving the accuracy of autonomous navigation of mobile robots is a hot issue to be solved. However, traditional navigation methods are unable to realize crash-free navigation in an environment with dynamic obstacles, more and more scholars are gradually using autonomous navigation based on deep reinforcement learning (DRL) to replace overly conservative traditional methods. But on the other hand, DRL's training time is too long, and the lack of long-term memory easily leads the robot to a dead end, which makes its application in the actual scene more difficult. To shorten training time and prevent mobile robots from getting stuck and spinning around, we design a new robot autonomous navigation framework which combines the traditional global planning and the local planning based on DRL. Therefore, the entire navigation process can be transformed into first using traditional navigation algorithms to find the global path, then searching for several high-value landmarks on the global path, and then using the DRL algorithm to move the mobile robot toward the designated landmarks to complete the final navigation, which makes the robot training difficulty greatly reduced. Furthermore, in order to improve the lack of long-term memory in deep reinforcement learning, we design a feature extraction network containing memory modules to preserve the long-term dependence of input features. Through comparing our methods with traditional navigation methods and reinforcement learning based on end-to-end depth navigation methods, it shows that while the number of dynamic obstacles is large and obstacles are rapidly moving, our proposed method is, on average, 20% better than the second ranked method in navigation efficiency (navigation time and navigation paths' length), 34% better than the second ranked method in safety (collision times), 26.6% higher than the second ranked method in success rate, and shows strong robustness. Frontiers Media S.A. 2023-08-22 /pmc/articles/PMC10477440/ /pubmed/37674856 http://dx.doi.org/10.3389/fnbot.2023.1200214 Text en Copyright © 2023 Wang, Sun, Xie, Bin and Xiao. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Neuroscience
Wang, Xuanzhi
Sun, Yankang
Xie, Yuyang
Bin, Jiang
Xiao, Jian
Deep reinforcement learning-aided autonomous navigation with landmark generators
title Deep reinforcement learning-aided autonomous navigation with landmark generators
title_full Deep reinforcement learning-aided autonomous navigation with landmark generators
title_fullStr Deep reinforcement learning-aided autonomous navigation with landmark generators
title_full_unstemmed Deep reinforcement learning-aided autonomous navigation with landmark generators
title_short Deep reinforcement learning-aided autonomous navigation with landmark generators
title_sort deep reinforcement learning-aided autonomous navigation with landmark generators
topic Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10477440/
https://www.ncbi.nlm.nih.gov/pubmed/37674856
http://dx.doi.org/10.3389/fnbot.2023.1200214
work_keys_str_mv AT wangxuanzhi deepreinforcementlearningaidedautonomousnavigationwithlandmarkgenerators
AT sunyankang deepreinforcementlearningaidedautonomousnavigationwithlandmarkgenerators
AT xieyuyang deepreinforcementlearningaidedautonomousnavigationwithlandmarkgenerators
AT binjiang deepreinforcementlearningaidedautonomousnavigationwithlandmarkgenerators
AT xiaojian deepreinforcementlearningaidedautonomousnavigationwithlandmarkgenerators