Cargando…
Deep reinforcement learning-aided autonomous navigation with landmark generators
Mobile robots are playing an increasingly significant role in social life and industrial production, such as searching and rescuing robots, autonomous exploration of sweeping robots, and so on. Improving the accuracy of autonomous navigation of mobile robots is a hot issue to be solved. However, tra...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Frontiers Media S.A.
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10477440/ https://www.ncbi.nlm.nih.gov/pubmed/37674856 http://dx.doi.org/10.3389/fnbot.2023.1200214 |
_version_ | 1785101149516333056 |
---|---|
author | Wang, Xuanzhi Sun, Yankang Xie, Yuyang Bin, Jiang Xiao, Jian |
author_facet | Wang, Xuanzhi Sun, Yankang Xie, Yuyang Bin, Jiang Xiao, Jian |
author_sort | Wang, Xuanzhi |
collection | PubMed |
description | Mobile robots are playing an increasingly significant role in social life and industrial production, such as searching and rescuing robots, autonomous exploration of sweeping robots, and so on. Improving the accuracy of autonomous navigation of mobile robots is a hot issue to be solved. However, traditional navigation methods are unable to realize crash-free navigation in an environment with dynamic obstacles, more and more scholars are gradually using autonomous navigation based on deep reinforcement learning (DRL) to replace overly conservative traditional methods. But on the other hand, DRL's training time is too long, and the lack of long-term memory easily leads the robot to a dead end, which makes its application in the actual scene more difficult. To shorten training time and prevent mobile robots from getting stuck and spinning around, we design a new robot autonomous navigation framework which combines the traditional global planning and the local planning based on DRL. Therefore, the entire navigation process can be transformed into first using traditional navigation algorithms to find the global path, then searching for several high-value landmarks on the global path, and then using the DRL algorithm to move the mobile robot toward the designated landmarks to complete the final navigation, which makes the robot training difficulty greatly reduced. Furthermore, in order to improve the lack of long-term memory in deep reinforcement learning, we design a feature extraction network containing memory modules to preserve the long-term dependence of input features. Through comparing our methods with traditional navigation methods and reinforcement learning based on end-to-end depth navigation methods, it shows that while the number of dynamic obstacles is large and obstacles are rapidly moving, our proposed method is, on average, 20% better than the second ranked method in navigation efficiency (navigation time and navigation paths' length), 34% better than the second ranked method in safety (collision times), 26.6% higher than the second ranked method in success rate, and shows strong robustness. |
format | Online Article Text |
id | pubmed-10477440 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | Frontiers Media S.A. |
record_format | MEDLINE/PubMed |
spelling | pubmed-104774402023-09-06 Deep reinforcement learning-aided autonomous navigation with landmark generators Wang, Xuanzhi Sun, Yankang Xie, Yuyang Bin, Jiang Xiao, Jian Front Neurorobot Neuroscience Mobile robots are playing an increasingly significant role in social life and industrial production, such as searching and rescuing robots, autonomous exploration of sweeping robots, and so on. Improving the accuracy of autonomous navigation of mobile robots is a hot issue to be solved. However, traditional navigation methods are unable to realize crash-free navigation in an environment with dynamic obstacles, more and more scholars are gradually using autonomous navigation based on deep reinforcement learning (DRL) to replace overly conservative traditional methods. But on the other hand, DRL's training time is too long, and the lack of long-term memory easily leads the robot to a dead end, which makes its application in the actual scene more difficult. To shorten training time and prevent mobile robots from getting stuck and spinning around, we design a new robot autonomous navigation framework which combines the traditional global planning and the local planning based on DRL. Therefore, the entire navigation process can be transformed into first using traditional navigation algorithms to find the global path, then searching for several high-value landmarks on the global path, and then using the DRL algorithm to move the mobile robot toward the designated landmarks to complete the final navigation, which makes the robot training difficulty greatly reduced. Furthermore, in order to improve the lack of long-term memory in deep reinforcement learning, we design a feature extraction network containing memory modules to preserve the long-term dependence of input features. Through comparing our methods with traditional navigation methods and reinforcement learning based on end-to-end depth navigation methods, it shows that while the number of dynamic obstacles is large and obstacles are rapidly moving, our proposed method is, on average, 20% better than the second ranked method in navigation efficiency (navigation time and navigation paths' length), 34% better than the second ranked method in safety (collision times), 26.6% higher than the second ranked method in success rate, and shows strong robustness. Frontiers Media S.A. 2023-08-22 /pmc/articles/PMC10477440/ /pubmed/37674856 http://dx.doi.org/10.3389/fnbot.2023.1200214 Text en Copyright © 2023 Wang, Sun, Xie, Bin and Xiao. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. |
spellingShingle | Neuroscience Wang, Xuanzhi Sun, Yankang Xie, Yuyang Bin, Jiang Xiao, Jian Deep reinforcement learning-aided autonomous navigation with landmark generators |
title | Deep reinforcement learning-aided autonomous navigation with landmark generators |
title_full | Deep reinforcement learning-aided autonomous navigation with landmark generators |
title_fullStr | Deep reinforcement learning-aided autonomous navigation with landmark generators |
title_full_unstemmed | Deep reinforcement learning-aided autonomous navigation with landmark generators |
title_short | Deep reinforcement learning-aided autonomous navigation with landmark generators |
title_sort | deep reinforcement learning-aided autonomous navigation with landmark generators |
topic | Neuroscience |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10477440/ https://www.ncbi.nlm.nih.gov/pubmed/37674856 http://dx.doi.org/10.3389/fnbot.2023.1200214 |
work_keys_str_mv | AT wangxuanzhi deepreinforcementlearningaidedautonomousnavigationwithlandmarkgenerators AT sunyankang deepreinforcementlearningaidedautonomousnavigationwithlandmarkgenerators AT xieyuyang deepreinforcementlearningaidedautonomousnavigationwithlandmarkgenerators AT binjiang deepreinforcementlearningaidedautonomousnavigationwithlandmarkgenerators AT xiaojian deepreinforcementlearningaidedautonomousnavigationwithlandmarkgenerators |