Cargando…

Assistive Navigation Using Deep Reinforcement Learning Guiding Robot With UWB/Voice Beacons and Semantic Feedbacks for Blind and Visually Impaired People

Facilitating navigation in pedestrian environments is critical for enabling people who are blind and visually impaired (BVI) to achieve independent mobility. A deep reinforcement learning (DRL)–based assistive guiding robot with ultrawide-bandwidth (UWB) beacons that can navigate through routes with...

Descripción completa

Detalles Bibliográficos
Autores principales: Lu, Chen-Lung, Liu, Zi-Yan, Huang, Jui-Te, Huang, Ching-I, Wang, Bo-Hui, Chen, Yi, Wu, Nien-Hsin, Wang, Hsueh-Cheng, Giarré, Laura, Kuo, Pei-Yi
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8258111/
https://www.ncbi.nlm.nih.gov/pubmed/34239900
http://dx.doi.org/10.3389/frobt.2021.654132
_version_ 1783718440042758144
author Lu, Chen-Lung
Liu, Zi-Yan
Huang, Jui-Te
Huang, Ching-I
Wang, Bo-Hui
Chen, Yi
Wu, Nien-Hsin
Wang, Hsueh-Cheng
Giarré, Laura
Kuo, Pei-Yi
author_facet Lu, Chen-Lung
Liu, Zi-Yan
Huang, Jui-Te
Huang, Ching-I
Wang, Bo-Hui
Chen, Yi
Wu, Nien-Hsin
Wang, Hsueh-Cheng
Giarré, Laura
Kuo, Pei-Yi
author_sort Lu, Chen-Lung
collection PubMed
description Facilitating navigation in pedestrian environments is critical for enabling people who are blind and visually impaired (BVI) to achieve independent mobility. A deep reinforcement learning (DRL)–based assistive guiding robot with ultrawide-bandwidth (UWB) beacons that can navigate through routes with designated waypoints was designed in this study. Typically, a simultaneous localization and mapping (SLAM) framework is used to estimate the robot pose and navigational goal; however, SLAM frameworks are vulnerable in certain dynamic environments. The proposed navigation method is a learning approach based on state-of-the-art DRL and can effectively avoid obstacles. When used with UWB beacons, the proposed strategy is suitable for environments with dynamic pedestrians. We also designed a handle device with an audio interface that enables BVI users to interact with the guiding robot through intuitive feedback. The UWB beacons were installed with an audio interface to obtain environmental information. The on-handle and on-beacon verbal feedback provides points of interests and turn-by-turn information to BVI users. BVI users were recruited in this study to conduct navigation tasks in different scenarios. A route was designed in a simulated ward to represent daily activities. In real-world situations, SLAM-based state estimation might be affected by dynamic obstacles, and the visual-based trail may suffer from occlusions from pedestrians or other obstacles. The proposed system successfully navigated through environments with dynamic pedestrians, in which systems based on existing SLAM algorithms have failed.
format Online
Article
Text
id pubmed-8258111
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-82581112021-07-07 Assistive Navigation Using Deep Reinforcement Learning Guiding Robot With UWB/Voice Beacons and Semantic Feedbacks for Blind and Visually Impaired People Lu, Chen-Lung Liu, Zi-Yan Huang, Jui-Te Huang, Ching-I Wang, Bo-Hui Chen, Yi Wu, Nien-Hsin Wang, Hsueh-Cheng Giarré, Laura Kuo, Pei-Yi Front Robot AI Robotics and AI Facilitating navigation in pedestrian environments is critical for enabling people who are blind and visually impaired (BVI) to achieve independent mobility. A deep reinforcement learning (DRL)–based assistive guiding robot with ultrawide-bandwidth (UWB) beacons that can navigate through routes with designated waypoints was designed in this study. Typically, a simultaneous localization and mapping (SLAM) framework is used to estimate the robot pose and navigational goal; however, SLAM frameworks are vulnerable in certain dynamic environments. The proposed navigation method is a learning approach based on state-of-the-art DRL and can effectively avoid obstacles. When used with UWB beacons, the proposed strategy is suitable for environments with dynamic pedestrians. We also designed a handle device with an audio interface that enables BVI users to interact with the guiding robot through intuitive feedback. The UWB beacons were installed with an audio interface to obtain environmental information. The on-handle and on-beacon verbal feedback provides points of interests and turn-by-turn information to BVI users. BVI users were recruited in this study to conduct navigation tasks in different scenarios. A route was designed in a simulated ward to represent daily activities. In real-world situations, SLAM-based state estimation might be affected by dynamic obstacles, and the visual-based trail may suffer from occlusions from pedestrians or other obstacles. The proposed system successfully navigated through environments with dynamic pedestrians, in which systems based on existing SLAM algorithms have failed. Frontiers Media S.A. 2021-06-22 /pmc/articles/PMC8258111/ /pubmed/34239900 http://dx.doi.org/10.3389/frobt.2021.654132 Text en Copyright © 2021 Lu, Liu, Huang, Huang, Wang, Chen, Wu, Wang, Giarré and Kuo. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Robotics and AI
Lu, Chen-Lung
Liu, Zi-Yan
Huang, Jui-Te
Huang, Ching-I
Wang, Bo-Hui
Chen, Yi
Wu, Nien-Hsin
Wang, Hsueh-Cheng
Giarré, Laura
Kuo, Pei-Yi
Assistive Navigation Using Deep Reinforcement Learning Guiding Robot With UWB/Voice Beacons and Semantic Feedbacks for Blind and Visually Impaired People
title Assistive Navigation Using Deep Reinforcement Learning Guiding Robot With UWB/Voice Beacons and Semantic Feedbacks for Blind and Visually Impaired People
title_full Assistive Navigation Using Deep Reinforcement Learning Guiding Robot With UWB/Voice Beacons and Semantic Feedbacks for Blind and Visually Impaired People
title_fullStr Assistive Navigation Using Deep Reinforcement Learning Guiding Robot With UWB/Voice Beacons and Semantic Feedbacks for Blind and Visually Impaired People
title_full_unstemmed Assistive Navigation Using Deep Reinforcement Learning Guiding Robot With UWB/Voice Beacons and Semantic Feedbacks for Blind and Visually Impaired People
title_short Assistive Navigation Using Deep Reinforcement Learning Guiding Robot With UWB/Voice Beacons and Semantic Feedbacks for Blind and Visually Impaired People
title_sort assistive navigation using deep reinforcement learning guiding robot with uwb/voice beacons and semantic feedbacks for blind and visually impaired people
topic Robotics and AI
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8258111/
https://www.ncbi.nlm.nih.gov/pubmed/34239900
http://dx.doi.org/10.3389/frobt.2021.654132
work_keys_str_mv AT luchenlung assistivenavigationusingdeepreinforcementlearningguidingrobotwithuwbvoicebeaconsandsemanticfeedbacksforblindandvisuallyimpairedpeople
AT liuziyan assistivenavigationusingdeepreinforcementlearningguidingrobotwithuwbvoicebeaconsandsemanticfeedbacksforblindandvisuallyimpairedpeople
AT huangjuite assistivenavigationusingdeepreinforcementlearningguidingrobotwithuwbvoicebeaconsandsemanticfeedbacksforblindandvisuallyimpairedpeople
AT huangchingi assistivenavigationusingdeepreinforcementlearningguidingrobotwithuwbvoicebeaconsandsemanticfeedbacksforblindandvisuallyimpairedpeople
AT wangbohui assistivenavigationusingdeepreinforcementlearningguidingrobotwithuwbvoicebeaconsandsemanticfeedbacksforblindandvisuallyimpairedpeople
AT chenyi assistivenavigationusingdeepreinforcementlearningguidingrobotwithuwbvoicebeaconsandsemanticfeedbacksforblindandvisuallyimpairedpeople
AT wunienhsin assistivenavigationusingdeepreinforcementlearningguidingrobotwithuwbvoicebeaconsandsemanticfeedbacksforblindandvisuallyimpairedpeople
AT wanghsuehcheng assistivenavigationusingdeepreinforcementlearningguidingrobotwithuwbvoicebeaconsandsemanticfeedbacksforblindandvisuallyimpairedpeople
AT giarrelaura assistivenavigationusingdeepreinforcementlearningguidingrobotwithuwbvoicebeaconsandsemanticfeedbacksforblindandvisuallyimpairedpeople
AT kuopeiyi assistivenavigationusingdeepreinforcementlearningguidingrobotwithuwbvoicebeaconsandsemanticfeedbacksforblindandvisuallyimpairedpeople