Cargando…

Intention Understanding in Human–Robot Interaction Based on Visual-NLP Semantics

With the rapid development of robotic and AI technology in recent years, human–robot interaction has made great advancement, making practical social impact. Verbal commands are one of the most direct and frequently used means for human–robot interaction. Currently, such technology can enable robots...

Descripción completa

Detalles Bibliográficos
Autores principales: Li, Zhihao, Mu, Yishan, Sun, Zhenglong, Song, Sifan, Su, Jionglong, Zhang, Jiaming
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7888278/
https://www.ncbi.nlm.nih.gov/pubmed/33613223
http://dx.doi.org/10.3389/fnbot.2020.610139
_version_ 1783652129644216320
author Li, Zhihao
Mu, Yishan
Sun, Zhenglong
Song, Sifan
Su, Jionglong
Zhang, Jiaming
author_facet Li, Zhihao
Mu, Yishan
Sun, Zhenglong
Song, Sifan
Su, Jionglong
Zhang, Jiaming
author_sort Li, Zhihao
collection PubMed
description With the rapid development of robotic and AI technology in recent years, human–robot interaction has made great advancement, making practical social impact. Verbal commands are one of the most direct and frequently used means for human–robot interaction. Currently, such technology can enable robots to execute pre-defined tasks based on simple and direct and explicit language instructions, e.g., certain keywords must be used and detected. However, that is not the natural way for human to communicate. In this paper, we propose a novel task-based framework to enable the robot to comprehend human intentions using visual semantics information, such that the robot is able to satisfy human intentions based on natural language instructions (total three types, namely clear, vague, and feeling, are defined and tested). The proposed framework includes a language semantics module to extract the keywords despite the explicitly of the command instruction, a visual object recognition module to identify the objects in front of the robot, and a similarity computation algorithm to infer the intention based on the given task. The task is then translated into the commands for the robot accordingly. Experiments are performed and validated on a humanoid robot with a defined task: to pick the desired item out of multiple objects on the table, and hand over to one desired user out of multiple human participants. The results show that our algorithm can interact with different types of instructions, even with unseen sentence structures.
format Online
Article
Text
id pubmed-7888278
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-78882782021-02-18 Intention Understanding in Human–Robot Interaction Based on Visual-NLP Semantics Li, Zhihao Mu, Yishan Sun, Zhenglong Song, Sifan Su, Jionglong Zhang, Jiaming Front Neurorobot Neuroscience With the rapid development of robotic and AI technology in recent years, human–robot interaction has made great advancement, making practical social impact. Verbal commands are one of the most direct and frequently used means for human–robot interaction. Currently, such technology can enable robots to execute pre-defined tasks based on simple and direct and explicit language instructions, e.g., certain keywords must be used and detected. However, that is not the natural way for human to communicate. In this paper, we propose a novel task-based framework to enable the robot to comprehend human intentions using visual semantics information, such that the robot is able to satisfy human intentions based on natural language instructions (total three types, namely clear, vague, and feeling, are defined and tested). The proposed framework includes a language semantics module to extract the keywords despite the explicitly of the command instruction, a visual object recognition module to identify the objects in front of the robot, and a similarity computation algorithm to infer the intention based on the given task. The task is then translated into the commands for the robot accordingly. Experiments are performed and validated on a humanoid robot with a defined task: to pick the desired item out of multiple objects on the table, and hand over to one desired user out of multiple human participants. The results show that our algorithm can interact with different types of instructions, even with unseen sentence structures. Frontiers Media S.A. 2021-02-02 /pmc/articles/PMC7888278/ /pubmed/33613223 http://dx.doi.org/10.3389/fnbot.2020.610139 Text en Copyright © 2021 Li, Mu, Sun, Song, Su and Zhang. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Neuroscience
Li, Zhihao
Mu, Yishan
Sun, Zhenglong
Song, Sifan
Su, Jionglong
Zhang, Jiaming
Intention Understanding in Human–Robot Interaction Based on Visual-NLP Semantics
title Intention Understanding in Human–Robot Interaction Based on Visual-NLP Semantics
title_full Intention Understanding in Human–Robot Interaction Based on Visual-NLP Semantics
title_fullStr Intention Understanding in Human–Robot Interaction Based on Visual-NLP Semantics
title_full_unstemmed Intention Understanding in Human–Robot Interaction Based on Visual-NLP Semantics
title_short Intention Understanding in Human–Robot Interaction Based on Visual-NLP Semantics
title_sort intention understanding in human–robot interaction based on visual-nlp semantics
topic Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7888278/
https://www.ncbi.nlm.nih.gov/pubmed/33613223
http://dx.doi.org/10.3389/fnbot.2020.610139
work_keys_str_mv AT lizhihao intentionunderstandinginhumanrobotinteractionbasedonvisualnlpsemantics
AT muyishan intentionunderstandinginhumanrobotinteractionbasedonvisualnlpsemantics
AT sunzhenglong intentionunderstandinginhumanrobotinteractionbasedonvisualnlpsemantics
AT songsifan intentionunderstandinginhumanrobotinteractionbasedonvisualnlpsemantics
AT sujionglong intentionunderstandinginhumanrobotinteractionbasedonvisualnlpsemantics
AT zhangjiaming intentionunderstandinginhumanrobotinteractionbasedonvisualnlpsemantics