Cargando…

Teaching Human Poses Interactively to a Social Robot

The main activity of social robots is to interact with people. In order to do that, the robot must be able to understand what the user is saying or doing. Typically, this capability consists of pre-programmed behaviors or is acquired through controlled learning processes, which are executed before t...

Descripción completa

Detalles Bibliográficos
Autores principales: Gonzalez-Pacheco, Victor, Malfaz, Maria, Fernandez, Fernando, Salichs, Miguel A.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2013
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3821356/
https://www.ncbi.nlm.nih.gov/pubmed/24048336
http://dx.doi.org/10.3390/s130912406
_version_ 1782290291577847808
author Gonzalez-Pacheco, Victor
Malfaz, Maria
Fernandez, Fernando
Salichs, Miguel A.
author_facet Gonzalez-Pacheco, Victor
Malfaz, Maria
Fernandez, Fernando
Salichs, Miguel A.
author_sort Gonzalez-Pacheco, Victor
collection PubMed
description The main activity of social robots is to interact with people. In order to do that, the robot must be able to understand what the user is saying or doing. Typically, this capability consists of pre-programmed behaviors or is acquired through controlled learning processes, which are executed before the social interaction begins. This paper presents a software architecture that enables a robot to learn poses in a similar way as people do. That is, hearing its teacher's explanations and acquiring new knowledge in real time. The architecture leans on two main components: an RGB-D (Red-, Green-, Blue- Depth) -based visual system, which gathers the user examples, and an Automatic Speech Recognition (ASR) system, which processes the speech describing those examples. The robot is able to naturally learn the poses the teacher is showing to it by maintaining a natural interaction with the teacher. We evaluate our system with 24 users who teach the robot a predetermined set of poses. The experimental results show that, with a few training examples, the system reaches high accuracy and robustness. This method shows how to combine data from the visual and auditory systems for the acquisition of new knowledge in a natural manner. Such a natural way of training enables robots to learn from users, even if they are not experts in robotics.
format Online
Article
Text
id pubmed-3821356
institution National Center for Biotechnology Information
language English
publishDate 2013
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-38213562013-11-09 Teaching Human Poses Interactively to a Social Robot Gonzalez-Pacheco, Victor Malfaz, Maria Fernandez, Fernando Salichs, Miguel A. Sensors (Basel) Article The main activity of social robots is to interact with people. In order to do that, the robot must be able to understand what the user is saying or doing. Typically, this capability consists of pre-programmed behaviors or is acquired through controlled learning processes, which are executed before the social interaction begins. This paper presents a software architecture that enables a robot to learn poses in a similar way as people do. That is, hearing its teacher's explanations and acquiring new knowledge in real time. The architecture leans on two main components: an RGB-D (Red-, Green-, Blue- Depth) -based visual system, which gathers the user examples, and an Automatic Speech Recognition (ASR) system, which processes the speech describing those examples. The robot is able to naturally learn the poses the teacher is showing to it by maintaining a natural interaction with the teacher. We evaluate our system with 24 users who teach the robot a predetermined set of poses. The experimental results show that, with a few training examples, the system reaches high accuracy and robustness. This method shows how to combine data from the visual and auditory systems for the acquisition of new knowledge in a natural manner. Such a natural way of training enables robots to learn from users, even if they are not experts in robotics. MDPI 2013-09-17 /pmc/articles/PMC3821356/ /pubmed/24048336 http://dx.doi.org/10.3390/s130912406 Text en © 2013 by the authors; licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution license (http://creativecommons.org/licenses/by/3.0/).
spellingShingle Article
Gonzalez-Pacheco, Victor
Malfaz, Maria
Fernandez, Fernando
Salichs, Miguel A.
Teaching Human Poses Interactively to a Social Robot
title Teaching Human Poses Interactively to a Social Robot
title_full Teaching Human Poses Interactively to a Social Robot
title_fullStr Teaching Human Poses Interactively to a Social Robot
title_full_unstemmed Teaching Human Poses Interactively to a Social Robot
title_short Teaching Human Poses Interactively to a Social Robot
title_sort teaching human poses interactively to a social robot
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3821356/
https://www.ncbi.nlm.nih.gov/pubmed/24048336
http://dx.doi.org/10.3390/s130912406
work_keys_str_mv AT gonzalezpachecovictor teachinghumanposesinteractivelytoasocialrobot
AT malfazmaria teachinghumanposesinteractivelytoasocialrobot
AT fernandezfernando teachinghumanposesinteractivelytoasocialrobot
AT salichsmiguela teachinghumanposesinteractivelytoasocialrobot