Cargando…

Muecas: A Multi-Sensor Robotic Head for Affective Human Robot Interaction and Imitation

This paper presents a multi-sensor humanoid robotic head for human robot interaction. The design of the robotic head, Muecas, is based on ongoing research on the mechanisms of perception and imitation of human expressions and emotions. These mechanisms allow direct interaction between the robot and...

Descripción completa

Detalles Bibliográficos
Autores principales: Cid, Felipe, Moreno, Jose, Bustos, Pablo, Núñez, Pedro
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Molecular Diversity Preservation International (MDPI) 2014
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4063071/
https://www.ncbi.nlm.nih.gov/pubmed/24787636
http://dx.doi.org/10.3390/s140507711
_version_ 1782321742390231040
author Cid, Felipe
Moreno, Jose
Bustos, Pablo
Núñez, Pedro
author_facet Cid, Felipe
Moreno, Jose
Bustos, Pablo
Núñez, Pedro
author_sort Cid, Felipe
collection PubMed
description This paper presents a multi-sensor humanoid robotic head for human robot interaction. The design of the robotic head, Muecas, is based on ongoing research on the mechanisms of perception and imitation of human expressions and emotions. These mechanisms allow direct interaction between the robot and its human companion through the different natural language modalities: speech, body language and facial expressions. The robotic head has 12 degrees of freedom, in a human-like configuration, including eyes, eyebrows, mouth and neck, and has been designed and built entirely by IADeX (Engineering, Automation and Design of Extremadura) and RoboLab. A detailed description of its kinematics is provided along with the design of the most complex controllers. Muecas can be directly controlled by FACS (Facial Action Coding System), the de facto standard for facial expression recognition and synthesis. This feature facilitates its use by third party platforms and encourages the development of imitation and of goal-based systems. Imitation systems learn from the user, while goal-based ones use planning techniques to drive the user towards a final desired state. To show the flexibility and reliability of the robotic head, the paper presents a software architecture that is able to detect, recognize, classify and generate facial expressions in real time using FACS. This system has been implemented using the robotics framework, RoboComp, which provides hardware-independent access to the sensors in the head. Finally, the paper presents experimental results showing the real-time functioning of the whole system, including recognition and imitation of human facial expressions.
format Online
Article
Text
id pubmed-4063071
institution National Center for Biotechnology Information
language English
publishDate 2014
publisher Molecular Diversity Preservation International (MDPI)
record_format MEDLINE/PubMed
spelling pubmed-40630712014-06-19 Muecas: A Multi-Sensor Robotic Head for Affective Human Robot Interaction and Imitation Cid, Felipe Moreno, Jose Bustos, Pablo Núñez, Pedro Sensors (Basel) Article This paper presents a multi-sensor humanoid robotic head for human robot interaction. The design of the robotic head, Muecas, is based on ongoing research on the mechanisms of perception and imitation of human expressions and emotions. These mechanisms allow direct interaction between the robot and its human companion through the different natural language modalities: speech, body language and facial expressions. The robotic head has 12 degrees of freedom, in a human-like configuration, including eyes, eyebrows, mouth and neck, and has been designed and built entirely by IADeX (Engineering, Automation and Design of Extremadura) and RoboLab. A detailed description of its kinematics is provided along with the design of the most complex controllers. Muecas can be directly controlled by FACS (Facial Action Coding System), the de facto standard for facial expression recognition and synthesis. This feature facilitates its use by third party platforms and encourages the development of imitation and of goal-based systems. Imitation systems learn from the user, while goal-based ones use planning techniques to drive the user towards a final desired state. To show the flexibility and reliability of the robotic head, the paper presents a software architecture that is able to detect, recognize, classify and generate facial expressions in real time using FACS. This system has been implemented using the robotics framework, RoboComp, which provides hardware-independent access to the sensors in the head. Finally, the paper presents experimental results showing the real-time functioning of the whole system, including recognition and imitation of human facial expressions. Molecular Diversity Preservation International (MDPI) 2014-04-28 /pmc/articles/PMC4063071/ /pubmed/24787636 http://dx.doi.org/10.3390/s140507711 Text en © 2014 by the authors; licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution license (http://creativecommons.org/licenses/by/3.0/).
spellingShingle Article
Cid, Felipe
Moreno, Jose
Bustos, Pablo
Núñez, Pedro
Muecas: A Multi-Sensor Robotic Head for Affective Human Robot Interaction and Imitation
title Muecas: A Multi-Sensor Robotic Head for Affective Human Robot Interaction and Imitation
title_full Muecas: A Multi-Sensor Robotic Head for Affective Human Robot Interaction and Imitation
title_fullStr Muecas: A Multi-Sensor Robotic Head for Affective Human Robot Interaction and Imitation
title_full_unstemmed Muecas: A Multi-Sensor Robotic Head for Affective Human Robot Interaction and Imitation
title_short Muecas: A Multi-Sensor Robotic Head for Affective Human Robot Interaction and Imitation
title_sort muecas: a multi-sensor robotic head for affective human robot interaction and imitation
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4063071/
https://www.ncbi.nlm.nih.gov/pubmed/24787636
http://dx.doi.org/10.3390/s140507711
work_keys_str_mv AT cidfelipe muecasamultisensorroboticheadforaffectivehumanrobotinteractionandimitation
AT morenojose muecasamultisensorroboticheadforaffectivehumanrobotinteractionandimitation
AT bustospablo muecasamultisensorroboticheadforaffectivehumanrobotinteractionandimitation
AT nunezpedro muecasamultisensorroboticheadforaffectivehumanrobotinteractionandimitation