Cargando…

Group Emotion Detection Based on Social Robot Perception

Social robotics is an emerging area that is becoming present in social spaces, by introducing autonomous social robots. Social robots offer services, perform tasks, and interact with people in such social environments, demanding more efficient and complex Human–Robot Interaction (HRI) designs. A str...

Descripción completa

Detalles Bibliográficos
Autores principales: Quiroz, Marco, Patiño, Raquel, Diaz-Amado, José, Cardinale, Yudith
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9145339/
https://www.ncbi.nlm.nih.gov/pubmed/35632160
http://dx.doi.org/10.3390/s22103749
_version_ 1784716263108378624
author Quiroz, Marco
Patiño, Raquel
Diaz-Amado, José
Cardinale, Yudith
author_facet Quiroz, Marco
Patiño, Raquel
Diaz-Amado, José
Cardinale, Yudith
author_sort Quiroz, Marco
collection PubMed
description Social robotics is an emerging area that is becoming present in social spaces, by introducing autonomous social robots. Social robots offer services, perform tasks, and interact with people in such social environments, demanding more efficient and complex Human–Robot Interaction (HRI) designs. A strategy to improve HRI is to provide robots with the capacity of detecting the emotions of the people around them to plan a trajectory, modify their behaviour, and generate an appropriate interaction with people based on the analysed information. However, in social environments in which it is common to find a group of persons, new approaches are needed in order to make robots able to recognise groups of people and the emotion of the groups, which can be also associated with a scene in which the group is participating. Some existing studies are focused on detecting group cohesion and the recognition of group emotions; nevertheless, these works do not focus on performing the recognition tasks from a robocentric perspective, considering the sensory capacity of robots. In this context, a system to recognise scenes in terms of groups of people, to then detect global (prevailing) emotions in a scene, is presented. The approach proposed to visualise and recognise emotions in typical HRI is based on the face size of people recognised by the robot during its navigation (face sizes decrease when the robot moves away from a group of people). On each frame of the video stream of the visual sensor, individual emotions are recognised based on the Visual Geometry Group (VGG) neural network pre-trained to recognise faces (VGGFace); then, to detect the emotion of the frame, individual emotions are aggregated with a fusion method, and consequently, to detect global (prevalent) emotion in the scene (group of people), the emotions of its constituent frames are also aggregated. Additionally, this work proposes a strategy to create datasets with images/videos in order to validate the estimation of emotions in scenes and personal emotions. Both datasets are generated in a simulated environment based on the Robot Operating System (ROS) from videos captured by robots through their sensory capabilities. Tests are performed in two simulated environments in ROS/Gazebo: a museum and a cafeteria. Results show that the accuracy in the detection of individual emotions is 99.79% and the detection of group emotion (scene emotion) in each frame is 90.84% and 89.78% in the cafeteria and the museum scenarios, respectively.
format Online
Article
Text
id pubmed-9145339
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-91453392022-05-29 Group Emotion Detection Based on Social Robot Perception Quiroz, Marco Patiño, Raquel Diaz-Amado, José Cardinale, Yudith Sensors (Basel) Article Social robotics is an emerging area that is becoming present in social spaces, by introducing autonomous social robots. Social robots offer services, perform tasks, and interact with people in such social environments, demanding more efficient and complex Human–Robot Interaction (HRI) designs. A strategy to improve HRI is to provide robots with the capacity of detecting the emotions of the people around them to plan a trajectory, modify their behaviour, and generate an appropriate interaction with people based on the analysed information. However, in social environments in which it is common to find a group of persons, new approaches are needed in order to make robots able to recognise groups of people and the emotion of the groups, which can be also associated with a scene in which the group is participating. Some existing studies are focused on detecting group cohesion and the recognition of group emotions; nevertheless, these works do not focus on performing the recognition tasks from a robocentric perspective, considering the sensory capacity of robots. In this context, a system to recognise scenes in terms of groups of people, to then detect global (prevailing) emotions in a scene, is presented. The approach proposed to visualise and recognise emotions in typical HRI is based on the face size of people recognised by the robot during its navigation (face sizes decrease when the robot moves away from a group of people). On each frame of the video stream of the visual sensor, individual emotions are recognised based on the Visual Geometry Group (VGG) neural network pre-trained to recognise faces (VGGFace); then, to detect the emotion of the frame, individual emotions are aggregated with a fusion method, and consequently, to detect global (prevalent) emotion in the scene (group of people), the emotions of its constituent frames are also aggregated. Additionally, this work proposes a strategy to create datasets with images/videos in order to validate the estimation of emotions in scenes and personal emotions. Both datasets are generated in a simulated environment based on the Robot Operating System (ROS) from videos captured by robots through their sensory capabilities. Tests are performed in two simulated environments in ROS/Gazebo: a museum and a cafeteria. Results show that the accuracy in the detection of individual emotions is 99.79% and the detection of group emotion (scene emotion) in each frame is 90.84% and 89.78% in the cafeteria and the museum scenarios, respectively. MDPI 2022-05-14 /pmc/articles/PMC9145339/ /pubmed/35632160 http://dx.doi.org/10.3390/s22103749 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Quiroz, Marco
Patiño, Raquel
Diaz-Amado, José
Cardinale, Yudith
Group Emotion Detection Based on Social Robot Perception
title Group Emotion Detection Based on Social Robot Perception
title_full Group Emotion Detection Based on Social Robot Perception
title_fullStr Group Emotion Detection Based on Social Robot Perception
title_full_unstemmed Group Emotion Detection Based on Social Robot Perception
title_short Group Emotion Detection Based on Social Robot Perception
title_sort group emotion detection based on social robot perception
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9145339/
https://www.ncbi.nlm.nih.gov/pubmed/35632160
http://dx.doi.org/10.3390/s22103749
work_keys_str_mv AT quirozmarco groupemotiondetectionbasedonsocialrobotperception
AT patinoraquel groupemotiondetectionbasedonsocialrobotperception
AT diazamadojose groupemotiondetectionbasedonsocialrobotperception
AT cardinaleyudith groupemotiondetectionbasedonsocialrobotperception