Cargando…

Detecting and Classifying Human Touches in a Social Robot Through Acoustic Sensing and Machine Learning

An important aspect in Human–Robot Interaction is responding to different kinds of touch stimuli. To date, several technologies have been explored to determine how a touch is perceived by a social robot, usually placing a large number of sensors throughout the robot’s shell. In this work, we introdu...

Descripción completa

Detalles Bibliográficos
Autores principales: Alonso-Martín, Fernando, Gamboa-Montero, Juan José, Castillo, José Carlos, Castro-González, Álvaro, Salichs, Miguel Ángel
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2017
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5470814/
https://www.ncbi.nlm.nih.gov/pubmed/28509865
http://dx.doi.org/10.3390/s17051138
_version_ 1783243829014429696
author Alonso-Martín, Fernando
Gamboa-Montero, Juan José
Castillo, José Carlos
Castro-González, Álvaro
Salichs, Miguel Ángel
author_facet Alonso-Martín, Fernando
Gamboa-Montero, Juan José
Castillo, José Carlos
Castro-González, Álvaro
Salichs, Miguel Ángel
author_sort Alonso-Martín, Fernando
collection PubMed
description An important aspect in Human–Robot Interaction is responding to different kinds of touch stimuli. To date, several technologies have been explored to determine how a touch is perceived by a social robot, usually placing a large number of sensors throughout the robot’s shell. In this work, we introduce a novel approach, where the audio acquired from contact microphones located in the robot’s shell is processed using machine learning techniques to distinguish between different types of touches. The system is able to determine when the robot is touched (touch detection), and to ascertain the kind of touch performed among a set of possibilities: stroke, tap, slap, and tickle (touch classification). This proposal is cost-effective since just a few microphones are able to cover the whole robot’s shell since a single microphone is enough to cover each solid part of the robot. Besides, it is easy to install and configure as it just requires a contact surface to attach the microphone to the robot’s shell and plug it into the robot’s computer. Results show the high accuracy scores in touch gesture recognition. The testing phase revealed that Logistic Model Trees achieved the best performance, with an F-score of 0.81. The dataset was built with information from 25 participants performing a total of 1981 touch gestures.
format Online
Article
Text
id pubmed-5470814
institution National Center for Biotechnology Information
language English
publishDate 2017
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-54708142017-06-16 Detecting and Classifying Human Touches in a Social Robot Through Acoustic Sensing and Machine Learning Alonso-Martín, Fernando Gamboa-Montero, Juan José Castillo, José Carlos Castro-González, Álvaro Salichs, Miguel Ángel Sensors (Basel) Article An important aspect in Human–Robot Interaction is responding to different kinds of touch stimuli. To date, several technologies have been explored to determine how a touch is perceived by a social robot, usually placing a large number of sensors throughout the robot’s shell. In this work, we introduce a novel approach, where the audio acquired from contact microphones located in the robot’s shell is processed using machine learning techniques to distinguish between different types of touches. The system is able to determine when the robot is touched (touch detection), and to ascertain the kind of touch performed among a set of possibilities: stroke, tap, slap, and tickle (touch classification). This proposal is cost-effective since just a few microphones are able to cover the whole robot’s shell since a single microphone is enough to cover each solid part of the robot. Besides, it is easy to install and configure as it just requires a contact surface to attach the microphone to the robot’s shell and plug it into the robot’s computer. Results show the high accuracy scores in touch gesture recognition. The testing phase revealed that Logistic Model Trees achieved the best performance, with an F-score of 0.81. The dataset was built with information from 25 participants performing a total of 1981 touch gestures. MDPI 2017-05-16 /pmc/articles/PMC5470814/ /pubmed/28509865 http://dx.doi.org/10.3390/s17051138 Text en © 2017 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Alonso-Martín, Fernando
Gamboa-Montero, Juan José
Castillo, José Carlos
Castro-González, Álvaro
Salichs, Miguel Ángel
Detecting and Classifying Human Touches in a Social Robot Through Acoustic Sensing and Machine Learning
title Detecting and Classifying Human Touches in a Social Robot Through Acoustic Sensing and Machine Learning
title_full Detecting and Classifying Human Touches in a Social Robot Through Acoustic Sensing and Machine Learning
title_fullStr Detecting and Classifying Human Touches in a Social Robot Through Acoustic Sensing and Machine Learning
title_full_unstemmed Detecting and Classifying Human Touches in a Social Robot Through Acoustic Sensing and Machine Learning
title_short Detecting and Classifying Human Touches in a Social Robot Through Acoustic Sensing and Machine Learning
title_sort detecting and classifying human touches in a social robot through acoustic sensing and machine learning
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5470814/
https://www.ncbi.nlm.nih.gov/pubmed/28509865
http://dx.doi.org/10.3390/s17051138
work_keys_str_mv AT alonsomartinfernando detectingandclassifyinghumantouchesinasocialrobotthroughacousticsensingandmachinelearning
AT gamboamonterojuanjose detectingandclassifyinghumantouchesinasocialrobotthroughacousticsensingandmachinelearning
AT castillojosecarlos detectingandclassifyinghumantouchesinasocialrobotthroughacousticsensingandmachinelearning
AT castrogonzalezalvaro detectingandclassifyinghumantouchesinasocialrobotthroughacousticsensingandmachinelearning
AT salichsmiguelangel detectingandclassifyinghumantouchesinasocialrobotthroughacousticsensingandmachinelearning