Cargando…
Facial Muscle Activity Recognition with Reconfigurable Differential Stethoscope-Microphones
Many human activities and states are related to the facial muscles’ actions: from the expression of emotions, stress, and non-verbal communication through health-related actions, such as coughing and sneezing to nutrition and drinking. In this work, we describe, in detail, the design and evaluation...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2020
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7506891/ https://www.ncbi.nlm.nih.gov/pubmed/32872633 http://dx.doi.org/10.3390/s20174904 |
_version_ | 1783585116576022528 |
---|---|
author | Bello, Hymalai Zhou, Bo Lukowicz, Paul |
author_facet | Bello, Hymalai Zhou, Bo Lukowicz, Paul |
author_sort | Bello, Hymalai |
collection | PubMed |
description | Many human activities and states are related to the facial muscles’ actions: from the expression of emotions, stress, and non-verbal communication through health-related actions, such as coughing and sneezing to nutrition and drinking. In this work, we describe, in detail, the design and evaluation of a wearable system for facial muscle activity monitoring based on a re-configurable differential array of stethoscope-microphones. In our system, six stethoscopes are placed at locations that could easily be integrated into the frame of smart glasses. The paper describes the detailed hardware design and selection and adaptation of appropriate signal processing and machine learning methods. For the evaluation, we asked eight participants to imitate a set of facial actions, such as expressions of happiness, anger, surprise, sadness, upset, and disgust, and gestures, like kissing, winkling, sticking the tongue out, and taking a pill. An evaluation of a complete data set of 2640 events with 66% training and a 33% testing rate has been performed. Although we encountered high variability of the volunteers’ expressions, our approach shows a recall = 55%, precision = 56%, and f1-score of 54% for the user-independent scenario(9% chance-level). On a user-dependent basis, our worst result has an f1-score = 60% and best result with f1-score = 89%. Having a recall [Formula: see text] for expressions like happiness, anger, kissing, sticking the tongue out, and neutral(Null-class). |
format | Online Article Text |
id | pubmed-7506891 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2020 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-75068912020-09-26 Facial Muscle Activity Recognition with Reconfigurable Differential Stethoscope-Microphones Bello, Hymalai Zhou, Bo Lukowicz, Paul Sensors (Basel) Article Many human activities and states are related to the facial muscles’ actions: from the expression of emotions, stress, and non-verbal communication through health-related actions, such as coughing and sneezing to nutrition and drinking. In this work, we describe, in detail, the design and evaluation of a wearable system for facial muscle activity monitoring based on a re-configurable differential array of stethoscope-microphones. In our system, six stethoscopes are placed at locations that could easily be integrated into the frame of smart glasses. The paper describes the detailed hardware design and selection and adaptation of appropriate signal processing and machine learning methods. For the evaluation, we asked eight participants to imitate a set of facial actions, such as expressions of happiness, anger, surprise, sadness, upset, and disgust, and gestures, like kissing, winkling, sticking the tongue out, and taking a pill. An evaluation of a complete data set of 2640 events with 66% training and a 33% testing rate has been performed. Although we encountered high variability of the volunteers’ expressions, our approach shows a recall = 55%, precision = 56%, and f1-score of 54% for the user-independent scenario(9% chance-level). On a user-dependent basis, our worst result has an f1-score = 60% and best result with f1-score = 89%. Having a recall [Formula: see text] for expressions like happiness, anger, kissing, sticking the tongue out, and neutral(Null-class). MDPI 2020-08-30 /pmc/articles/PMC7506891/ /pubmed/32872633 http://dx.doi.org/10.3390/s20174904 Text en © 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Bello, Hymalai Zhou, Bo Lukowicz, Paul Facial Muscle Activity Recognition with Reconfigurable Differential Stethoscope-Microphones |
title | Facial Muscle Activity Recognition with Reconfigurable Differential Stethoscope-Microphones |
title_full | Facial Muscle Activity Recognition with Reconfigurable Differential Stethoscope-Microphones |
title_fullStr | Facial Muscle Activity Recognition with Reconfigurable Differential Stethoscope-Microphones |
title_full_unstemmed | Facial Muscle Activity Recognition with Reconfigurable Differential Stethoscope-Microphones |
title_short | Facial Muscle Activity Recognition with Reconfigurable Differential Stethoscope-Microphones |
title_sort | facial muscle activity recognition with reconfigurable differential stethoscope-microphones |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7506891/ https://www.ncbi.nlm.nih.gov/pubmed/32872633 http://dx.doi.org/10.3390/s20174904 |
work_keys_str_mv | AT bellohymalai facialmuscleactivityrecognitionwithreconfigurabledifferentialstethoscopemicrophones AT zhoubo facialmuscleactivityrecognitionwithreconfigurabledifferentialstethoscopemicrophones AT lukowiczpaul facialmuscleactivityrecognitionwithreconfigurabledifferentialstethoscopemicrophones |