Cargando…

Developing crossmodal expression recognition based on a deep neural model

A robot capable of understanding emotion expressions can increase its own capability of solving problems by using emotion expressions as part of its own decision-making, in a similar way to humans. Evidence shows that the perception of human interaction starts with an innate perception mechanism, wh...

Descripción completa

Detalles Bibliográficos
Autores principales: Barros, Pablo, Wermter, Stefan
Formato: Online Artículo Texto
Lenguaje:English
Publicado: SAGE Publications 2016
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5098700/
https://www.ncbi.nlm.nih.gov/pubmed/27853349
http://dx.doi.org/10.1177/1059712316664017
_version_ 1782465812484849664
author Barros, Pablo
Wermter, Stefan
author_facet Barros, Pablo
Wermter, Stefan
author_sort Barros, Pablo
collection PubMed
description A robot capable of understanding emotion expressions can increase its own capability of solving problems by using emotion expressions as part of its own decision-making, in a similar way to humans. Evidence shows that the perception of human interaction starts with an innate perception mechanism, where the interaction between different entities is perceived and categorized into two very clear directions: positive or negative. While the person is developing during childhood, the perception evolves and is shaped based on the observation of human interaction, creating the capability to learn different categories of expressions. In the context of human–robot interaction, we propose a model that simulates the innate perception of audio–visual emotion expressions with deep neural networks, that learns new expressions by categorizing them into emotional clusters with a self-organizing layer. The proposed model is evaluated with three different corpora: The Surrey Audio–Visual Expressed Emotion (SAVEE) database, the visual Bi-modal Face and Body benchmark (FABO) database, and the multimodal corpus of the Emotion Recognition in the Wild (EmotiW) challenge. We use these corpora to evaluate the performance of the model to recognize emotional expressions, and compare it to state-of-the-art research.
format Online
Article
Text
id pubmed-5098700
institution National Center for Biotechnology Information
language English
publishDate 2016
publisher SAGE Publications
record_format MEDLINE/PubMed
spelling pubmed-50987002016-11-14 Developing crossmodal expression recognition based on a deep neural model Barros, Pablo Wermter, Stefan Adapt Behav Special Issue Articles A robot capable of understanding emotion expressions can increase its own capability of solving problems by using emotion expressions as part of its own decision-making, in a similar way to humans. Evidence shows that the perception of human interaction starts with an innate perception mechanism, where the interaction between different entities is perceived and categorized into two very clear directions: positive or negative. While the person is developing during childhood, the perception evolves and is shaped based on the observation of human interaction, creating the capability to learn different categories of expressions. In the context of human–robot interaction, we propose a model that simulates the innate perception of audio–visual emotion expressions with deep neural networks, that learns new expressions by categorizing them into emotional clusters with a self-organizing layer. The proposed model is evaluated with three different corpora: The Surrey Audio–Visual Expressed Emotion (SAVEE) database, the visual Bi-modal Face and Body benchmark (FABO) database, and the multimodal corpus of the Emotion Recognition in the Wild (EmotiW) challenge. We use these corpora to evaluate the performance of the model to recognize emotional expressions, and compare it to state-of-the-art research. SAGE Publications 2016-10-10 2016-10 /pmc/articles/PMC5098700/ /pubmed/27853349 http://dx.doi.org/10.1177/1059712316664017 Text en © The Author(s) 2016 http://creativecommons.org/licenses/by-nc/3.0/ This article is distributed under the terms of the Creative Commons Attribution-NonCommercial 3.0 License (http://www.creativecommons.org/licenses/by-nc/3.0/) which permits non-commercial use, reproduction and distribution of the work without further permission provided the original work is attributed as specified on the SAGE and Open Access page (https://us.sagepub.com/en-us/nam/open-access-at-sage).
spellingShingle Special Issue Articles
Barros, Pablo
Wermter, Stefan
Developing crossmodal expression recognition based on a deep neural model
title Developing crossmodal expression recognition based on a deep neural model
title_full Developing crossmodal expression recognition based on a deep neural model
title_fullStr Developing crossmodal expression recognition based on a deep neural model
title_full_unstemmed Developing crossmodal expression recognition based on a deep neural model
title_short Developing crossmodal expression recognition based on a deep neural model
title_sort developing crossmodal expression recognition based on a deep neural model
topic Special Issue Articles
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5098700/
https://www.ncbi.nlm.nih.gov/pubmed/27853349
http://dx.doi.org/10.1177/1059712316664017
work_keys_str_mv AT barrospablo developingcrossmodalexpressionrecognitionbasedonadeepneuralmodel
AT wermterstefan developingcrossmodalexpressionrecognitionbasedonadeepneuralmodel