Cargando…

Strength Is in Numbers: Can Concordant Artificial Listeners Improve Prediction of Emotion from Speech?

Humans can communicate their emotions by modulating facial expressions or the tone of their voice. Albeit numerous applications exist that enable machines to read facial emotions and recognize the content of verbal messages, methods for speech emotion recognition are still in their infancy. Yet, fas...

Descripción completa

Detalles Bibliográficos
Autores principales: Martinelli, Eugenio, Mencattini, Arianna, Daprati, Elena, Di Natale, Corrado
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Public Library of Science 2016
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5001724/
https://www.ncbi.nlm.nih.gov/pubmed/27563724
http://dx.doi.org/10.1371/journal.pone.0161752
_version_ 1782450472549875712
author Martinelli, Eugenio
Mencattini, Arianna
Daprati, Elena
Di Natale, Corrado
author_facet Martinelli, Eugenio
Mencattini, Arianna
Daprati, Elena
Di Natale, Corrado
author_sort Martinelli, Eugenio
collection PubMed
description Humans can communicate their emotions by modulating facial expressions or the tone of their voice. Albeit numerous applications exist that enable machines to read facial emotions and recognize the content of verbal messages, methods for speech emotion recognition are still in their infancy. Yet, fast and reliable applications for emotion recognition are the obvious advancement of present ‘intelligent personal assistants’, and may have countless applications in diagnostics, rehabilitation and research. Taking inspiration from the dynamics of human group decision-making, we devised a novel speech emotion recognition system that applies, for the first time, a semi-supervised prediction model based on consensus. Three tests were carried out to compare this algorithm with traditional approaches. Labeling performances relative to a public database of spontaneous speeches are reported. The novel system appears to be fast, robust and less computationally demanding than traditional methods, allowing for easier implementation in portable voice-analyzers (as used in rehabilitation, research, industry, etc.) and for applications in the research domain (such as real-time pairing of stimuli to participants’ emotional state, selective/differential data collection based on emotional content, etc.).
format Online
Article
Text
id pubmed-5001724
institution National Center for Biotechnology Information
language English
publishDate 2016
publisher Public Library of Science
record_format MEDLINE/PubMed
spelling pubmed-50017242016-09-12 Strength Is in Numbers: Can Concordant Artificial Listeners Improve Prediction of Emotion from Speech? Martinelli, Eugenio Mencattini, Arianna Daprati, Elena Di Natale, Corrado PLoS One Research Article Humans can communicate their emotions by modulating facial expressions or the tone of their voice. Albeit numerous applications exist that enable machines to read facial emotions and recognize the content of verbal messages, methods for speech emotion recognition are still in their infancy. Yet, fast and reliable applications for emotion recognition are the obvious advancement of present ‘intelligent personal assistants’, and may have countless applications in diagnostics, rehabilitation and research. Taking inspiration from the dynamics of human group decision-making, we devised a novel speech emotion recognition system that applies, for the first time, a semi-supervised prediction model based on consensus. Three tests were carried out to compare this algorithm with traditional approaches. Labeling performances relative to a public database of spontaneous speeches are reported. The novel system appears to be fast, robust and less computationally demanding than traditional methods, allowing for easier implementation in portable voice-analyzers (as used in rehabilitation, research, industry, etc.) and for applications in the research domain (such as real-time pairing of stimuli to participants’ emotional state, selective/differential data collection based on emotional content, etc.). Public Library of Science 2016-08-26 /pmc/articles/PMC5001724/ /pubmed/27563724 http://dx.doi.org/10.1371/journal.pone.0161752 Text en © 2016 Martinelli et al http://creativecommons.org/licenses/by/4.0/ This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
spellingShingle Research Article
Martinelli, Eugenio
Mencattini, Arianna
Daprati, Elena
Di Natale, Corrado
Strength Is in Numbers: Can Concordant Artificial Listeners Improve Prediction of Emotion from Speech?
title Strength Is in Numbers: Can Concordant Artificial Listeners Improve Prediction of Emotion from Speech?
title_full Strength Is in Numbers: Can Concordant Artificial Listeners Improve Prediction of Emotion from Speech?
title_fullStr Strength Is in Numbers: Can Concordant Artificial Listeners Improve Prediction of Emotion from Speech?
title_full_unstemmed Strength Is in Numbers: Can Concordant Artificial Listeners Improve Prediction of Emotion from Speech?
title_short Strength Is in Numbers: Can Concordant Artificial Listeners Improve Prediction of Emotion from Speech?
title_sort strength is in numbers: can concordant artificial listeners improve prediction of emotion from speech?
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5001724/
https://www.ncbi.nlm.nih.gov/pubmed/27563724
http://dx.doi.org/10.1371/journal.pone.0161752
work_keys_str_mv AT martinellieugenio strengthisinnumberscanconcordantartificiallistenersimprovepredictionofemotionfromspeech
AT mencattiniarianna strengthisinnumberscanconcordantartificiallistenersimprovepredictionofemotionfromspeech
AT dapratielena strengthisinnumberscanconcordantartificiallistenersimprovepredictionofemotionfromspeech
AT dinatalecorrado strengthisinnumberscanconcordantartificiallistenersimprovepredictionofemotionfromspeech