Cargando…

SAE+LSTM: A New Framework for Emotion Recognition From Multi-Channel EEG

EEG-based automatic emotion recognition can help brain-inspired robots in improving their interactions with humans. This paper presents a novel framework for emotion recognition using multi-channel electroencephalogram (EEG). The framework consists of a linear EEG mixing model and an emotion timing...

Descripción completa

Detalles Bibliográficos
Autores principales: Xing, Xiaofen, Li, Zhenqi, Xu, Tianyuan, Shu, Lin, Hu, Bin, Xu, Xiangmin
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2019
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6581731/
https://www.ncbi.nlm.nih.gov/pubmed/31244638
http://dx.doi.org/10.3389/fnbot.2019.00037
_version_ 1783428200368439296
author Xing, Xiaofen
Li, Zhenqi
Xu, Tianyuan
Shu, Lin
Hu, Bin
Xu, Xiangmin
author_facet Xing, Xiaofen
Li, Zhenqi
Xu, Tianyuan
Shu, Lin
Hu, Bin
Xu, Xiangmin
author_sort Xing, Xiaofen
collection PubMed
description EEG-based automatic emotion recognition can help brain-inspired robots in improving their interactions with humans. This paper presents a novel framework for emotion recognition using multi-channel electroencephalogram (EEG). The framework consists of a linear EEG mixing model and an emotion timing model. Our proposed framework considerably decomposes the EEG source signals from the collected EEG signals and improves classification accuracy by using the context correlations of the EEG feature sequences. Specially, Stack AutoEncoder (SAE) is used to build and solve the linear EEG mixing model and the emotion timing model is based on the Long Short-Term Memory Recurrent Neural Network (LSTM-RNN). The framework was implemented on the DEAP dataset for an emotion recognition experiment, where the mean accuracy of emotion recognition achieved 81.10% in valence and 74.38% in arousal, and the effectiveness of our framework was verified. Our framework exhibited a better performance in emotion recognition using multi-channel EEG than the compared conventional approaches in the experiments.
format Online
Article
Text
id pubmed-6581731
institution National Center for Biotechnology Information
language English
publishDate 2019
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-65817312019-06-26 SAE+LSTM: A New Framework for Emotion Recognition From Multi-Channel EEG Xing, Xiaofen Li, Zhenqi Xu, Tianyuan Shu, Lin Hu, Bin Xu, Xiangmin Front Neurorobot Robotics and AI EEG-based automatic emotion recognition can help brain-inspired robots in improving their interactions with humans. This paper presents a novel framework for emotion recognition using multi-channel electroencephalogram (EEG). The framework consists of a linear EEG mixing model and an emotion timing model. Our proposed framework considerably decomposes the EEG source signals from the collected EEG signals and improves classification accuracy by using the context correlations of the EEG feature sequences. Specially, Stack AutoEncoder (SAE) is used to build and solve the linear EEG mixing model and the emotion timing model is based on the Long Short-Term Memory Recurrent Neural Network (LSTM-RNN). The framework was implemented on the DEAP dataset for an emotion recognition experiment, where the mean accuracy of emotion recognition achieved 81.10% in valence and 74.38% in arousal, and the effectiveness of our framework was verified. Our framework exhibited a better performance in emotion recognition using multi-channel EEG than the compared conventional approaches in the experiments. Frontiers Media S.A. 2019-06-12 /pmc/articles/PMC6581731/ /pubmed/31244638 http://dx.doi.org/10.3389/fnbot.2019.00037 Text en Copyright © 2019 Xing, Li, Xu, Shu, Hu and Xu. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Robotics and AI
Xing, Xiaofen
Li, Zhenqi
Xu, Tianyuan
Shu, Lin
Hu, Bin
Xu, Xiangmin
SAE+LSTM: A New Framework for Emotion Recognition From Multi-Channel EEG
title SAE+LSTM: A New Framework for Emotion Recognition From Multi-Channel EEG
title_full SAE+LSTM: A New Framework for Emotion Recognition From Multi-Channel EEG
title_fullStr SAE+LSTM: A New Framework for Emotion Recognition From Multi-Channel EEG
title_full_unstemmed SAE+LSTM: A New Framework for Emotion Recognition From Multi-Channel EEG
title_short SAE+LSTM: A New Framework for Emotion Recognition From Multi-Channel EEG
title_sort sae+lstm: a new framework for emotion recognition from multi-channel eeg
topic Robotics and AI
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6581731/
https://www.ncbi.nlm.nih.gov/pubmed/31244638
http://dx.doi.org/10.3389/fnbot.2019.00037
work_keys_str_mv AT xingxiaofen saelstmanewframeworkforemotionrecognitionfrommultichanneleeg
AT lizhenqi saelstmanewframeworkforemotionrecognitionfrommultichanneleeg
AT xutianyuan saelstmanewframeworkforemotionrecognitionfrommultichanneleeg
AT shulin saelstmanewframeworkforemotionrecognitionfrommultichanneleeg
AT hubin saelstmanewframeworkforemotionrecognitionfrommultichanneleeg
AT xuxiangmin saelstmanewframeworkforemotionrecognitionfrommultichanneleeg