Cargando…

Multi-method Fusion of Cross-Subject Emotion Recognition Based on High-Dimensional EEG Features

Emotion recognition using electroencephalogram (EEG) signals has attracted significant research attention. However, it is difficult to improve the emotional recognition effect across subjects. In response to this difficulty, in this study, multiple features were extracted for the formation of high-d...

Descripción completa

Detalles Bibliográficos
Autores principales: Yang, Fu, Zhao, Xingcong, Jiang, Wenge, Gao, Pengfei, Liu, Guangyuan
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2019
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6714862/
https://www.ncbi.nlm.nih.gov/pubmed/31507396
http://dx.doi.org/10.3389/fncom.2019.00053
_version_ 1783447134360567808
author Yang, Fu
Zhao, Xingcong
Jiang, Wenge
Gao, Pengfei
Liu, Guangyuan
author_facet Yang, Fu
Zhao, Xingcong
Jiang, Wenge
Gao, Pengfei
Liu, Guangyuan
author_sort Yang, Fu
collection PubMed
description Emotion recognition using electroencephalogram (EEG) signals has attracted significant research attention. However, it is difficult to improve the emotional recognition effect across subjects. In response to this difficulty, in this study, multiple features were extracted for the formation of high-dimensional features. Based on the high-dimensional features, an effective method for cross-subject emotion recognition was then developed, which integrated the significance test/sequential backward selection and the support vector machine (ST-SBSSVM). The effectiveness of the ST-SBSSVM was validated on a dataset for emotion analysis using physiological signals (DEAP) and the SJTU Emotion EEG Dataset (SEED). With respect to high-dimensional features, the ST-SBSSVM average improved the accuracy of cross-subject emotion recognition by 12.4% on the DEAP and 26.5% on the SEED when compared with common emotion recognition methods. The recognition accuracy obtained using ST-SBSSVM was as high as that obtained using sequential backward selection (SBS) on the DEAP dataset. However, on the SEED dataset, the recognition accuracy increased by ~6% using ST-SBSSVM from that using the SBS. Using the ST-SBSSVM, ~97% (DEAP) and 91% (SEED) of the program runtime was eliminated when compared with the SBS. Compared with recent similar works, the method developed in this study for emotion recognition across all subjects was found to be effective, and its accuracy was 72% (DEAP) and 89% (SEED).
format Online
Article
Text
id pubmed-6714862
institution National Center for Biotechnology Information
language English
publishDate 2019
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-67148622019-09-10 Multi-method Fusion of Cross-Subject Emotion Recognition Based on High-Dimensional EEG Features Yang, Fu Zhao, Xingcong Jiang, Wenge Gao, Pengfei Liu, Guangyuan Front Comput Neurosci Neuroscience Emotion recognition using electroencephalogram (EEG) signals has attracted significant research attention. However, it is difficult to improve the emotional recognition effect across subjects. In response to this difficulty, in this study, multiple features were extracted for the formation of high-dimensional features. Based on the high-dimensional features, an effective method for cross-subject emotion recognition was then developed, which integrated the significance test/sequential backward selection and the support vector machine (ST-SBSSVM). The effectiveness of the ST-SBSSVM was validated on a dataset for emotion analysis using physiological signals (DEAP) and the SJTU Emotion EEG Dataset (SEED). With respect to high-dimensional features, the ST-SBSSVM average improved the accuracy of cross-subject emotion recognition by 12.4% on the DEAP and 26.5% on the SEED when compared with common emotion recognition methods. The recognition accuracy obtained using ST-SBSSVM was as high as that obtained using sequential backward selection (SBS) on the DEAP dataset. However, on the SEED dataset, the recognition accuracy increased by ~6% using ST-SBSSVM from that using the SBS. Using the ST-SBSSVM, ~97% (DEAP) and 91% (SEED) of the program runtime was eliminated when compared with the SBS. Compared with recent similar works, the method developed in this study for emotion recognition across all subjects was found to be effective, and its accuracy was 72% (DEAP) and 89% (SEED). Frontiers Media S.A. 2019-08-20 /pmc/articles/PMC6714862/ /pubmed/31507396 http://dx.doi.org/10.3389/fncom.2019.00053 Text en Copyright © 2019 Yang, Zhao, Jiang, Gao and Liu. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Neuroscience
Yang, Fu
Zhao, Xingcong
Jiang, Wenge
Gao, Pengfei
Liu, Guangyuan
Multi-method Fusion of Cross-Subject Emotion Recognition Based on High-Dimensional EEG Features
title Multi-method Fusion of Cross-Subject Emotion Recognition Based on High-Dimensional EEG Features
title_full Multi-method Fusion of Cross-Subject Emotion Recognition Based on High-Dimensional EEG Features
title_fullStr Multi-method Fusion of Cross-Subject Emotion Recognition Based on High-Dimensional EEG Features
title_full_unstemmed Multi-method Fusion of Cross-Subject Emotion Recognition Based on High-Dimensional EEG Features
title_short Multi-method Fusion of Cross-Subject Emotion Recognition Based on High-Dimensional EEG Features
title_sort multi-method fusion of cross-subject emotion recognition based on high-dimensional eeg features
topic Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6714862/
https://www.ncbi.nlm.nih.gov/pubmed/31507396
http://dx.doi.org/10.3389/fncom.2019.00053
work_keys_str_mv AT yangfu multimethodfusionofcrosssubjectemotionrecognitionbasedonhighdimensionaleegfeatures
AT zhaoxingcong multimethodfusionofcrosssubjectemotionrecognitionbasedonhighdimensionaleegfeatures
AT jiangwenge multimethodfusionofcrosssubjectemotionrecognitionbasedonhighdimensionaleegfeatures
AT gaopengfei multimethodfusionofcrosssubjectemotionrecognitionbasedonhighdimensionaleegfeatures
AT liuguangyuan multimethodfusionofcrosssubjectemotionrecognitionbasedonhighdimensionaleegfeatures