Cargando…

A novel feature fusion network for multimodal emotion recognition from EEG and eye movement signals

Emotion recognition is a challenging task, and the use of multimodal fusion methods for emotion recognition has become a trend. Fusion vectors can provide a more comprehensive representation of changes in the subject's emotional state, leading to more accurate emotion recognition results. Diffe...

Descripción completa

Detalles Bibliográficos
Autores principales: Fu, Baole, Gu, Chunrui, Fu, Ming, Xia, Yuxiao, Liu, Yinhua
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10436100/
https://www.ncbi.nlm.nih.gov/pubmed/37600016
http://dx.doi.org/10.3389/fnins.2023.1234162
_version_ 1785092251984068608
author Fu, Baole
Gu, Chunrui
Fu, Ming
Xia, Yuxiao
Liu, Yinhua
author_facet Fu, Baole
Gu, Chunrui
Fu, Ming
Xia, Yuxiao
Liu, Yinhua
author_sort Fu, Baole
collection PubMed
description Emotion recognition is a challenging task, and the use of multimodal fusion methods for emotion recognition has become a trend. Fusion vectors can provide a more comprehensive representation of changes in the subject's emotional state, leading to more accurate emotion recognition results. Different fusion inputs or feature fusion methods have varying effects on the final fusion outcome. In this paper, we propose a novel Multimodal Feature Fusion Neural Network model (MFFNN) that effectively extracts complementary information from eye movement signals and performs feature fusion with EEG signals. We construct a dual-branch feature extraction module to extract features from both modalities while ensuring temporal alignment. A multi-scale feature fusion module is introduced, which utilizes cross-channel soft attention to adaptively select information from different spatial scales, enabling the acquisition of features at different spatial scales for effective fusion. We conduct experiments on the publicly available SEED-IV dataset, and our model achieves an accuracy of 87.32% in recognizing four emotions (happiness, sadness, fear, and neutrality). The results demonstrate that the proposed model can better explore complementary information from EEG and eye movement signals, thereby improving accuracy, and stability in emotion recognition.
format Online
Article
Text
id pubmed-10436100
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-104361002023-08-19 A novel feature fusion network for multimodal emotion recognition from EEG and eye movement signals Fu, Baole Gu, Chunrui Fu, Ming Xia, Yuxiao Liu, Yinhua Front Neurosci Neuroscience Emotion recognition is a challenging task, and the use of multimodal fusion methods for emotion recognition has become a trend. Fusion vectors can provide a more comprehensive representation of changes in the subject's emotional state, leading to more accurate emotion recognition results. Different fusion inputs or feature fusion methods have varying effects on the final fusion outcome. In this paper, we propose a novel Multimodal Feature Fusion Neural Network model (MFFNN) that effectively extracts complementary information from eye movement signals and performs feature fusion with EEG signals. We construct a dual-branch feature extraction module to extract features from both modalities while ensuring temporal alignment. A multi-scale feature fusion module is introduced, which utilizes cross-channel soft attention to adaptively select information from different spatial scales, enabling the acquisition of features at different spatial scales for effective fusion. We conduct experiments on the publicly available SEED-IV dataset, and our model achieves an accuracy of 87.32% in recognizing four emotions (happiness, sadness, fear, and neutrality). The results demonstrate that the proposed model can better explore complementary information from EEG and eye movement signals, thereby improving accuracy, and stability in emotion recognition. Frontiers Media S.A. 2023-08-03 /pmc/articles/PMC10436100/ /pubmed/37600016 http://dx.doi.org/10.3389/fnins.2023.1234162 Text en Copyright © 2023 Fu, Gu, Fu, Xia and Liu. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Neuroscience
Fu, Baole
Gu, Chunrui
Fu, Ming
Xia, Yuxiao
Liu, Yinhua
A novel feature fusion network for multimodal emotion recognition from EEG and eye movement signals
title A novel feature fusion network for multimodal emotion recognition from EEG and eye movement signals
title_full A novel feature fusion network for multimodal emotion recognition from EEG and eye movement signals
title_fullStr A novel feature fusion network for multimodal emotion recognition from EEG and eye movement signals
title_full_unstemmed A novel feature fusion network for multimodal emotion recognition from EEG and eye movement signals
title_short A novel feature fusion network for multimodal emotion recognition from EEG and eye movement signals
title_sort novel feature fusion network for multimodal emotion recognition from eeg and eye movement signals
topic Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10436100/
https://www.ncbi.nlm.nih.gov/pubmed/37600016
http://dx.doi.org/10.3389/fnins.2023.1234162
work_keys_str_mv AT fubaole anovelfeaturefusionnetworkformultimodalemotionrecognitionfromeegandeyemovementsignals
AT guchunrui anovelfeaturefusionnetworkformultimodalemotionrecognitionfromeegandeyemovementsignals
AT fuming anovelfeaturefusionnetworkformultimodalemotionrecognitionfromeegandeyemovementsignals
AT xiayuxiao anovelfeaturefusionnetworkformultimodalemotionrecognitionfromeegandeyemovementsignals
AT liuyinhua anovelfeaturefusionnetworkformultimodalemotionrecognitionfromeegandeyemovementsignals
AT fubaole novelfeaturefusionnetworkformultimodalemotionrecognitionfromeegandeyemovementsignals
AT guchunrui novelfeaturefusionnetworkformultimodalemotionrecognitionfromeegandeyemovementsignals
AT fuming novelfeaturefusionnetworkformultimodalemotionrecognitionfromeegandeyemovementsignals
AT xiayuxiao novelfeaturefusionnetworkformultimodalemotionrecognitionfromeegandeyemovementsignals
AT liuyinhua novelfeaturefusionnetworkformultimodalemotionrecognitionfromeegandeyemovementsignals