Cargando…
Dual attentive fusion for EEG-based brain-computer interfaces
The classification based on Electroencephalogram (EEG) is a challenging task in the brain-computer interface (BCI) field due to data with a low signal-to-noise ratio. Most current deep learning based studies in this challenge focus on designing a desired convolutional neural network (CNN) to learn a...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Frontiers Media S.A.
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9727253/ https://www.ncbi.nlm.nih.gov/pubmed/36507325 http://dx.doi.org/10.3389/fnins.2022.1044631 |
_version_ | 1784844972314329088 |
---|---|
author | Du, Yuanhua Huang, Jian Huang, Xiuyu Shi, Kaibo Zhou, Nan |
author_facet | Du, Yuanhua Huang, Jian Huang, Xiuyu Shi, Kaibo Zhou, Nan |
author_sort | Du, Yuanhua |
collection | PubMed |
description | The classification based on Electroencephalogram (EEG) is a challenging task in the brain-computer interface (BCI) field due to data with a low signal-to-noise ratio. Most current deep learning based studies in this challenge focus on designing a desired convolutional neural network (CNN) to learn and classify the raw EEG signals. However, only CNN itself may not capture the highly discriminative patterns of EEG due to a lack of exploration of attentive spatial and temporal dynamics. To improve information utilization, this study proposes a Dual Attentive Fusion Model (DAFM) for the EEG-based BCI. DAFM is employed to capture the spatial and temporal information by modeling the interdependencies between the features from the EEG signals. To our best knowledge, our method is the first to fuse spatial and temporal dimensions in an interactive attention module. This module improves the expression ability of the extracted features. Extensive experiments implemented on four publicly available datasets demonstrate that our method outperforms state-of-the-art methods. Meanwhile, this work also indicates the effectiveness of Dual Attentive Fusion Module. |
format | Online Article Text |
id | pubmed-9727253 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Frontiers Media S.A. |
record_format | MEDLINE/PubMed |
spelling | pubmed-97272532022-12-08 Dual attentive fusion for EEG-based brain-computer interfaces Du, Yuanhua Huang, Jian Huang, Xiuyu Shi, Kaibo Zhou, Nan Front Neurosci Neuroscience The classification based on Electroencephalogram (EEG) is a challenging task in the brain-computer interface (BCI) field due to data with a low signal-to-noise ratio. Most current deep learning based studies in this challenge focus on designing a desired convolutional neural network (CNN) to learn and classify the raw EEG signals. However, only CNN itself may not capture the highly discriminative patterns of EEG due to a lack of exploration of attentive spatial and temporal dynamics. To improve information utilization, this study proposes a Dual Attentive Fusion Model (DAFM) for the EEG-based BCI. DAFM is employed to capture the spatial and temporal information by modeling the interdependencies between the features from the EEG signals. To our best knowledge, our method is the first to fuse spatial and temporal dimensions in an interactive attention module. This module improves the expression ability of the extracted features. Extensive experiments implemented on four publicly available datasets demonstrate that our method outperforms state-of-the-art methods. Meanwhile, this work also indicates the effectiveness of Dual Attentive Fusion Module. Frontiers Media S.A. 2022-11-23 /pmc/articles/PMC9727253/ /pubmed/36507325 http://dx.doi.org/10.3389/fnins.2022.1044631 Text en Copyright © 2022 Du, Huang, Huang, Shi and Zhou. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. |
spellingShingle | Neuroscience Du, Yuanhua Huang, Jian Huang, Xiuyu Shi, Kaibo Zhou, Nan Dual attentive fusion for EEG-based brain-computer interfaces |
title | Dual attentive fusion for EEG-based brain-computer interfaces |
title_full | Dual attentive fusion for EEG-based brain-computer interfaces |
title_fullStr | Dual attentive fusion for EEG-based brain-computer interfaces |
title_full_unstemmed | Dual attentive fusion for EEG-based brain-computer interfaces |
title_short | Dual attentive fusion for EEG-based brain-computer interfaces |
title_sort | dual attentive fusion for eeg-based brain-computer interfaces |
topic | Neuroscience |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9727253/ https://www.ncbi.nlm.nih.gov/pubmed/36507325 http://dx.doi.org/10.3389/fnins.2022.1044631 |
work_keys_str_mv | AT duyuanhua dualattentivefusionforeegbasedbraincomputerinterfaces AT huangjian dualattentivefusionforeegbasedbraincomputerinterfaces AT huangxiuyu dualattentivefusionforeegbasedbraincomputerinterfaces AT shikaibo dualattentivefusionforeegbasedbraincomputerinterfaces AT zhounan dualattentivefusionforeegbasedbraincomputerinterfaces |