Cargando…

Multimodal Autoencoder Predicts fNIRS Resting State From EEG Signals

In this work, we introduce a deep learning architecture for evaluation on multimodal electroencephalographic (EEG) and functional near-infrared spectroscopy (fNIRS) recordings from 40 epileptic patients. Long short-term memory units and convolutional neural networks are integrated within a multimoda...

Descripción completa

Detalles Bibliográficos
Autores principales: Sirpal, Parikshat, Damseh, Rafat, Peng, Ke, Nguyen, Dang Khoa, Lesage, Frédéric
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Springer US 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9547786/
https://www.ncbi.nlm.nih.gov/pubmed/34378155
http://dx.doi.org/10.1007/s12021-021-09538-3
_version_ 1784805340513042432
author Sirpal, Parikshat
Damseh, Rafat
Peng, Ke
Nguyen, Dang Khoa
Lesage, Frédéric
author_facet Sirpal, Parikshat
Damseh, Rafat
Peng, Ke
Nguyen, Dang Khoa
Lesage, Frédéric
author_sort Sirpal, Parikshat
collection PubMed
description In this work, we introduce a deep learning architecture for evaluation on multimodal electroencephalographic (EEG) and functional near-infrared spectroscopy (fNIRS) recordings from 40 epileptic patients. Long short-term memory units and convolutional neural networks are integrated within a multimodal sequence-to-sequence autoencoder. The trained neural network predicts fNIRS signals from EEG, sans a priori, by hierarchically extracting deep features from EEG full spectra and specific EEG frequency bands. Results show that higher frequency EEG ranges are predictive of fNIRS signals with the gamma band inputs dominating fNIRS prediction as compared to other frequency envelopes. Seed based functional connectivity validates similar patterns between experimental fNIRS and our model’s fNIRS reconstructions. This is the first study that shows it is possible to predict brain hemodynamics (fNIRS) from encoded neural data (EEG) in the resting human epileptic brain based on power spectrum amplitude modulation of frequency oscillations in the context of specific hypotheses about how EEG frequency bands decode fNIRS signals.
format Online
Article
Text
id pubmed-9547786
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher Springer US
record_format MEDLINE/PubMed
spelling pubmed-95477862022-10-10 Multimodal Autoencoder Predicts fNIRS Resting State From EEG Signals Sirpal, Parikshat Damseh, Rafat Peng, Ke Nguyen, Dang Khoa Lesage, Frédéric Neuroinformatics Original Article In this work, we introduce a deep learning architecture for evaluation on multimodal electroencephalographic (EEG) and functional near-infrared spectroscopy (fNIRS) recordings from 40 epileptic patients. Long short-term memory units and convolutional neural networks are integrated within a multimodal sequence-to-sequence autoencoder. The trained neural network predicts fNIRS signals from EEG, sans a priori, by hierarchically extracting deep features from EEG full spectra and specific EEG frequency bands. Results show that higher frequency EEG ranges are predictive of fNIRS signals with the gamma band inputs dominating fNIRS prediction as compared to other frequency envelopes. Seed based functional connectivity validates similar patterns between experimental fNIRS and our model’s fNIRS reconstructions. This is the first study that shows it is possible to predict brain hemodynamics (fNIRS) from encoded neural data (EEG) in the resting human epileptic brain based on power spectrum amplitude modulation of frequency oscillations in the context of specific hypotheses about how EEG frequency bands decode fNIRS signals. Springer US 2021-08-10 2022 /pmc/articles/PMC9547786/ /pubmed/34378155 http://dx.doi.org/10.1007/s12021-021-09538-3 Text en © The Author(s) 2021 https://creativecommons.org/licenses/by/4.0/Open AccessThis article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) .
spellingShingle Original Article
Sirpal, Parikshat
Damseh, Rafat
Peng, Ke
Nguyen, Dang Khoa
Lesage, Frédéric
Multimodal Autoencoder Predicts fNIRS Resting State From EEG Signals
title Multimodal Autoencoder Predicts fNIRS Resting State From EEG Signals
title_full Multimodal Autoencoder Predicts fNIRS Resting State From EEG Signals
title_fullStr Multimodal Autoencoder Predicts fNIRS Resting State From EEG Signals
title_full_unstemmed Multimodal Autoencoder Predicts fNIRS Resting State From EEG Signals
title_short Multimodal Autoencoder Predicts fNIRS Resting State From EEG Signals
title_sort multimodal autoencoder predicts fnirs resting state from eeg signals
topic Original Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9547786/
https://www.ncbi.nlm.nih.gov/pubmed/34378155
http://dx.doi.org/10.1007/s12021-021-09538-3
work_keys_str_mv AT sirpalparikshat multimodalautoencoderpredictsfnirsrestingstatefromeegsignals
AT damsehrafat multimodalautoencoderpredictsfnirsrestingstatefromeegsignals
AT pengke multimodalautoencoderpredictsfnirsrestingstatefromeegsignals
AT nguyendangkhoa multimodalautoencoderpredictsfnirsrestingstatefromeegsignals
AT lesagefrederic multimodalautoencoderpredictsfnirsrestingstatefromeegsignals