Cargando…

Impact of dataset size and long-term ECoG-based BCI usage on deep learning decoders performance

INTRODUCTION: In brain-computer interfaces (BCI) research, recording data is time-consuming and expensive, which limits access to big datasets. This may influence the BCI system performance as machine learning methods depend strongly on the training dataset size. Important questions arise: taking in...

Descripción completa

Detalles Bibliográficos
Autores principales: Śliwowski, Maciej, Martin, Matthieu, Souloumiac, Antoine, Blanchart, Pierre, Aksenova, Tetiana
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10061076/
https://www.ncbi.nlm.nih.gov/pubmed/37007675
http://dx.doi.org/10.3389/fnhum.2023.1111645
_version_ 1785017219618439168
author Śliwowski, Maciej
Martin, Matthieu
Souloumiac, Antoine
Blanchart, Pierre
Aksenova, Tetiana
author_facet Śliwowski, Maciej
Martin, Matthieu
Souloumiac, Antoine
Blanchart, Pierre
Aksenova, Tetiana
author_sort Śliwowski, Maciej
collection PubMed
description INTRODUCTION: In brain-computer interfaces (BCI) research, recording data is time-consuming and expensive, which limits access to big datasets. This may influence the BCI system performance as machine learning methods depend strongly on the training dataset size. Important questions arise: taking into account neuronal signal characteristics (e.g., non-stationarity), can we achieve higher decoding performance with more data to train decoders? What is the perspective for further improvement with time in the case of long-term BCI studies? In this study, we investigated the impact of long-term recordings on motor imagery decoding from two main perspectives: model requirements regarding dataset size and potential for patient adaptation. METHODS: We evaluated the multilinear model and two deep learning (DL) models on a long-term BCI & Tetraplegia (ClinicalTrials.gov identifier: NCT02550522) clinical trial dataset containing 43 sessions of ECoG recordings performed with a tetraplegic patient. In the experiment, a participant executed 3D virtual hand translation using motor imagery patterns. We designed multiple computational experiments in which training datasets were increased or translated to investigate the relationship between models' performance and different factors influencing recordings. RESULTS: Our results showed that DL decoders showed similar requirements regarding the dataset size compared to the multilinear model while demonstrating higher decoding performance. Moreover, high decoding performance was obtained with relatively small datasets recorded later in the experiment, suggesting motor imagery patterns improvement and patient adaptation during the long-term experiment. Finally, we proposed UMAP embeddings and local intrinsic dimensionality as a way to visualize the data and potentially evaluate data quality. DISCUSSION: DL-based decoding is a prospective approach in BCI which may be efficiently applied with real-life dataset size. Patient-decoder co-adaptation is an important factor to consider in long-term clinical BCI.
format Online
Article
Text
id pubmed-10061076
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-100610762023-03-31 Impact of dataset size and long-term ECoG-based BCI usage on deep learning decoders performance Śliwowski, Maciej Martin, Matthieu Souloumiac, Antoine Blanchart, Pierre Aksenova, Tetiana Front Hum Neurosci Human Neuroscience INTRODUCTION: In brain-computer interfaces (BCI) research, recording data is time-consuming and expensive, which limits access to big datasets. This may influence the BCI system performance as machine learning methods depend strongly on the training dataset size. Important questions arise: taking into account neuronal signal characteristics (e.g., non-stationarity), can we achieve higher decoding performance with more data to train decoders? What is the perspective for further improvement with time in the case of long-term BCI studies? In this study, we investigated the impact of long-term recordings on motor imagery decoding from two main perspectives: model requirements regarding dataset size and potential for patient adaptation. METHODS: We evaluated the multilinear model and two deep learning (DL) models on a long-term BCI & Tetraplegia (ClinicalTrials.gov identifier: NCT02550522) clinical trial dataset containing 43 sessions of ECoG recordings performed with a tetraplegic patient. In the experiment, a participant executed 3D virtual hand translation using motor imagery patterns. We designed multiple computational experiments in which training datasets were increased or translated to investigate the relationship between models' performance and different factors influencing recordings. RESULTS: Our results showed that DL decoders showed similar requirements regarding the dataset size compared to the multilinear model while demonstrating higher decoding performance. Moreover, high decoding performance was obtained with relatively small datasets recorded later in the experiment, suggesting motor imagery patterns improvement and patient adaptation during the long-term experiment. Finally, we proposed UMAP embeddings and local intrinsic dimensionality as a way to visualize the data and potentially evaluate data quality. DISCUSSION: DL-based decoding is a prospective approach in BCI which may be efficiently applied with real-life dataset size. Patient-decoder co-adaptation is an important factor to consider in long-term clinical BCI. Frontiers Media S.A. 2023-03-16 /pmc/articles/PMC10061076/ /pubmed/37007675 http://dx.doi.org/10.3389/fnhum.2023.1111645 Text en Copyright © 2023 Śliwowski, Martin, Souloumiac, Blanchart and Aksenova. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Human Neuroscience
Śliwowski, Maciej
Martin, Matthieu
Souloumiac, Antoine
Blanchart, Pierre
Aksenova, Tetiana
Impact of dataset size and long-term ECoG-based BCI usage on deep learning decoders performance
title Impact of dataset size and long-term ECoG-based BCI usage on deep learning decoders performance
title_full Impact of dataset size and long-term ECoG-based BCI usage on deep learning decoders performance
title_fullStr Impact of dataset size and long-term ECoG-based BCI usage on deep learning decoders performance
title_full_unstemmed Impact of dataset size and long-term ECoG-based BCI usage on deep learning decoders performance
title_short Impact of dataset size and long-term ECoG-based BCI usage on deep learning decoders performance
title_sort impact of dataset size and long-term ecog-based bci usage on deep learning decoders performance
topic Human Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10061076/
https://www.ncbi.nlm.nih.gov/pubmed/37007675
http://dx.doi.org/10.3389/fnhum.2023.1111645
work_keys_str_mv AT sliwowskimaciej impactofdatasetsizeandlongtermecogbasedbciusageondeeplearningdecodersperformance
AT martinmatthieu impactofdatasetsizeandlongtermecogbasedbciusageondeeplearningdecodersperformance
AT souloumiacantoine impactofdatasetsizeandlongtermecogbasedbciusageondeeplearningdecodersperformance
AT blanchartpierre impactofdatasetsizeandlongtermecogbasedbciusageondeeplearningdecodersperformance
AT aksenovatetiana impactofdatasetsizeandlongtermecogbasedbciusageondeeplearningdecodersperformance