Cargando…
Impact of dataset size and long-term ECoG-based BCI usage on deep learning decoders performance
INTRODUCTION: In brain-computer interfaces (BCI) research, recording data is time-consuming and expensive, which limits access to big datasets. This may influence the BCI system performance as machine learning methods depend strongly on the training dataset size. Important questions arise: taking in...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Frontiers Media S.A.
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10061076/ https://www.ncbi.nlm.nih.gov/pubmed/37007675 http://dx.doi.org/10.3389/fnhum.2023.1111645 |
_version_ | 1785017219618439168 |
---|---|
author | Śliwowski, Maciej Martin, Matthieu Souloumiac, Antoine Blanchart, Pierre Aksenova, Tetiana |
author_facet | Śliwowski, Maciej Martin, Matthieu Souloumiac, Antoine Blanchart, Pierre Aksenova, Tetiana |
author_sort | Śliwowski, Maciej |
collection | PubMed |
description | INTRODUCTION: In brain-computer interfaces (BCI) research, recording data is time-consuming and expensive, which limits access to big datasets. This may influence the BCI system performance as machine learning methods depend strongly on the training dataset size. Important questions arise: taking into account neuronal signal characteristics (e.g., non-stationarity), can we achieve higher decoding performance with more data to train decoders? What is the perspective for further improvement with time in the case of long-term BCI studies? In this study, we investigated the impact of long-term recordings on motor imagery decoding from two main perspectives: model requirements regarding dataset size and potential for patient adaptation. METHODS: We evaluated the multilinear model and two deep learning (DL) models on a long-term BCI & Tetraplegia (ClinicalTrials.gov identifier: NCT02550522) clinical trial dataset containing 43 sessions of ECoG recordings performed with a tetraplegic patient. In the experiment, a participant executed 3D virtual hand translation using motor imagery patterns. We designed multiple computational experiments in which training datasets were increased or translated to investigate the relationship between models' performance and different factors influencing recordings. RESULTS: Our results showed that DL decoders showed similar requirements regarding the dataset size compared to the multilinear model while demonstrating higher decoding performance. Moreover, high decoding performance was obtained with relatively small datasets recorded later in the experiment, suggesting motor imagery patterns improvement and patient adaptation during the long-term experiment. Finally, we proposed UMAP embeddings and local intrinsic dimensionality as a way to visualize the data and potentially evaluate data quality. DISCUSSION: DL-based decoding is a prospective approach in BCI which may be efficiently applied with real-life dataset size. Patient-decoder co-adaptation is an important factor to consider in long-term clinical BCI. |
format | Online Article Text |
id | pubmed-10061076 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | Frontiers Media S.A. |
record_format | MEDLINE/PubMed |
spelling | pubmed-100610762023-03-31 Impact of dataset size and long-term ECoG-based BCI usage on deep learning decoders performance Śliwowski, Maciej Martin, Matthieu Souloumiac, Antoine Blanchart, Pierre Aksenova, Tetiana Front Hum Neurosci Human Neuroscience INTRODUCTION: In brain-computer interfaces (BCI) research, recording data is time-consuming and expensive, which limits access to big datasets. This may influence the BCI system performance as machine learning methods depend strongly on the training dataset size. Important questions arise: taking into account neuronal signal characteristics (e.g., non-stationarity), can we achieve higher decoding performance with more data to train decoders? What is the perspective for further improvement with time in the case of long-term BCI studies? In this study, we investigated the impact of long-term recordings on motor imagery decoding from two main perspectives: model requirements regarding dataset size and potential for patient adaptation. METHODS: We evaluated the multilinear model and two deep learning (DL) models on a long-term BCI & Tetraplegia (ClinicalTrials.gov identifier: NCT02550522) clinical trial dataset containing 43 sessions of ECoG recordings performed with a tetraplegic patient. In the experiment, a participant executed 3D virtual hand translation using motor imagery patterns. We designed multiple computational experiments in which training datasets were increased or translated to investigate the relationship between models' performance and different factors influencing recordings. RESULTS: Our results showed that DL decoders showed similar requirements regarding the dataset size compared to the multilinear model while demonstrating higher decoding performance. Moreover, high decoding performance was obtained with relatively small datasets recorded later in the experiment, suggesting motor imagery patterns improvement and patient adaptation during the long-term experiment. Finally, we proposed UMAP embeddings and local intrinsic dimensionality as a way to visualize the data and potentially evaluate data quality. DISCUSSION: DL-based decoding is a prospective approach in BCI which may be efficiently applied with real-life dataset size. Patient-decoder co-adaptation is an important factor to consider in long-term clinical BCI. Frontiers Media S.A. 2023-03-16 /pmc/articles/PMC10061076/ /pubmed/37007675 http://dx.doi.org/10.3389/fnhum.2023.1111645 Text en Copyright © 2023 Śliwowski, Martin, Souloumiac, Blanchart and Aksenova. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. |
spellingShingle | Human Neuroscience Śliwowski, Maciej Martin, Matthieu Souloumiac, Antoine Blanchart, Pierre Aksenova, Tetiana Impact of dataset size and long-term ECoG-based BCI usage on deep learning decoders performance |
title | Impact of dataset size and long-term ECoG-based BCI usage on deep learning decoders performance |
title_full | Impact of dataset size and long-term ECoG-based BCI usage on deep learning decoders performance |
title_fullStr | Impact of dataset size and long-term ECoG-based BCI usage on deep learning decoders performance |
title_full_unstemmed | Impact of dataset size and long-term ECoG-based BCI usage on deep learning decoders performance |
title_short | Impact of dataset size and long-term ECoG-based BCI usage on deep learning decoders performance |
title_sort | impact of dataset size and long-term ecog-based bci usage on deep learning decoders performance |
topic | Human Neuroscience |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10061076/ https://www.ncbi.nlm.nih.gov/pubmed/37007675 http://dx.doi.org/10.3389/fnhum.2023.1111645 |
work_keys_str_mv | AT sliwowskimaciej impactofdatasetsizeandlongtermecogbasedbciusageondeeplearningdecodersperformance AT martinmatthieu impactofdatasetsizeandlongtermecogbasedbciusageondeeplearningdecodersperformance AT souloumiacantoine impactofdatasetsizeandlongtermecogbasedbciusageondeeplearningdecodersperformance AT blanchartpierre impactofdatasetsizeandlongtermecogbasedbciusageondeeplearningdecodersperformance AT aksenovatetiana impactofdatasetsizeandlongtermecogbasedbciusageondeeplearningdecodersperformance |