Cargando…

Evaluating Convolutional Neural Networks as a Method of EEG–EMG Fusion

Wearable robotic exoskeletons have emerged as an exciting new treatment tool for disorders affecting mobility; however, the human–machine interface, used by the patient for device control, requires further improvement before robotic assistance and rehabilitation can be widely adopted. One method, ma...

Descripción completa

Detalles Bibliográficos
Autores principales: Tryon, Jacob, Trejos, Ana Luisa
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8649783/
https://www.ncbi.nlm.nih.gov/pubmed/34887739
http://dx.doi.org/10.3389/fnbot.2021.692183
_version_ 1784611072374734848
author Tryon, Jacob
Trejos, Ana Luisa
author_facet Tryon, Jacob
Trejos, Ana Luisa
author_sort Tryon, Jacob
collection PubMed
description Wearable robotic exoskeletons have emerged as an exciting new treatment tool for disorders affecting mobility; however, the human–machine interface, used by the patient for device control, requires further improvement before robotic assistance and rehabilitation can be widely adopted. One method, made possible through advancements in machine learning technology, is the use of bioelectrical signals, such as electroencephalography (EEG) and electromyography (EMG), to classify the user's actions and intentions. While classification using these signals has been demonstrated for many relevant control tasks, such as motion intention detection and gesture recognition, challenges in decoding the bioelectrical signals have caused researchers to seek methods for improving the accuracy of these models. One such method is the use of EEG–EMG fusion, creating a classification model that decodes information from both EEG and EMG signals simultaneously to increase the amount of available information. So far, EEG–EMG fusion has been implemented using traditional machine learning methods that rely on manual feature extraction; however, new machine learning methods have emerged that can automatically extract relevant information from a dataset, which may prove beneficial during EEG–EMG fusion. In this study, Convolutional Neural Network (CNN) models were developed using combined EEG–EMG inputs to determine if they have potential as a method of EEG–EMG fusion that automatically extracts relevant information from both signals simultaneously. EEG and EMG signals were recorded during elbow flexion–extension and used to develop CNN models based on time–frequency (spectrogram) and time (filtered signal) domain image inputs. The results show a mean accuracy of 80.51 ± 8.07% for a three-class output (33.33% chance level), with an F-score of 80.74%, using time–frequency domain-based models. This work demonstrates the viability of CNNs as a new method of EEG–EMG fusion and evaluates different signal representations to determine the best implementation of a combined EEG–EMG CNN. It leverages modern machine learning methods to advance EEG–EMG fusion, which will ultimately lead to improvements in the usability of wearable robotic exoskeletons.
format Online
Article
Text
id pubmed-8649783
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-86497832021-12-08 Evaluating Convolutional Neural Networks as a Method of EEG–EMG Fusion Tryon, Jacob Trejos, Ana Luisa Front Neurorobot Neuroscience Wearable robotic exoskeletons have emerged as an exciting new treatment tool for disorders affecting mobility; however, the human–machine interface, used by the patient for device control, requires further improvement before robotic assistance and rehabilitation can be widely adopted. One method, made possible through advancements in machine learning technology, is the use of bioelectrical signals, such as electroencephalography (EEG) and electromyography (EMG), to classify the user's actions and intentions. While classification using these signals has been demonstrated for many relevant control tasks, such as motion intention detection and gesture recognition, challenges in decoding the bioelectrical signals have caused researchers to seek methods for improving the accuracy of these models. One such method is the use of EEG–EMG fusion, creating a classification model that decodes information from both EEG and EMG signals simultaneously to increase the amount of available information. So far, EEG–EMG fusion has been implemented using traditional machine learning methods that rely on manual feature extraction; however, new machine learning methods have emerged that can automatically extract relevant information from a dataset, which may prove beneficial during EEG–EMG fusion. In this study, Convolutional Neural Network (CNN) models were developed using combined EEG–EMG inputs to determine if they have potential as a method of EEG–EMG fusion that automatically extracts relevant information from both signals simultaneously. EEG and EMG signals were recorded during elbow flexion–extension and used to develop CNN models based on time–frequency (spectrogram) and time (filtered signal) domain image inputs. The results show a mean accuracy of 80.51 ± 8.07% for a three-class output (33.33% chance level), with an F-score of 80.74%, using time–frequency domain-based models. This work demonstrates the viability of CNNs as a new method of EEG–EMG fusion and evaluates different signal representations to determine the best implementation of a combined EEG–EMG CNN. It leverages modern machine learning methods to advance EEG–EMG fusion, which will ultimately lead to improvements in the usability of wearable robotic exoskeletons. Frontiers Media S.A. 2021-11-23 /pmc/articles/PMC8649783/ /pubmed/34887739 http://dx.doi.org/10.3389/fnbot.2021.692183 Text en Copyright © 2021 Tryon and Trejos. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Neuroscience
Tryon, Jacob
Trejos, Ana Luisa
Evaluating Convolutional Neural Networks as a Method of EEG–EMG Fusion
title Evaluating Convolutional Neural Networks as a Method of EEG–EMG Fusion
title_full Evaluating Convolutional Neural Networks as a Method of EEG–EMG Fusion
title_fullStr Evaluating Convolutional Neural Networks as a Method of EEG–EMG Fusion
title_full_unstemmed Evaluating Convolutional Neural Networks as a Method of EEG–EMG Fusion
title_short Evaluating Convolutional Neural Networks as a Method of EEG–EMG Fusion
title_sort evaluating convolutional neural networks as a method of eeg–emg fusion
topic Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8649783/
https://www.ncbi.nlm.nih.gov/pubmed/34887739
http://dx.doi.org/10.3389/fnbot.2021.692183
work_keys_str_mv AT tryonjacob evaluatingconvolutionalneuralnetworksasamethodofeegemgfusion
AT trejosanaluisa evaluatingconvolutionalneuralnetworksasamethodofeegemgfusion