Cargando…

Deep learning-based framework for real-time upper limb motion intention classification using combined bio-signals

This research study proposes a unique framework that takes input from a surface electromyogram (sEMG) and functional near-infrared spectroscopy (fNIRS) bio-signals. These signals are trained using convolutional neural networks (CNN). The framework entails a real-time neuro-machine interface to decod...

Descripción completa

Detalles Bibliográficos
Autores principales: Syed, A. Usama, Sattar, Neelum Y., Ganiyu, Ismaila, Sanjay, Chintakindi, Alkhatib, Soliman, Salah, Bashir
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10413572/
https://www.ncbi.nlm.nih.gov/pubmed/37575360
http://dx.doi.org/10.3389/fnbot.2023.1174613
_version_ 1785087158656172032
author Syed, A. Usama
Sattar, Neelum Y.
Ganiyu, Ismaila
Sanjay, Chintakindi
Alkhatib, Soliman
Salah, Bashir
author_facet Syed, A. Usama
Sattar, Neelum Y.
Ganiyu, Ismaila
Sanjay, Chintakindi
Alkhatib, Soliman
Salah, Bashir
author_sort Syed, A. Usama
collection PubMed
description This research study proposes a unique framework that takes input from a surface electromyogram (sEMG) and functional near-infrared spectroscopy (fNIRS) bio-signals. These signals are trained using convolutional neural networks (CNN). The framework entails a real-time neuro-machine interface to decode the human intention of upper limb motions. The bio-signals from the two modalities are recorded for eight movements simultaneously for prosthetic arm functions focusing on trans-humeral amputees. The fNIRS signals are acquired from the human motor cortex, while sEMG is recorded from the human bicep muscles. The selected classification and command generation features are the peak, minimum, and mean ΔHbO and ΔHbR values within a 2-s moving window. In the case of sEMG, wavelength, peak, and mean were extracted with a 150-ms moving window. It was found that this scheme generates eight motions with an enhanced average accuracy of 94.5%. The obtained results validate the adopted research methodology and potential for future real-time neural-machine interfaces to control prosthetic arms.
format Online
Article
Text
id pubmed-10413572
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-104135722023-08-11 Deep learning-based framework for real-time upper limb motion intention classification using combined bio-signals Syed, A. Usama Sattar, Neelum Y. Ganiyu, Ismaila Sanjay, Chintakindi Alkhatib, Soliman Salah, Bashir Front Neurorobot Neuroscience This research study proposes a unique framework that takes input from a surface electromyogram (sEMG) and functional near-infrared spectroscopy (fNIRS) bio-signals. These signals are trained using convolutional neural networks (CNN). The framework entails a real-time neuro-machine interface to decode the human intention of upper limb motions. The bio-signals from the two modalities are recorded for eight movements simultaneously for prosthetic arm functions focusing on trans-humeral amputees. The fNIRS signals are acquired from the human motor cortex, while sEMG is recorded from the human bicep muscles. The selected classification and command generation features are the peak, minimum, and mean ΔHbO and ΔHbR values within a 2-s moving window. In the case of sEMG, wavelength, peak, and mean were extracted with a 150-ms moving window. It was found that this scheme generates eight motions with an enhanced average accuracy of 94.5%. The obtained results validate the adopted research methodology and potential for future real-time neural-machine interfaces to control prosthetic arms. Frontiers Media S.A. 2023-07-27 /pmc/articles/PMC10413572/ /pubmed/37575360 http://dx.doi.org/10.3389/fnbot.2023.1174613 Text en Copyright © 2023 Syed, Sattar, Ganiyu, Sanjay, Alkhatib and Salah. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Neuroscience
Syed, A. Usama
Sattar, Neelum Y.
Ganiyu, Ismaila
Sanjay, Chintakindi
Alkhatib, Soliman
Salah, Bashir
Deep learning-based framework for real-time upper limb motion intention classification using combined bio-signals
title Deep learning-based framework for real-time upper limb motion intention classification using combined bio-signals
title_full Deep learning-based framework for real-time upper limb motion intention classification using combined bio-signals
title_fullStr Deep learning-based framework for real-time upper limb motion intention classification using combined bio-signals
title_full_unstemmed Deep learning-based framework for real-time upper limb motion intention classification using combined bio-signals
title_short Deep learning-based framework for real-time upper limb motion intention classification using combined bio-signals
title_sort deep learning-based framework for real-time upper limb motion intention classification using combined bio-signals
topic Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10413572/
https://www.ncbi.nlm.nih.gov/pubmed/37575360
http://dx.doi.org/10.3389/fnbot.2023.1174613
work_keys_str_mv AT syedausama deeplearningbasedframeworkforrealtimeupperlimbmotionintentionclassificationusingcombinedbiosignals
AT sattarneelumy deeplearningbasedframeworkforrealtimeupperlimbmotionintentionclassificationusingcombinedbiosignals
AT ganiyuismaila deeplearningbasedframeworkforrealtimeupperlimbmotionintentionclassificationusingcombinedbiosignals
AT sanjaychintakindi deeplearningbasedframeworkforrealtimeupperlimbmotionintentionclassificationusingcombinedbiosignals
AT alkhatibsoliman deeplearningbasedframeworkforrealtimeupperlimbmotionintentionclassificationusingcombinedbiosignals
AT salahbashir deeplearningbasedframeworkforrealtimeupperlimbmotionintentionclassificationusingcombinedbiosignals