Cargando…
Deep learning-based framework for real-time upper limb motion intention classification using combined bio-signals
This research study proposes a unique framework that takes input from a surface electromyogram (sEMG) and functional near-infrared spectroscopy (fNIRS) bio-signals. These signals are trained using convolutional neural networks (CNN). The framework entails a real-time neuro-machine interface to decod...
Autores principales: | , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Frontiers Media S.A.
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10413572/ https://www.ncbi.nlm.nih.gov/pubmed/37575360 http://dx.doi.org/10.3389/fnbot.2023.1174613 |
_version_ | 1785087158656172032 |
---|---|
author | Syed, A. Usama Sattar, Neelum Y. Ganiyu, Ismaila Sanjay, Chintakindi Alkhatib, Soliman Salah, Bashir |
author_facet | Syed, A. Usama Sattar, Neelum Y. Ganiyu, Ismaila Sanjay, Chintakindi Alkhatib, Soliman Salah, Bashir |
author_sort | Syed, A. Usama |
collection | PubMed |
description | This research study proposes a unique framework that takes input from a surface electromyogram (sEMG) and functional near-infrared spectroscopy (fNIRS) bio-signals. These signals are trained using convolutional neural networks (CNN). The framework entails a real-time neuro-machine interface to decode the human intention of upper limb motions. The bio-signals from the two modalities are recorded for eight movements simultaneously for prosthetic arm functions focusing on trans-humeral amputees. The fNIRS signals are acquired from the human motor cortex, while sEMG is recorded from the human bicep muscles. The selected classification and command generation features are the peak, minimum, and mean ΔHbO and ΔHbR values within a 2-s moving window. In the case of sEMG, wavelength, peak, and mean were extracted with a 150-ms moving window. It was found that this scheme generates eight motions with an enhanced average accuracy of 94.5%. The obtained results validate the adopted research methodology and potential for future real-time neural-machine interfaces to control prosthetic arms. |
format | Online Article Text |
id | pubmed-10413572 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | Frontiers Media S.A. |
record_format | MEDLINE/PubMed |
spelling | pubmed-104135722023-08-11 Deep learning-based framework for real-time upper limb motion intention classification using combined bio-signals Syed, A. Usama Sattar, Neelum Y. Ganiyu, Ismaila Sanjay, Chintakindi Alkhatib, Soliman Salah, Bashir Front Neurorobot Neuroscience This research study proposes a unique framework that takes input from a surface electromyogram (sEMG) and functional near-infrared spectroscopy (fNIRS) bio-signals. These signals are trained using convolutional neural networks (CNN). The framework entails a real-time neuro-machine interface to decode the human intention of upper limb motions. The bio-signals from the two modalities are recorded for eight movements simultaneously for prosthetic arm functions focusing on trans-humeral amputees. The fNIRS signals are acquired from the human motor cortex, while sEMG is recorded from the human bicep muscles. The selected classification and command generation features are the peak, minimum, and mean ΔHbO and ΔHbR values within a 2-s moving window. In the case of sEMG, wavelength, peak, and mean were extracted with a 150-ms moving window. It was found that this scheme generates eight motions with an enhanced average accuracy of 94.5%. The obtained results validate the adopted research methodology and potential for future real-time neural-machine interfaces to control prosthetic arms. Frontiers Media S.A. 2023-07-27 /pmc/articles/PMC10413572/ /pubmed/37575360 http://dx.doi.org/10.3389/fnbot.2023.1174613 Text en Copyright © 2023 Syed, Sattar, Ganiyu, Sanjay, Alkhatib and Salah. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. |
spellingShingle | Neuroscience Syed, A. Usama Sattar, Neelum Y. Ganiyu, Ismaila Sanjay, Chintakindi Alkhatib, Soliman Salah, Bashir Deep learning-based framework for real-time upper limb motion intention classification using combined bio-signals |
title | Deep learning-based framework for real-time upper limb motion intention classification using combined bio-signals |
title_full | Deep learning-based framework for real-time upper limb motion intention classification using combined bio-signals |
title_fullStr | Deep learning-based framework for real-time upper limb motion intention classification using combined bio-signals |
title_full_unstemmed | Deep learning-based framework for real-time upper limb motion intention classification using combined bio-signals |
title_short | Deep learning-based framework for real-time upper limb motion intention classification using combined bio-signals |
title_sort | deep learning-based framework for real-time upper limb motion intention classification using combined bio-signals |
topic | Neuroscience |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10413572/ https://www.ncbi.nlm.nih.gov/pubmed/37575360 http://dx.doi.org/10.3389/fnbot.2023.1174613 |
work_keys_str_mv | AT syedausama deeplearningbasedframeworkforrealtimeupperlimbmotionintentionclassificationusingcombinedbiosignals AT sattarneelumy deeplearningbasedframeworkforrealtimeupperlimbmotionintentionclassificationusingcombinedbiosignals AT ganiyuismaila deeplearningbasedframeworkforrealtimeupperlimbmotionintentionclassificationusingcombinedbiosignals AT sanjaychintakindi deeplearningbasedframeworkforrealtimeupperlimbmotionintentionclassificationusingcombinedbiosignals AT alkhatibsoliman deeplearningbasedframeworkforrealtimeupperlimbmotionintentionclassificationusingcombinedbiosignals AT salahbashir deeplearningbasedframeworkforrealtimeupperlimbmotionintentionclassificationusingcombinedbiosignals |