Cargando…
Virtual Integration Environment as an Advanced Prosthetic Limb Training Platform
Background: Despite advances in prosthetic development and neurorehabilitation, individuals with upper extremity (UE) loss continue to face functional and psychosocial challenges following amputation. Recent advanced myoelectric prostheses offer intuitive control over multiple, simultaneous degrees...
Autores principales: | , , , , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Frontiers Media S.A.
2018
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6232892/ https://www.ncbi.nlm.nih.gov/pubmed/30459696 http://dx.doi.org/10.3389/fneur.2018.00785 |
_version_ | 1783370478302265344 |
---|---|
author | Perry, Briana N. Armiger, Robert S. Yu, Kristin E. Alattar, Ali A. Moran, Courtney W. Wolde, Mikias McFarland, Kayla Pasquina, Paul F. Tsao, Jack W. |
author_facet | Perry, Briana N. Armiger, Robert S. Yu, Kristin E. Alattar, Ali A. Moran, Courtney W. Wolde, Mikias McFarland, Kayla Pasquina, Paul F. Tsao, Jack W. |
author_sort | Perry, Briana N. |
collection | PubMed |
description | Background: Despite advances in prosthetic development and neurorehabilitation, individuals with upper extremity (UE) loss continue to face functional and psychosocial challenges following amputation. Recent advanced myoelectric prostheses offer intuitive control over multiple, simultaneous degrees of motion and promise sensory feedback integration, but require complex training to effectively manipulate. We explored whether a virtual reality simulator could be used to teach dexterous prosthetic control paradigms to individuals with UE loss. Methods: Thirteen active-duty military personnel with UE loss (14 limbs) completed twenty, 30-min passive motor training sessions over 1–2 months. Participants were asked to follow the motions of a virtual avatar using residual and phantom limbs, and electrical activity from the residual limb was recorded using surface electromyography. Eight participants (nine limbs), also completed twenty, 30-min active motor training sessions. Participants controlled a virtual avatar through three motion sets of increasing complexity (Basic, Advanced, and Digit) and were scored on how accurately they performed requested motions. Score trajectory was assessed as a function of time using longitudinal mixed effects linear regression. Results: Mean classification accuracy for passive motor training was 43.8 ± 10.7% (14 limbs, 277 passive sessions). In active motor sessions, >95% classification accuracy (which we used as the threshold for prosthetic acceptance) was achieved by all participants for Basic sets and by 50% of participants in Advanced and Digit sets. Significant improvement in active motor scores over time was observed in Basic and Advanced sets (per additional session: β-coefficient 0.125, p = 0.022; β-coefficient 0.45, p = 0.001, respectively), and trended toward significance for Digit sets (β-coefficient 0.594, p = 0.077). Conclusions: These results offer robust evidence that a virtual reality training platform can be used to quickly and efficiently train individuals with UE loss to operate advanced prosthetic control paradigms. Participants can be trained to generate muscle contraction patterns in residual limbs that are interpreted with high accuracy by computer software as distinct active motion commands. These results support the potential viability of advanced myoelectric prostheses relying on pattern recognition feedback or similar controls systems. |
format | Online Article Text |
id | pubmed-6232892 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2018 |
publisher | Frontiers Media S.A. |
record_format | MEDLINE/PubMed |
spelling | pubmed-62328922018-11-20 Virtual Integration Environment as an Advanced Prosthetic Limb Training Platform Perry, Briana N. Armiger, Robert S. Yu, Kristin E. Alattar, Ali A. Moran, Courtney W. Wolde, Mikias McFarland, Kayla Pasquina, Paul F. Tsao, Jack W. Front Neurol Neurology Background: Despite advances in prosthetic development and neurorehabilitation, individuals with upper extremity (UE) loss continue to face functional and psychosocial challenges following amputation. Recent advanced myoelectric prostheses offer intuitive control over multiple, simultaneous degrees of motion and promise sensory feedback integration, but require complex training to effectively manipulate. We explored whether a virtual reality simulator could be used to teach dexterous prosthetic control paradigms to individuals with UE loss. Methods: Thirteen active-duty military personnel with UE loss (14 limbs) completed twenty, 30-min passive motor training sessions over 1–2 months. Participants were asked to follow the motions of a virtual avatar using residual and phantom limbs, and electrical activity from the residual limb was recorded using surface electromyography. Eight participants (nine limbs), also completed twenty, 30-min active motor training sessions. Participants controlled a virtual avatar through three motion sets of increasing complexity (Basic, Advanced, and Digit) and were scored on how accurately they performed requested motions. Score trajectory was assessed as a function of time using longitudinal mixed effects linear regression. Results: Mean classification accuracy for passive motor training was 43.8 ± 10.7% (14 limbs, 277 passive sessions). In active motor sessions, >95% classification accuracy (which we used as the threshold for prosthetic acceptance) was achieved by all participants for Basic sets and by 50% of participants in Advanced and Digit sets. Significant improvement in active motor scores over time was observed in Basic and Advanced sets (per additional session: β-coefficient 0.125, p = 0.022; β-coefficient 0.45, p = 0.001, respectively), and trended toward significance for Digit sets (β-coefficient 0.594, p = 0.077). Conclusions: These results offer robust evidence that a virtual reality training platform can be used to quickly and efficiently train individuals with UE loss to operate advanced prosthetic control paradigms. Participants can be trained to generate muscle contraction patterns in residual limbs that are interpreted with high accuracy by computer software as distinct active motion commands. These results support the potential viability of advanced myoelectric prostheses relying on pattern recognition feedback or similar controls systems. Frontiers Media S.A. 2018-10-17 /pmc/articles/PMC6232892/ /pubmed/30459696 http://dx.doi.org/10.3389/fneur.2018.00785 Text en Copyright © 2018 Perry, Armiger, Yu, Alattar, Moran, Wolde, McFarland, Pasquina and Tsao. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. |
spellingShingle | Neurology Perry, Briana N. Armiger, Robert S. Yu, Kristin E. Alattar, Ali A. Moran, Courtney W. Wolde, Mikias McFarland, Kayla Pasquina, Paul F. Tsao, Jack W. Virtual Integration Environment as an Advanced Prosthetic Limb Training Platform |
title | Virtual Integration Environment as an Advanced Prosthetic Limb Training Platform |
title_full | Virtual Integration Environment as an Advanced Prosthetic Limb Training Platform |
title_fullStr | Virtual Integration Environment as an Advanced Prosthetic Limb Training Platform |
title_full_unstemmed | Virtual Integration Environment as an Advanced Prosthetic Limb Training Platform |
title_short | Virtual Integration Environment as an Advanced Prosthetic Limb Training Platform |
title_sort | virtual integration environment as an advanced prosthetic limb training platform |
topic | Neurology |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6232892/ https://www.ncbi.nlm.nih.gov/pubmed/30459696 http://dx.doi.org/10.3389/fneur.2018.00785 |
work_keys_str_mv | AT perrybrianan virtualintegrationenvironmentasanadvancedprostheticlimbtrainingplatform AT armigerroberts virtualintegrationenvironmentasanadvancedprostheticlimbtrainingplatform AT yukristine virtualintegrationenvironmentasanadvancedprostheticlimbtrainingplatform AT alattaralia virtualintegrationenvironmentasanadvancedprostheticlimbtrainingplatform AT morancourtneyw virtualintegrationenvironmentasanadvancedprostheticlimbtrainingplatform AT woldemikias virtualintegrationenvironmentasanadvancedprostheticlimbtrainingplatform AT mcfarlandkayla virtualintegrationenvironmentasanadvancedprostheticlimbtrainingplatform AT pasquinapaulf virtualintegrationenvironmentasanadvancedprostheticlimbtrainingplatform AT tsaojackw virtualintegrationenvironmentasanadvancedprostheticlimbtrainingplatform |