Cargando…

Surgical Hand Gesture Recognition Utilizing Electroencephalogram as Input to the Machine Learning and Network Neuroscience Algorithms

Surgical gestures detection can provide targeted, automated surgical skill assessment and feedback during surgical training for robot-assisted surgery (RAS). Several sources including surgical videos, robot tool kinematics, and an electromyogram (EMG) have been proposed to reach this goal. We aimed...

Descripción completa

Detalles Bibliográficos
Autores principales: Shafiei, Somayeh B., Durrani, Mohammad, Jing, Zhe, Mostowy, Michael, Doherty, Philippa, Hussein, Ahmed A., Elsayed, Ahmed S., Iqbal, Umar, Guru, Khurshid
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7959280/
https://www.ncbi.nlm.nih.gov/pubmed/33802372
http://dx.doi.org/10.3390/s21051733
_version_ 1783664936963014656
author Shafiei, Somayeh B.
Durrani, Mohammad
Jing, Zhe
Mostowy, Michael
Doherty, Philippa
Hussein, Ahmed A.
Elsayed, Ahmed S.
Iqbal, Umar
Guru, Khurshid
author_facet Shafiei, Somayeh B.
Durrani, Mohammad
Jing, Zhe
Mostowy, Michael
Doherty, Philippa
Hussein, Ahmed A.
Elsayed, Ahmed S.
Iqbal, Umar
Guru, Khurshid
author_sort Shafiei, Somayeh B.
collection PubMed
description Surgical gestures detection can provide targeted, automated surgical skill assessment and feedback during surgical training for robot-assisted surgery (RAS). Several sources including surgical videos, robot tool kinematics, and an electromyogram (EMG) have been proposed to reach this goal. We aimed to extract features from electroencephalogram (EEG) data and use them in machine learning algorithms to classify robot-assisted surgical gestures. EEG was collected from five RAS surgeons with varying experience while performing 34 robot-assisted radical prostatectomies over the course of three years. Eight dominant hand and six non-dominant hand gesture types were extracted and synchronized with associated EEG data. Network neuroscience algorithms were utilized to extract functional brain network and power spectral density features. Sixty extracted features were used as input to machine learning algorithms to classify gesture types. The analysis of variance (ANOVA) F-value statistical method was used for feature selection and 10-fold cross-validation was used to validate the proposed method. The proposed feature set used in the extra trees (ET) algorithm classified eight gesture types performed by the dominant hand of five RAS surgeons with an accuracy of 90%, precision: 90%, sensitivity: 88%, and also classified six gesture types performed by the non-dominant hand with an accuracy of 93%, precision: 94%, sensitivity: 94%.
format Online
Article
Text
id pubmed-7959280
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-79592802021-03-16 Surgical Hand Gesture Recognition Utilizing Electroencephalogram as Input to the Machine Learning and Network Neuroscience Algorithms Shafiei, Somayeh B. Durrani, Mohammad Jing, Zhe Mostowy, Michael Doherty, Philippa Hussein, Ahmed A. Elsayed, Ahmed S. Iqbal, Umar Guru, Khurshid Sensors (Basel) Article Surgical gestures detection can provide targeted, automated surgical skill assessment and feedback during surgical training for robot-assisted surgery (RAS). Several sources including surgical videos, robot tool kinematics, and an electromyogram (EMG) have been proposed to reach this goal. We aimed to extract features from electroencephalogram (EEG) data and use them in machine learning algorithms to classify robot-assisted surgical gestures. EEG was collected from five RAS surgeons with varying experience while performing 34 robot-assisted radical prostatectomies over the course of three years. Eight dominant hand and six non-dominant hand gesture types were extracted and synchronized with associated EEG data. Network neuroscience algorithms were utilized to extract functional brain network and power spectral density features. Sixty extracted features were used as input to machine learning algorithms to classify gesture types. The analysis of variance (ANOVA) F-value statistical method was used for feature selection and 10-fold cross-validation was used to validate the proposed method. The proposed feature set used in the extra trees (ET) algorithm classified eight gesture types performed by the dominant hand of five RAS surgeons with an accuracy of 90%, precision: 90%, sensitivity: 88%, and also classified six gesture types performed by the non-dominant hand with an accuracy of 93%, precision: 94%, sensitivity: 94%. MDPI 2021-03-03 /pmc/articles/PMC7959280/ /pubmed/33802372 http://dx.doi.org/10.3390/s21051733 Text en © 2021 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Shafiei, Somayeh B.
Durrani, Mohammad
Jing, Zhe
Mostowy, Michael
Doherty, Philippa
Hussein, Ahmed A.
Elsayed, Ahmed S.
Iqbal, Umar
Guru, Khurshid
Surgical Hand Gesture Recognition Utilizing Electroencephalogram as Input to the Machine Learning and Network Neuroscience Algorithms
title Surgical Hand Gesture Recognition Utilizing Electroencephalogram as Input to the Machine Learning and Network Neuroscience Algorithms
title_full Surgical Hand Gesture Recognition Utilizing Electroencephalogram as Input to the Machine Learning and Network Neuroscience Algorithms
title_fullStr Surgical Hand Gesture Recognition Utilizing Electroencephalogram as Input to the Machine Learning and Network Neuroscience Algorithms
title_full_unstemmed Surgical Hand Gesture Recognition Utilizing Electroencephalogram as Input to the Machine Learning and Network Neuroscience Algorithms
title_short Surgical Hand Gesture Recognition Utilizing Electroencephalogram as Input to the Machine Learning and Network Neuroscience Algorithms
title_sort surgical hand gesture recognition utilizing electroencephalogram as input to the machine learning and network neuroscience algorithms
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7959280/
https://www.ncbi.nlm.nih.gov/pubmed/33802372
http://dx.doi.org/10.3390/s21051733
work_keys_str_mv AT shafieisomayehb surgicalhandgesturerecognitionutilizingelectroencephalogramasinputtothemachinelearningandnetworkneurosciencealgorithms
AT durranimohammad surgicalhandgesturerecognitionutilizingelectroencephalogramasinputtothemachinelearningandnetworkneurosciencealgorithms
AT jingzhe surgicalhandgesturerecognitionutilizingelectroencephalogramasinputtothemachinelearningandnetworkneurosciencealgorithms
AT mostowymichael surgicalhandgesturerecognitionutilizingelectroencephalogramasinputtothemachinelearningandnetworkneurosciencealgorithms
AT dohertyphilippa surgicalhandgesturerecognitionutilizingelectroencephalogramasinputtothemachinelearningandnetworkneurosciencealgorithms
AT husseinahmeda surgicalhandgesturerecognitionutilizingelectroencephalogramasinputtothemachinelearningandnetworkneurosciencealgorithms
AT elsayedahmeds surgicalhandgesturerecognitionutilizingelectroencephalogramasinputtothemachinelearningandnetworkneurosciencealgorithms
AT iqbalumar surgicalhandgesturerecognitionutilizingelectroencephalogramasinputtothemachinelearningandnetworkneurosciencealgorithms
AT gurukhurshid surgicalhandgesturerecognitionutilizingelectroencephalogramasinputtothemachinelearningandnetworkneurosciencealgorithms