Cargando…

Discrimination, Reliability, Sensitivity, and Specificity of Robotic Surgical Proficiency Assessment With Global Evaluative Assessment of Robotic Skills and Binary Scoring Metrics: Results From a Randomized Controlled Trial

OBJECTIVE: To compare binary metrics and Global Evaluative Assessment of Robotic Skills (GEARS) evaluations of training outcome assessments for reliability, sensitivity, and specificity. BACKGROUND: GEARS–Likert-scale skills assessment are a widely accepted tool for robotic surgical training outcome...

Descripción completa

Detalles Bibliográficos
Autores principales: De Groote, Ruben, Puliatti, Stefano, Amato, Marco, Mazzone, Elio, Larcher, Alessandro, Farinha, Rui, Paludo, Artur, Desender, Liesbeth, Hubert, Nicolas, Cleynenbreugel, Ben Van, Bunting, Brendan P., Mottrie, Alexandre, Gallagher, Anthony G., Rosiello, Giuseppe, Uvin, Pieter, Decoene, Jasper, Tuyten, Tom, D’Hondt, Mathieu, Chatzopoulos, Charles, De Troyer, Bart, Turri, Filippo, Dell’Oglio, Paolo, Liakos, Nikolaos, Andrea Bravi, Carlo, Lambert, Edward, Andras, Iulia, Di Maida, Fabrizio, Everaerts, Wouter
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Wolters Kluwer Health, Inc. 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10513364/
https://www.ncbi.nlm.nih.gov/pubmed/37746611
http://dx.doi.org/10.1097/AS9.0000000000000307
_version_ 1785108553853304832
author De Groote, Ruben
Puliatti, Stefano
Amato, Marco
Mazzone, Elio
Larcher, Alessandro
Farinha, Rui
Paludo, Artur
Desender, Liesbeth
Hubert, Nicolas
Cleynenbreugel, Ben Van
Bunting, Brendan P.
Mottrie, Alexandre
Gallagher, Anthony G.
Rosiello, Giuseppe
Uvin, Pieter
Decoene, Jasper
Tuyten, Tom
D’Hondt, Mathieu
Chatzopoulos, Charles
De Troyer, Bart
Turri, Filippo
Dell’Oglio, Paolo
Liakos, Nikolaos
Andrea Bravi, Carlo
Lambert, Edward
Andras, Iulia
Di Maida, Fabrizio
Everaerts, Wouter
author_facet De Groote, Ruben
Puliatti, Stefano
Amato, Marco
Mazzone, Elio
Larcher, Alessandro
Farinha, Rui
Paludo, Artur
Desender, Liesbeth
Hubert, Nicolas
Cleynenbreugel, Ben Van
Bunting, Brendan P.
Mottrie, Alexandre
Gallagher, Anthony G.
Rosiello, Giuseppe
Uvin, Pieter
Decoene, Jasper
Tuyten, Tom
D’Hondt, Mathieu
Chatzopoulos, Charles
De Troyer, Bart
Turri, Filippo
Dell’Oglio, Paolo
Liakos, Nikolaos
Andrea Bravi, Carlo
Lambert, Edward
Andras, Iulia
Di Maida, Fabrizio
Everaerts, Wouter
author_sort De Groote, Ruben
collection PubMed
description OBJECTIVE: To compare binary metrics and Global Evaluative Assessment of Robotic Skills (GEARS) evaluations of training outcome assessments for reliability, sensitivity, and specificity. BACKGROUND: GEARS–Likert-scale skills assessment are a widely accepted tool for robotic surgical training outcome evaluations. Proficiency-based progression (PBP) training is another methodology but uses binary performance metrics for evaluations. METHODS: In a prospective, randomized, and blinded study, we compared conventional with PBP training for a robotic suturing, knot-tying anastomosis task. Thirty-six surgical residents from 16 Belgium residency programs were randomized. In the skills laboratory, the PBP group trained until they demonstrated a quantitatively defined proficiency benchmark. The conventional group were yoked to the same training time but without the proficiency requirement. The final trial was video recorded and assessed with binary metrics and GEARS by robotic surgeons blinded to individual, group, and residency program. Sensitivity and specificity of the two assessment methods were evaluated with area under the curve (AUC) and receiver operating characteristics (ROC) curves. RESULTS: The PBP group made 42% fewer objectively assessed performance errors than the conventional group (P < 0.001) and scored 15% better on the GEARS assessment (P = 0.033). The mean interrater reliability for binary metrics and GEARS was 0.87 and 0.38, respectively. Binary total error metrics AUC was 97% and for GEARS 85%. With a sensitivity threshold of 0.8, false positives rates were 3% and 25% for, respectively, the binary and GEARS assessments. CONCLUSIONS: Binary metrics for scoring a robotic VUA task demonstrated better psychometric properties than the GEARS assessment.
format Online
Article
Text
id pubmed-10513364
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Wolters Kluwer Health, Inc.
record_format MEDLINE/PubMed
spelling pubmed-105133642023-09-22 Discrimination, Reliability, Sensitivity, and Specificity of Robotic Surgical Proficiency Assessment With Global Evaluative Assessment of Robotic Skills and Binary Scoring Metrics: Results From a Randomized Controlled Trial De Groote, Ruben Puliatti, Stefano Amato, Marco Mazzone, Elio Larcher, Alessandro Farinha, Rui Paludo, Artur Desender, Liesbeth Hubert, Nicolas Cleynenbreugel, Ben Van Bunting, Brendan P. Mottrie, Alexandre Gallagher, Anthony G. Rosiello, Giuseppe Uvin, Pieter Decoene, Jasper Tuyten, Tom D’Hondt, Mathieu Chatzopoulos, Charles De Troyer, Bart Turri, Filippo Dell’Oglio, Paolo Liakos, Nikolaos Andrea Bravi, Carlo Lambert, Edward Andras, Iulia Di Maida, Fabrizio Everaerts, Wouter Ann Surg Open Original Study OBJECTIVE: To compare binary metrics and Global Evaluative Assessment of Robotic Skills (GEARS) evaluations of training outcome assessments for reliability, sensitivity, and specificity. BACKGROUND: GEARS–Likert-scale skills assessment are a widely accepted tool for robotic surgical training outcome evaluations. Proficiency-based progression (PBP) training is another methodology but uses binary performance metrics for evaluations. METHODS: In a prospective, randomized, and blinded study, we compared conventional with PBP training for a robotic suturing, knot-tying anastomosis task. Thirty-six surgical residents from 16 Belgium residency programs were randomized. In the skills laboratory, the PBP group trained until they demonstrated a quantitatively defined proficiency benchmark. The conventional group were yoked to the same training time but without the proficiency requirement. The final trial was video recorded and assessed with binary metrics and GEARS by robotic surgeons blinded to individual, group, and residency program. Sensitivity and specificity of the two assessment methods were evaluated with area under the curve (AUC) and receiver operating characteristics (ROC) curves. RESULTS: The PBP group made 42% fewer objectively assessed performance errors than the conventional group (P < 0.001) and scored 15% better on the GEARS assessment (P = 0.033). The mean interrater reliability for binary metrics and GEARS was 0.87 and 0.38, respectively. Binary total error metrics AUC was 97% and for GEARS 85%. With a sensitivity threshold of 0.8, false positives rates were 3% and 25% for, respectively, the binary and GEARS assessments. CONCLUSIONS: Binary metrics for scoring a robotic VUA task demonstrated better psychometric properties than the GEARS assessment. Wolters Kluwer Health, Inc. 2023-08-16 /pmc/articles/PMC10513364/ /pubmed/37746611 http://dx.doi.org/10.1097/AS9.0000000000000307 Text en Copyright © 2023 The Author(s). Published by Wolters Kluwer Health, Inc. https://creativecommons.org/licenses/by-nc-nd/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution-Non Commercial-No Derivatives License 4.0 (CCBY-NC-ND (https://creativecommons.org/licenses/by-nc-nd/4.0/) ), where it is permissible to download and share the work provided it is properly cited. The work cannot be changed in any way or used commercially without permission from the journal.
spellingShingle Original Study
De Groote, Ruben
Puliatti, Stefano
Amato, Marco
Mazzone, Elio
Larcher, Alessandro
Farinha, Rui
Paludo, Artur
Desender, Liesbeth
Hubert, Nicolas
Cleynenbreugel, Ben Van
Bunting, Brendan P.
Mottrie, Alexandre
Gallagher, Anthony G.
Rosiello, Giuseppe
Uvin, Pieter
Decoene, Jasper
Tuyten, Tom
D’Hondt, Mathieu
Chatzopoulos, Charles
De Troyer, Bart
Turri, Filippo
Dell’Oglio, Paolo
Liakos, Nikolaos
Andrea Bravi, Carlo
Lambert, Edward
Andras, Iulia
Di Maida, Fabrizio
Everaerts, Wouter
Discrimination, Reliability, Sensitivity, and Specificity of Robotic Surgical Proficiency Assessment With Global Evaluative Assessment of Robotic Skills and Binary Scoring Metrics: Results From a Randomized Controlled Trial
title Discrimination, Reliability, Sensitivity, and Specificity of Robotic Surgical Proficiency Assessment With Global Evaluative Assessment of Robotic Skills and Binary Scoring Metrics: Results From a Randomized Controlled Trial
title_full Discrimination, Reliability, Sensitivity, and Specificity of Robotic Surgical Proficiency Assessment With Global Evaluative Assessment of Robotic Skills and Binary Scoring Metrics: Results From a Randomized Controlled Trial
title_fullStr Discrimination, Reliability, Sensitivity, and Specificity of Robotic Surgical Proficiency Assessment With Global Evaluative Assessment of Robotic Skills and Binary Scoring Metrics: Results From a Randomized Controlled Trial
title_full_unstemmed Discrimination, Reliability, Sensitivity, and Specificity of Robotic Surgical Proficiency Assessment With Global Evaluative Assessment of Robotic Skills and Binary Scoring Metrics: Results From a Randomized Controlled Trial
title_short Discrimination, Reliability, Sensitivity, and Specificity of Robotic Surgical Proficiency Assessment With Global Evaluative Assessment of Robotic Skills and Binary Scoring Metrics: Results From a Randomized Controlled Trial
title_sort discrimination, reliability, sensitivity, and specificity of robotic surgical proficiency assessment with global evaluative assessment of robotic skills and binary scoring metrics: results from a randomized controlled trial
topic Original Study
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10513364/
https://www.ncbi.nlm.nih.gov/pubmed/37746611
http://dx.doi.org/10.1097/AS9.0000000000000307
work_keys_str_mv AT degrooteruben discriminationreliabilitysensitivityandspecificityofroboticsurgicalproficiencyassessmentwithglobalevaluativeassessmentofroboticskillsandbinaryscoringmetricsresultsfromarandomizedcontrolledtrial
AT puliattistefano discriminationreliabilitysensitivityandspecificityofroboticsurgicalproficiencyassessmentwithglobalevaluativeassessmentofroboticskillsandbinaryscoringmetricsresultsfromarandomizedcontrolledtrial
AT amatomarco discriminationreliabilitysensitivityandspecificityofroboticsurgicalproficiencyassessmentwithglobalevaluativeassessmentofroboticskillsandbinaryscoringmetricsresultsfromarandomizedcontrolledtrial
AT mazzoneelio discriminationreliabilitysensitivityandspecificityofroboticsurgicalproficiencyassessmentwithglobalevaluativeassessmentofroboticskillsandbinaryscoringmetricsresultsfromarandomizedcontrolledtrial
AT larcheralessandro discriminationreliabilitysensitivityandspecificityofroboticsurgicalproficiencyassessmentwithglobalevaluativeassessmentofroboticskillsandbinaryscoringmetricsresultsfromarandomizedcontrolledtrial
AT farinharui discriminationreliabilitysensitivityandspecificityofroboticsurgicalproficiencyassessmentwithglobalevaluativeassessmentofroboticskillsandbinaryscoringmetricsresultsfromarandomizedcontrolledtrial
AT paludoartur discriminationreliabilitysensitivityandspecificityofroboticsurgicalproficiencyassessmentwithglobalevaluativeassessmentofroboticskillsandbinaryscoringmetricsresultsfromarandomizedcontrolledtrial
AT desenderliesbeth discriminationreliabilitysensitivityandspecificityofroboticsurgicalproficiencyassessmentwithglobalevaluativeassessmentofroboticskillsandbinaryscoringmetricsresultsfromarandomizedcontrolledtrial
AT hubertnicolas discriminationreliabilitysensitivityandspecificityofroboticsurgicalproficiencyassessmentwithglobalevaluativeassessmentofroboticskillsandbinaryscoringmetricsresultsfromarandomizedcontrolledtrial
AT cleynenbreugelbenvan discriminationreliabilitysensitivityandspecificityofroboticsurgicalproficiencyassessmentwithglobalevaluativeassessmentofroboticskillsandbinaryscoringmetricsresultsfromarandomizedcontrolledtrial
AT buntingbrendanp discriminationreliabilitysensitivityandspecificityofroboticsurgicalproficiencyassessmentwithglobalevaluativeassessmentofroboticskillsandbinaryscoringmetricsresultsfromarandomizedcontrolledtrial
AT mottriealexandre discriminationreliabilitysensitivityandspecificityofroboticsurgicalproficiencyassessmentwithglobalevaluativeassessmentofroboticskillsandbinaryscoringmetricsresultsfromarandomizedcontrolledtrial
AT gallagheranthonyg discriminationreliabilitysensitivityandspecificityofroboticsurgicalproficiencyassessmentwithglobalevaluativeassessmentofroboticskillsandbinaryscoringmetricsresultsfromarandomizedcontrolledtrial
AT rosiellogiuseppe discriminationreliabilitysensitivityandspecificityofroboticsurgicalproficiencyassessmentwithglobalevaluativeassessmentofroboticskillsandbinaryscoringmetricsresultsfromarandomizedcontrolledtrial
AT uvinpieter discriminationreliabilitysensitivityandspecificityofroboticsurgicalproficiencyassessmentwithglobalevaluativeassessmentofroboticskillsandbinaryscoringmetricsresultsfromarandomizedcontrolledtrial
AT decoenejasper discriminationreliabilitysensitivityandspecificityofroboticsurgicalproficiencyassessmentwithglobalevaluativeassessmentofroboticskillsandbinaryscoringmetricsresultsfromarandomizedcontrolledtrial
AT tuytentom discriminationreliabilitysensitivityandspecificityofroboticsurgicalproficiencyassessmentwithglobalevaluativeassessmentofroboticskillsandbinaryscoringmetricsresultsfromarandomizedcontrolledtrial
AT dhondtmathieu discriminationreliabilitysensitivityandspecificityofroboticsurgicalproficiencyassessmentwithglobalevaluativeassessmentofroboticskillsandbinaryscoringmetricsresultsfromarandomizedcontrolledtrial
AT chatzopouloscharles discriminationreliabilitysensitivityandspecificityofroboticsurgicalproficiencyassessmentwithglobalevaluativeassessmentofroboticskillsandbinaryscoringmetricsresultsfromarandomizedcontrolledtrial
AT detroyerbart discriminationreliabilitysensitivityandspecificityofroboticsurgicalproficiencyassessmentwithglobalevaluativeassessmentofroboticskillsandbinaryscoringmetricsresultsfromarandomizedcontrolledtrial
AT turrifilippo discriminationreliabilitysensitivityandspecificityofroboticsurgicalproficiencyassessmentwithglobalevaluativeassessmentofroboticskillsandbinaryscoringmetricsresultsfromarandomizedcontrolledtrial
AT dellogliopaolo discriminationreliabilitysensitivityandspecificityofroboticsurgicalproficiencyassessmentwithglobalevaluativeassessmentofroboticskillsandbinaryscoringmetricsresultsfromarandomizedcontrolledtrial
AT liakosnikolaos discriminationreliabilitysensitivityandspecificityofroboticsurgicalproficiencyassessmentwithglobalevaluativeassessmentofroboticskillsandbinaryscoringmetricsresultsfromarandomizedcontrolledtrial
AT andreabravicarlo discriminationreliabilitysensitivityandspecificityofroboticsurgicalproficiencyassessmentwithglobalevaluativeassessmentofroboticskillsandbinaryscoringmetricsresultsfromarandomizedcontrolledtrial
AT lambertedward discriminationreliabilitysensitivityandspecificityofroboticsurgicalproficiencyassessmentwithglobalevaluativeassessmentofroboticskillsandbinaryscoringmetricsresultsfromarandomizedcontrolledtrial
AT andrasiulia discriminationreliabilitysensitivityandspecificityofroboticsurgicalproficiencyassessmentwithglobalevaluativeassessmentofroboticskillsandbinaryscoringmetricsresultsfromarandomizedcontrolledtrial
AT dimaidafabrizio discriminationreliabilitysensitivityandspecificityofroboticsurgicalproficiencyassessmentwithglobalevaluativeassessmentofroboticskillsandbinaryscoringmetricsresultsfromarandomizedcontrolledtrial
AT everaertswouter discriminationreliabilitysensitivityandspecificityofroboticsurgicalproficiencyassessmentwithglobalevaluativeassessmentofroboticskillsandbinaryscoringmetricsresultsfromarandomizedcontrolledtrial