Cargando…
Objective evaluation of laparoscopic surgical skills in wet lab training based on motion analysis and machine learning
BACKGROUND: Our aim was to build a skill assessment system, providing objective feedback to trainees based on the motion metrics of laparoscopic surgical instruments. METHODS: Participants performed tissue dissection around the aorta (tissue dissection task) and renal parenchymal closure (parenchyma...
Autores principales: | , , , , , , , , , , , , , , , , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Springer Berlin Heidelberg
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9399206/ https://www.ncbi.nlm.nih.gov/pubmed/35394212 http://dx.doi.org/10.1007/s00423-022-02505-9 |
_version_ | 1784772471670439936 |
---|---|
author | Ebina, Koki Abe, Takashige Hotta, Kiyohiko Higuchi, Madoka Furumido, Jun Iwahara, Naoya Kon, Masafumi Miyaji, Kou Shibuya, Sayaka Lingbo, Yan Komizunai, Shunsuke Kurashima, Yo Kikuchi, Hiroshi Matsumoto, Ryuji Osawa, Takahiro Murai, Sachiyo Tsujita, Teppei Sase, Kazuya Chen, Xiaoshuai Konno, Atsushi Shinohara, Nobuo |
author_facet | Ebina, Koki Abe, Takashige Hotta, Kiyohiko Higuchi, Madoka Furumido, Jun Iwahara, Naoya Kon, Masafumi Miyaji, Kou Shibuya, Sayaka Lingbo, Yan Komizunai, Shunsuke Kurashima, Yo Kikuchi, Hiroshi Matsumoto, Ryuji Osawa, Takahiro Murai, Sachiyo Tsujita, Teppei Sase, Kazuya Chen, Xiaoshuai Konno, Atsushi Shinohara, Nobuo |
author_sort | Ebina, Koki |
collection | PubMed |
description | BACKGROUND: Our aim was to build a skill assessment system, providing objective feedback to trainees based on the motion metrics of laparoscopic surgical instruments. METHODS: Participants performed tissue dissection around the aorta (tissue dissection task) and renal parenchymal closure (parenchymal-suturing task), using swine organs in a box trainer under a motion capture (Mocap) system. Two experts assessed the recorded movies, according to the formula of global operative assessment of laparoscopic skills (GOALS: score range, 5–25), and the mean scores were utilized as objective variables in the regression analyses. The correlations between mean GOALS scores and Mocap metrics were evaluated, and potential Mocap metrics with a Spearman’s rank correlation coefficient value exceeding 0.4 were selected for each GOALS item estimation. Four regression algorithms, support vector regression (SVR), principal component analysis (PCA)-SVR, ridge regression, and partial least squares regression, were utilized for automatic GOALS estimation. Model validation was conducted by nested and repeated k-fold cross validation, and the mean absolute error (MAE) was calculated to evaluate the accuracy of each regression model. RESULTS: Forty-five urologic, 9 gastroenterological, and 3 gynecologic surgeons, 4 junior residents, and 9 medical students participated in the training. In both tasks, a positive correlation was observed between the speed-related parameters (e.g., velocity, velocity range, acceleration, jerk) and mean GOALS scores, with a negative correlation between the efficiency-related parameters (e.g., task time, path length, number of opening/closing operations) and mean GOALS scores. Among the 4 algorithms, SVR showed the highest accuracy in the tissue dissection task ([Formula: see text] ), and PCA-SVR in the parenchymal-suturing task ([Formula: see text] ), based on 100 iterations of the validation process of automatic GOALS estimation. CONCLUSION: We developed a machine learning–based GOALS scoring system in wet lab training, with an error of approximately 1–2 points for the total score, and motion metrics that were explainable to trainees. Our future challenges are the further improvement of onsite GOALS feedback, exploring the educational benefit of our model and building an efficient training program. SUPPLEMENTARY INFORMATION: The online version contains supplementary material available at 10.1007/s00423-022-02505-9. |
format | Online Article Text |
id | pubmed-9399206 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Springer Berlin Heidelberg |
record_format | MEDLINE/PubMed |
spelling | pubmed-93992062022-08-25 Objective evaluation of laparoscopic surgical skills in wet lab training based on motion analysis and machine learning Ebina, Koki Abe, Takashige Hotta, Kiyohiko Higuchi, Madoka Furumido, Jun Iwahara, Naoya Kon, Masafumi Miyaji, Kou Shibuya, Sayaka Lingbo, Yan Komizunai, Shunsuke Kurashima, Yo Kikuchi, Hiroshi Matsumoto, Ryuji Osawa, Takahiro Murai, Sachiyo Tsujita, Teppei Sase, Kazuya Chen, Xiaoshuai Konno, Atsushi Shinohara, Nobuo Langenbecks Arch Surg Original Article BACKGROUND: Our aim was to build a skill assessment system, providing objective feedback to trainees based on the motion metrics of laparoscopic surgical instruments. METHODS: Participants performed tissue dissection around the aorta (tissue dissection task) and renal parenchymal closure (parenchymal-suturing task), using swine organs in a box trainer under a motion capture (Mocap) system. Two experts assessed the recorded movies, according to the formula of global operative assessment of laparoscopic skills (GOALS: score range, 5–25), and the mean scores were utilized as objective variables in the regression analyses. The correlations between mean GOALS scores and Mocap metrics were evaluated, and potential Mocap metrics with a Spearman’s rank correlation coefficient value exceeding 0.4 were selected for each GOALS item estimation. Four regression algorithms, support vector regression (SVR), principal component analysis (PCA)-SVR, ridge regression, and partial least squares regression, were utilized for automatic GOALS estimation. Model validation was conducted by nested and repeated k-fold cross validation, and the mean absolute error (MAE) was calculated to evaluate the accuracy of each regression model. RESULTS: Forty-five urologic, 9 gastroenterological, and 3 gynecologic surgeons, 4 junior residents, and 9 medical students participated in the training. In both tasks, a positive correlation was observed between the speed-related parameters (e.g., velocity, velocity range, acceleration, jerk) and mean GOALS scores, with a negative correlation between the efficiency-related parameters (e.g., task time, path length, number of opening/closing operations) and mean GOALS scores. Among the 4 algorithms, SVR showed the highest accuracy in the tissue dissection task ([Formula: see text] ), and PCA-SVR in the parenchymal-suturing task ([Formula: see text] ), based on 100 iterations of the validation process of automatic GOALS estimation. CONCLUSION: We developed a machine learning–based GOALS scoring system in wet lab training, with an error of approximately 1–2 points for the total score, and motion metrics that were explainable to trainees. Our future challenges are the further improvement of onsite GOALS feedback, exploring the educational benefit of our model and building an efficient training program. SUPPLEMENTARY INFORMATION: The online version contains supplementary material available at 10.1007/s00423-022-02505-9. Springer Berlin Heidelberg 2022-04-08 2022 /pmc/articles/PMC9399206/ /pubmed/35394212 http://dx.doi.org/10.1007/s00423-022-02505-9 Text en © The Author(s) 2022 https://creativecommons.org/licenses/by/4.0/Open AccessThis article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) . |
spellingShingle | Original Article Ebina, Koki Abe, Takashige Hotta, Kiyohiko Higuchi, Madoka Furumido, Jun Iwahara, Naoya Kon, Masafumi Miyaji, Kou Shibuya, Sayaka Lingbo, Yan Komizunai, Shunsuke Kurashima, Yo Kikuchi, Hiroshi Matsumoto, Ryuji Osawa, Takahiro Murai, Sachiyo Tsujita, Teppei Sase, Kazuya Chen, Xiaoshuai Konno, Atsushi Shinohara, Nobuo Objective evaluation of laparoscopic surgical skills in wet lab training based on motion analysis and machine learning |
title | Objective evaluation of laparoscopic surgical skills in wet lab training based on motion analysis and machine learning |
title_full | Objective evaluation of laparoscopic surgical skills in wet lab training based on motion analysis and machine learning |
title_fullStr | Objective evaluation of laparoscopic surgical skills in wet lab training based on motion analysis and machine learning |
title_full_unstemmed | Objective evaluation of laparoscopic surgical skills in wet lab training based on motion analysis and machine learning |
title_short | Objective evaluation of laparoscopic surgical skills in wet lab training based on motion analysis and machine learning |
title_sort | objective evaluation of laparoscopic surgical skills in wet lab training based on motion analysis and machine learning |
topic | Original Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9399206/ https://www.ncbi.nlm.nih.gov/pubmed/35394212 http://dx.doi.org/10.1007/s00423-022-02505-9 |
work_keys_str_mv | AT ebinakoki objectiveevaluationoflaparoscopicsurgicalskillsinwetlabtrainingbasedonmotionanalysisandmachinelearning AT abetakashige objectiveevaluationoflaparoscopicsurgicalskillsinwetlabtrainingbasedonmotionanalysisandmachinelearning AT hottakiyohiko objectiveevaluationoflaparoscopicsurgicalskillsinwetlabtrainingbasedonmotionanalysisandmachinelearning AT higuchimadoka objectiveevaluationoflaparoscopicsurgicalskillsinwetlabtrainingbasedonmotionanalysisandmachinelearning AT furumidojun objectiveevaluationoflaparoscopicsurgicalskillsinwetlabtrainingbasedonmotionanalysisandmachinelearning AT iwaharanaoya objectiveevaluationoflaparoscopicsurgicalskillsinwetlabtrainingbasedonmotionanalysisandmachinelearning AT konmasafumi objectiveevaluationoflaparoscopicsurgicalskillsinwetlabtrainingbasedonmotionanalysisandmachinelearning AT miyajikou objectiveevaluationoflaparoscopicsurgicalskillsinwetlabtrainingbasedonmotionanalysisandmachinelearning AT shibuyasayaka objectiveevaluationoflaparoscopicsurgicalskillsinwetlabtrainingbasedonmotionanalysisandmachinelearning AT lingboyan objectiveevaluationoflaparoscopicsurgicalskillsinwetlabtrainingbasedonmotionanalysisandmachinelearning AT komizunaishunsuke objectiveevaluationoflaparoscopicsurgicalskillsinwetlabtrainingbasedonmotionanalysisandmachinelearning AT kurashimayo objectiveevaluationoflaparoscopicsurgicalskillsinwetlabtrainingbasedonmotionanalysisandmachinelearning AT kikuchihiroshi objectiveevaluationoflaparoscopicsurgicalskillsinwetlabtrainingbasedonmotionanalysisandmachinelearning AT matsumotoryuji objectiveevaluationoflaparoscopicsurgicalskillsinwetlabtrainingbasedonmotionanalysisandmachinelearning AT osawatakahiro objectiveevaluationoflaparoscopicsurgicalskillsinwetlabtrainingbasedonmotionanalysisandmachinelearning AT muraisachiyo objectiveevaluationoflaparoscopicsurgicalskillsinwetlabtrainingbasedonmotionanalysisandmachinelearning AT tsujitateppei objectiveevaluationoflaparoscopicsurgicalskillsinwetlabtrainingbasedonmotionanalysisandmachinelearning AT sasekazuya objectiveevaluationoflaparoscopicsurgicalskillsinwetlabtrainingbasedonmotionanalysisandmachinelearning AT chenxiaoshuai objectiveevaluationoflaparoscopicsurgicalskillsinwetlabtrainingbasedonmotionanalysisandmachinelearning AT konnoatsushi objectiveevaluationoflaparoscopicsurgicalskillsinwetlabtrainingbasedonmotionanalysisandmachinelearning AT shinoharanobuo objectiveevaluationoflaparoscopicsurgicalskillsinwetlabtrainingbasedonmotionanalysisandmachinelearning |