Cargando…

Active learning for extracting surgomic features in robot-assisted minimally invasive esophagectomy: a prospective annotation study

BACKGROUND: With Surgomics, we aim for personalized prediction of the patient's surgical outcome using machine-learning (ML) on multimodal intraoperative data to extract surgomic features as surgical process characteristics. As high-quality annotations by medical experts are crucial, but still...

Descripción completa

Detalles Bibliográficos
Autores principales: Brandenburg, Johanna M., Jenke, Alexander C., Stern, Antonia, Daum, Marie T. J., Schulze, André, Younis, Rayan, Petrynowski, Philipp, Davitashvili, Tornike, Vanat, Vincent, Bhasker, Nithya, Schneider, Sophia, Mündermann, Lars, Reinke, Annika, Kolbinger, Fiona R., Jörns, Vanessa, Fritz-Kebede, Fleur, Dugas, Martin, Maier-Hein, Lena, Klotz, Rosa, Distler, Marius, Weitz, Jürgen, Müller-Stich, Beat P., Speidel, Stefanie, Bodenstedt, Sebastian, Wagner, Martin
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Springer US 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10615926/
https://www.ncbi.nlm.nih.gov/pubmed/37833509
http://dx.doi.org/10.1007/s00464-023-10447-6
_version_ 1785129287736623104
author Brandenburg, Johanna M.
Jenke, Alexander C.
Stern, Antonia
Daum, Marie T. J.
Schulze, André
Younis, Rayan
Petrynowski, Philipp
Davitashvili, Tornike
Vanat, Vincent
Bhasker, Nithya
Schneider, Sophia
Mündermann, Lars
Reinke, Annika
Kolbinger, Fiona R.
Jörns, Vanessa
Fritz-Kebede, Fleur
Dugas, Martin
Maier-Hein, Lena
Klotz, Rosa
Distler, Marius
Weitz, Jürgen
Müller-Stich, Beat P.
Speidel, Stefanie
Bodenstedt, Sebastian
Wagner, Martin
author_facet Brandenburg, Johanna M.
Jenke, Alexander C.
Stern, Antonia
Daum, Marie T. J.
Schulze, André
Younis, Rayan
Petrynowski, Philipp
Davitashvili, Tornike
Vanat, Vincent
Bhasker, Nithya
Schneider, Sophia
Mündermann, Lars
Reinke, Annika
Kolbinger, Fiona R.
Jörns, Vanessa
Fritz-Kebede, Fleur
Dugas, Martin
Maier-Hein, Lena
Klotz, Rosa
Distler, Marius
Weitz, Jürgen
Müller-Stich, Beat P.
Speidel, Stefanie
Bodenstedt, Sebastian
Wagner, Martin
author_sort Brandenburg, Johanna M.
collection PubMed
description BACKGROUND: With Surgomics, we aim for personalized prediction of the patient's surgical outcome using machine-learning (ML) on multimodal intraoperative data to extract surgomic features as surgical process characteristics. As high-quality annotations by medical experts are crucial, but still a bottleneck, we prospectively investigate active learning (AL) to reduce annotation effort and present automatic recognition of surgomic features. METHODS: To establish a process for development of surgomic features, ten video-based features related to bleeding, as highly relevant intraoperative complication, were chosen. They comprise the amount of blood and smoke in the surgical field, six instruments, and two anatomic structures. Annotation of selected frames from robot-assisted minimally invasive esophagectomies was performed by at least three independent medical experts. To test whether AL reduces annotation effort, we performed a prospective annotation study comparing AL with equidistant sampling (EQS) for frame selection. Multiple Bayesian ResNet18 architectures were trained on a multicentric dataset, consisting of 22 videos from two centers. RESULTS: In total, 14,004 frames were tag annotated. A mean F1-score of 0.75 ± 0.16 was achieved for all features. The highest F1-score was achieved for the instruments (mean 0.80 ± 0.17). This result is also reflected in the inter-rater-agreement (1-rater-kappa > 0.82). Compared to EQS, AL showed better recognition results for the instruments with a significant difference in the McNemar test comparing correctness of predictions. Moreover, in contrast to EQS, AL selected more frames of the four less common instruments (1512 vs. 607 frames) and achieved higher F1-scores for common instruments while requiring less training frames. CONCLUSION: We presented ten surgomic features relevant for bleeding events in esophageal surgery automatically extracted from surgical video using ML. AL showed the potential to reduce annotation effort while keeping ML performance high for selected features. The source code and the trained models are published open source. GRAPHICAL ABSTRACT: [Image: see text] SUPPLEMENTARY INFORMATION: The online version contains supplementary material available at 10.1007/s00464-023-10447-6.
format Online
Article
Text
id pubmed-10615926
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Springer US
record_format MEDLINE/PubMed
spelling pubmed-106159262023-11-01 Active learning for extracting surgomic features in robot-assisted minimally invasive esophagectomy: a prospective annotation study Brandenburg, Johanna M. Jenke, Alexander C. Stern, Antonia Daum, Marie T. J. Schulze, André Younis, Rayan Petrynowski, Philipp Davitashvili, Tornike Vanat, Vincent Bhasker, Nithya Schneider, Sophia Mündermann, Lars Reinke, Annika Kolbinger, Fiona R. Jörns, Vanessa Fritz-Kebede, Fleur Dugas, Martin Maier-Hein, Lena Klotz, Rosa Distler, Marius Weitz, Jürgen Müller-Stich, Beat P. Speidel, Stefanie Bodenstedt, Sebastian Wagner, Martin Surg Endosc 2023 EAES Oral BACKGROUND: With Surgomics, we aim for personalized prediction of the patient's surgical outcome using machine-learning (ML) on multimodal intraoperative data to extract surgomic features as surgical process characteristics. As high-quality annotations by medical experts are crucial, but still a bottleneck, we prospectively investigate active learning (AL) to reduce annotation effort and present automatic recognition of surgomic features. METHODS: To establish a process for development of surgomic features, ten video-based features related to bleeding, as highly relevant intraoperative complication, were chosen. They comprise the amount of blood and smoke in the surgical field, six instruments, and two anatomic structures. Annotation of selected frames from robot-assisted minimally invasive esophagectomies was performed by at least three independent medical experts. To test whether AL reduces annotation effort, we performed a prospective annotation study comparing AL with equidistant sampling (EQS) for frame selection. Multiple Bayesian ResNet18 architectures were trained on a multicentric dataset, consisting of 22 videos from two centers. RESULTS: In total, 14,004 frames were tag annotated. A mean F1-score of 0.75 ± 0.16 was achieved for all features. The highest F1-score was achieved for the instruments (mean 0.80 ± 0.17). This result is also reflected in the inter-rater-agreement (1-rater-kappa > 0.82). Compared to EQS, AL showed better recognition results for the instruments with a significant difference in the McNemar test comparing correctness of predictions. Moreover, in contrast to EQS, AL selected more frames of the four less common instruments (1512 vs. 607 frames) and achieved higher F1-scores for common instruments while requiring less training frames. CONCLUSION: We presented ten surgomic features relevant for bleeding events in esophageal surgery automatically extracted from surgical video using ML. AL showed the potential to reduce annotation effort while keeping ML performance high for selected features. The source code and the trained models are published open source. GRAPHICAL ABSTRACT: [Image: see text] SUPPLEMENTARY INFORMATION: The online version contains supplementary material available at 10.1007/s00464-023-10447-6. Springer US 2023-10-14 2023 /pmc/articles/PMC10615926/ /pubmed/37833509 http://dx.doi.org/10.1007/s00464-023-10447-6 Text en © The Author(s) 2023 https://creativecommons.org/licenses/by/4.0/Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) .
spellingShingle 2023 EAES Oral
Brandenburg, Johanna M.
Jenke, Alexander C.
Stern, Antonia
Daum, Marie T. J.
Schulze, André
Younis, Rayan
Petrynowski, Philipp
Davitashvili, Tornike
Vanat, Vincent
Bhasker, Nithya
Schneider, Sophia
Mündermann, Lars
Reinke, Annika
Kolbinger, Fiona R.
Jörns, Vanessa
Fritz-Kebede, Fleur
Dugas, Martin
Maier-Hein, Lena
Klotz, Rosa
Distler, Marius
Weitz, Jürgen
Müller-Stich, Beat P.
Speidel, Stefanie
Bodenstedt, Sebastian
Wagner, Martin
Active learning for extracting surgomic features in robot-assisted minimally invasive esophagectomy: a prospective annotation study
title Active learning for extracting surgomic features in robot-assisted minimally invasive esophagectomy: a prospective annotation study
title_full Active learning for extracting surgomic features in robot-assisted minimally invasive esophagectomy: a prospective annotation study
title_fullStr Active learning for extracting surgomic features in robot-assisted minimally invasive esophagectomy: a prospective annotation study
title_full_unstemmed Active learning for extracting surgomic features in robot-assisted minimally invasive esophagectomy: a prospective annotation study
title_short Active learning for extracting surgomic features in robot-assisted minimally invasive esophagectomy: a prospective annotation study
title_sort active learning for extracting surgomic features in robot-assisted minimally invasive esophagectomy: a prospective annotation study
topic 2023 EAES Oral
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10615926/
https://www.ncbi.nlm.nih.gov/pubmed/37833509
http://dx.doi.org/10.1007/s00464-023-10447-6
work_keys_str_mv AT brandenburgjohannam activelearningforextractingsurgomicfeaturesinrobotassistedminimallyinvasiveesophagectomyaprospectiveannotationstudy
AT jenkealexanderc activelearningforextractingsurgomicfeaturesinrobotassistedminimallyinvasiveesophagectomyaprospectiveannotationstudy
AT sternantonia activelearningforextractingsurgomicfeaturesinrobotassistedminimallyinvasiveesophagectomyaprospectiveannotationstudy
AT daummarietj activelearningforextractingsurgomicfeaturesinrobotassistedminimallyinvasiveesophagectomyaprospectiveannotationstudy
AT schulzeandre activelearningforextractingsurgomicfeaturesinrobotassistedminimallyinvasiveesophagectomyaprospectiveannotationstudy
AT younisrayan activelearningforextractingsurgomicfeaturesinrobotassistedminimallyinvasiveesophagectomyaprospectiveannotationstudy
AT petrynowskiphilipp activelearningforextractingsurgomicfeaturesinrobotassistedminimallyinvasiveesophagectomyaprospectiveannotationstudy
AT davitashvilitornike activelearningforextractingsurgomicfeaturesinrobotassistedminimallyinvasiveesophagectomyaprospectiveannotationstudy
AT vanatvincent activelearningforextractingsurgomicfeaturesinrobotassistedminimallyinvasiveesophagectomyaprospectiveannotationstudy
AT bhaskernithya activelearningforextractingsurgomicfeaturesinrobotassistedminimallyinvasiveesophagectomyaprospectiveannotationstudy
AT schneidersophia activelearningforextractingsurgomicfeaturesinrobotassistedminimallyinvasiveesophagectomyaprospectiveannotationstudy
AT mundermannlars activelearningforextractingsurgomicfeaturesinrobotassistedminimallyinvasiveesophagectomyaprospectiveannotationstudy
AT reinkeannika activelearningforextractingsurgomicfeaturesinrobotassistedminimallyinvasiveesophagectomyaprospectiveannotationstudy
AT kolbingerfionar activelearningforextractingsurgomicfeaturesinrobotassistedminimallyinvasiveesophagectomyaprospectiveannotationstudy
AT jornsvanessa activelearningforextractingsurgomicfeaturesinrobotassistedminimallyinvasiveesophagectomyaprospectiveannotationstudy
AT fritzkebedefleur activelearningforextractingsurgomicfeaturesinrobotassistedminimallyinvasiveesophagectomyaprospectiveannotationstudy
AT dugasmartin activelearningforextractingsurgomicfeaturesinrobotassistedminimallyinvasiveesophagectomyaprospectiveannotationstudy
AT maierheinlena activelearningforextractingsurgomicfeaturesinrobotassistedminimallyinvasiveesophagectomyaprospectiveannotationstudy
AT klotzrosa activelearningforextractingsurgomicfeaturesinrobotassistedminimallyinvasiveesophagectomyaprospectiveannotationstudy
AT distlermarius activelearningforextractingsurgomicfeaturesinrobotassistedminimallyinvasiveesophagectomyaprospectiveannotationstudy
AT weitzjurgen activelearningforextractingsurgomicfeaturesinrobotassistedminimallyinvasiveesophagectomyaprospectiveannotationstudy
AT mullerstichbeatp activelearningforextractingsurgomicfeaturesinrobotassistedminimallyinvasiveesophagectomyaprospectiveannotationstudy
AT speidelstefanie activelearningforextractingsurgomicfeaturesinrobotassistedminimallyinvasiveesophagectomyaprospectiveannotationstudy
AT bodenstedtsebastian activelearningforextractingsurgomicfeaturesinrobotassistedminimallyinvasiveesophagectomyaprospectiveannotationstudy
AT wagnermartin activelearningforextractingsurgomicfeaturesinrobotassistedminimallyinvasiveesophagectomyaprospectiveannotationstudy