Cargando…

Multilevel effective surgical workflow recognition in robotic left lateral sectionectomy with deep learning: experimental research

BACKGROUND: Automated surgical workflow recognition is the foundation for computational models of medical knowledge to interpret surgical procedures. The fine-grained segmentation of the surgical process and the improvement of the accuracy of surgical workflow recognition facilitate the realization...

Descripción completa

Detalles Bibliográficos
Autores principales: Liu, Yanzhe, Zhao, Shang, Zhang, Gong, Zhang, Xiuping, Hu, Minggen, Zhang, Xuan, Li, Chenggang, Zhou, S. Kevin, Liu, Rong
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Lippincott Williams & Wilkins 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10583914/
https://www.ncbi.nlm.nih.gov/pubmed/37318860
http://dx.doi.org/10.1097/JS9.0000000000000559
_version_ 1785122647683629056
author Liu, Yanzhe
Zhao, Shang
Zhang, Gong
Zhang, Xiuping
Hu, Minggen
Zhang, Xuan
Li, Chenggang
Zhou, S. Kevin
Liu, Rong
author_facet Liu, Yanzhe
Zhao, Shang
Zhang, Gong
Zhang, Xiuping
Hu, Minggen
Zhang, Xuan
Li, Chenggang
Zhou, S. Kevin
Liu, Rong
author_sort Liu, Yanzhe
collection PubMed
description BACKGROUND: Automated surgical workflow recognition is the foundation for computational models of medical knowledge to interpret surgical procedures. The fine-grained segmentation of the surgical process and the improvement of the accuracy of surgical workflow recognition facilitate the realization of autonomous robotic surgery. This study aimed to construct a multigranularity temporal annotation dataset of the standardized robotic left lateral sectionectomy (RLLS) and develop a deep learning-based automated model for multilevel overall and effective surgical workflow recognition. METHODS: From December 2016 to May 2019, 45 cases of RLLS videos were enrolled in our dataset. All frames of RLLS videos in this study are labeled with temporal annotations. The authors defined those activities that truly contribute to the surgery as effective frames, while other activities are labeled as under-effective frames. Effective frames of all RLLS videos are annotated with three hierarchical levels of 4 steps, 12 tasks, and 26 activities. A hybrid deep learning model were used for surgical workflow recognition of steps, tasks, activities, and under-effective frames. Moreover, the authors also carried out multilevel effective surgical workflow recognition after removing under-effective frames. RESULTS: The dataset comprises 4 383 516 annotated RLLS video frames with multilevel annotation, of which 2 418 468 frames are effective. The overall accuracies of automated recognition for Steps, Tasks, Activities, and under-effective frames are 0.82, 0.80, 0.79, and 0.85, respectively, with corresponding precision values of 0.81, 0.76, 0.60, and 0.85. In multilevel effective surgical workflow recognition, the overall accuracies were increased to 0.96, 0.88, and 0.82 for Steps, Tasks, and Activities, respectively, while the precision values were increased to 0.95, 0.80, and 0.68. CONCLUSION: In this study, the authors created a dataset of 45 RLLS cases with multilevel annotations and developed a hybrid deep learning model for surgical workflow recognition. The authors demonstrated a fairly higher accuracy in multilevel effective surgical workflow recognition when under-effective frames were removed. Our research could be helpful in the development of autonomous robotic surgery.
format Online
Article
Text
id pubmed-10583914
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Lippincott Williams & Wilkins
record_format MEDLINE/PubMed
spelling pubmed-105839142023-10-19 Multilevel effective surgical workflow recognition in robotic left lateral sectionectomy with deep learning: experimental research Liu, Yanzhe Zhao, Shang Zhang, Gong Zhang, Xiuping Hu, Minggen Zhang, Xuan Li, Chenggang Zhou, S. Kevin Liu, Rong Int J Surg Original Research BACKGROUND: Automated surgical workflow recognition is the foundation for computational models of medical knowledge to interpret surgical procedures. The fine-grained segmentation of the surgical process and the improvement of the accuracy of surgical workflow recognition facilitate the realization of autonomous robotic surgery. This study aimed to construct a multigranularity temporal annotation dataset of the standardized robotic left lateral sectionectomy (RLLS) and develop a deep learning-based automated model for multilevel overall and effective surgical workflow recognition. METHODS: From December 2016 to May 2019, 45 cases of RLLS videos were enrolled in our dataset. All frames of RLLS videos in this study are labeled with temporal annotations. The authors defined those activities that truly contribute to the surgery as effective frames, while other activities are labeled as under-effective frames. Effective frames of all RLLS videos are annotated with three hierarchical levels of 4 steps, 12 tasks, and 26 activities. A hybrid deep learning model were used for surgical workflow recognition of steps, tasks, activities, and under-effective frames. Moreover, the authors also carried out multilevel effective surgical workflow recognition after removing under-effective frames. RESULTS: The dataset comprises 4 383 516 annotated RLLS video frames with multilevel annotation, of which 2 418 468 frames are effective. The overall accuracies of automated recognition for Steps, Tasks, Activities, and under-effective frames are 0.82, 0.80, 0.79, and 0.85, respectively, with corresponding precision values of 0.81, 0.76, 0.60, and 0.85. In multilevel effective surgical workflow recognition, the overall accuracies were increased to 0.96, 0.88, and 0.82 for Steps, Tasks, and Activities, respectively, while the precision values were increased to 0.95, 0.80, and 0.68. CONCLUSION: In this study, the authors created a dataset of 45 RLLS cases with multilevel annotations and developed a hybrid deep learning model for surgical workflow recognition. The authors demonstrated a fairly higher accuracy in multilevel effective surgical workflow recognition when under-effective frames were removed. Our research could be helpful in the development of autonomous robotic surgery. Lippincott Williams & Wilkins 2023-06-14 /pmc/articles/PMC10583914/ /pubmed/37318860 http://dx.doi.org/10.1097/JS9.0000000000000559 Text en Copyright © 2023 The Author(s). Published by Wolters Kluwer Health, Inc. https://creativecommons.org/licenses/by-sa/4.0/This is an open access article distributed under the Creative Commons Attribution-ShareAlike License 4.0 (https://creativecommons.org/licenses/by-sa/4.0/) , which allows others to remix, tweak, and build upon the work, even for commercial purposes, as long as the author is credited and the new creations are licensed under the identical terms. http://creativecommons.org/licenses/by-sa/4.0/ (https://creativecommons.org/licenses/by-sa/4.0/)
spellingShingle Original Research
Liu, Yanzhe
Zhao, Shang
Zhang, Gong
Zhang, Xiuping
Hu, Minggen
Zhang, Xuan
Li, Chenggang
Zhou, S. Kevin
Liu, Rong
Multilevel effective surgical workflow recognition in robotic left lateral sectionectomy with deep learning: experimental research
title Multilevel effective surgical workflow recognition in robotic left lateral sectionectomy with deep learning: experimental research
title_full Multilevel effective surgical workflow recognition in robotic left lateral sectionectomy with deep learning: experimental research
title_fullStr Multilevel effective surgical workflow recognition in robotic left lateral sectionectomy with deep learning: experimental research
title_full_unstemmed Multilevel effective surgical workflow recognition in robotic left lateral sectionectomy with deep learning: experimental research
title_short Multilevel effective surgical workflow recognition in robotic left lateral sectionectomy with deep learning: experimental research
title_sort multilevel effective surgical workflow recognition in robotic left lateral sectionectomy with deep learning: experimental research
topic Original Research
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10583914/
https://www.ncbi.nlm.nih.gov/pubmed/37318860
http://dx.doi.org/10.1097/JS9.0000000000000559
work_keys_str_mv AT liuyanzhe multileveleffectivesurgicalworkflowrecognitioninroboticleftlateralsectionectomywithdeeplearningexperimentalresearch
AT zhaoshang multileveleffectivesurgicalworkflowrecognitioninroboticleftlateralsectionectomywithdeeplearningexperimentalresearch
AT zhanggong multileveleffectivesurgicalworkflowrecognitioninroboticleftlateralsectionectomywithdeeplearningexperimentalresearch
AT zhangxiuping multileveleffectivesurgicalworkflowrecognitioninroboticleftlateralsectionectomywithdeeplearningexperimentalresearch
AT huminggen multileveleffectivesurgicalworkflowrecognitioninroboticleftlateralsectionectomywithdeeplearningexperimentalresearch
AT zhangxuan multileveleffectivesurgicalworkflowrecognitioninroboticleftlateralsectionectomywithdeeplearningexperimentalresearch
AT lichenggang multileveleffectivesurgicalworkflowrecognitioninroboticleftlateralsectionectomywithdeeplearningexperimentalresearch
AT zhouskevin multileveleffectivesurgicalworkflowrecognitioninroboticleftlateralsectionectomywithdeeplearningexperimentalresearch
AT liurong multileveleffectivesurgicalworkflowrecognitioninroboticleftlateralsectionectomywithdeeplearningexperimentalresearch