Cargando…

Vitreoretinal Surgical Instrument Tracking in Three Dimensions Using Deep Learning

PURPOSE: To evaluate the potential for artificial intelligence-based video analysis to determine surgical instrument characteristics when moving in the three-dimensional vitreous space. METHODS: We designed and manufactured a model eye in which we recorded choreographed videos of many surgical instr...

Descripción completa

Detalles Bibliográficos
Autores principales: Baldi, Pierre F., Abdelkarim, Sherif, Liu, Junze, To, Josiah K., Ibarra, Marialejandra Diaz, Browne, Andrew W.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: The Association for Research in Vision and Ophthalmology 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9851279/
https://www.ncbi.nlm.nih.gov/pubmed/36648414
http://dx.doi.org/10.1167/tvst.12.1.20
_version_ 1784872363841552384
author Baldi, Pierre F.
Abdelkarim, Sherif
Liu, Junze
To, Josiah K.
Ibarra, Marialejandra Diaz
Browne, Andrew W.
author_facet Baldi, Pierre F.
Abdelkarim, Sherif
Liu, Junze
To, Josiah K.
Ibarra, Marialejandra Diaz
Browne, Andrew W.
author_sort Baldi, Pierre F.
collection PubMed
description PURPOSE: To evaluate the potential for artificial intelligence-based video analysis to determine surgical instrument characteristics when moving in the three-dimensional vitreous space. METHODS: We designed and manufactured a model eye in which we recorded choreographed videos of many surgical instruments moving throughout the eye. We labeled each frame of the videos to describe the surgical tool characteristics: tool type, location, depth, and insertional laterality. We trained two different deep learning models to predict each of the tool characteristics and evaluated model performances on a subset of images. RESULTS: The accuracy of the classification model on the training set is 84% for the x–y region, 97% for depth, 100% for instrument type, and 100% for laterality of insertion. The accuracy of the classification model on the validation dataset is 83% for the x–y region, 96% for depth, 100% for instrument type, and 100% for laterality of insertion. The close-up detection model performs at 67 frames per second, with precision for most instruments higher than 75%, achieving a mean average precision of 79.3%. CONCLUSIONS: We demonstrated that trained models can track surgical instrument movement in three-dimensional space and determine instrument depth, tip location, instrument insertional laterality, and instrument type. Model performance is nearly instantaneous and justifies further investigation into application to real-world surgical videos. TRANSLATIONAL RELEVANCE: Deep learning offers the potential for software-based safety feedback mechanisms during surgery or the ability to extract metrics of surgical technique that can direct research to optimize surgical outcomes.
format Online
Article
Text
id pubmed-9851279
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher The Association for Research in Vision and Ophthalmology
record_format MEDLINE/PubMed
spelling pubmed-98512792023-01-20 Vitreoretinal Surgical Instrument Tracking in Three Dimensions Using Deep Learning Baldi, Pierre F. Abdelkarim, Sherif Liu, Junze To, Josiah K. Ibarra, Marialejandra Diaz Browne, Andrew W. Transl Vis Sci Technol Artificial Intelligence PURPOSE: To evaluate the potential for artificial intelligence-based video analysis to determine surgical instrument characteristics when moving in the three-dimensional vitreous space. METHODS: We designed and manufactured a model eye in which we recorded choreographed videos of many surgical instruments moving throughout the eye. We labeled each frame of the videos to describe the surgical tool characteristics: tool type, location, depth, and insertional laterality. We trained two different deep learning models to predict each of the tool characteristics and evaluated model performances on a subset of images. RESULTS: The accuracy of the classification model on the training set is 84% for the x–y region, 97% for depth, 100% for instrument type, and 100% for laterality of insertion. The accuracy of the classification model on the validation dataset is 83% for the x–y region, 96% for depth, 100% for instrument type, and 100% for laterality of insertion. The close-up detection model performs at 67 frames per second, with precision for most instruments higher than 75%, achieving a mean average precision of 79.3%. CONCLUSIONS: We demonstrated that trained models can track surgical instrument movement in three-dimensional space and determine instrument depth, tip location, instrument insertional laterality, and instrument type. Model performance is nearly instantaneous and justifies further investigation into application to real-world surgical videos. TRANSLATIONAL RELEVANCE: Deep learning offers the potential for software-based safety feedback mechanisms during surgery or the ability to extract metrics of surgical technique that can direct research to optimize surgical outcomes. The Association for Research in Vision and Ophthalmology 2023-01-17 /pmc/articles/PMC9851279/ /pubmed/36648414 http://dx.doi.org/10.1167/tvst.12.1.20 Text en Copyright 2023 The Authors https://creativecommons.org/licenses/by-nc-nd/4.0/This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.
spellingShingle Artificial Intelligence
Baldi, Pierre F.
Abdelkarim, Sherif
Liu, Junze
To, Josiah K.
Ibarra, Marialejandra Diaz
Browne, Andrew W.
Vitreoretinal Surgical Instrument Tracking in Three Dimensions Using Deep Learning
title Vitreoretinal Surgical Instrument Tracking in Three Dimensions Using Deep Learning
title_full Vitreoretinal Surgical Instrument Tracking in Three Dimensions Using Deep Learning
title_fullStr Vitreoretinal Surgical Instrument Tracking in Three Dimensions Using Deep Learning
title_full_unstemmed Vitreoretinal Surgical Instrument Tracking in Three Dimensions Using Deep Learning
title_short Vitreoretinal Surgical Instrument Tracking in Three Dimensions Using Deep Learning
title_sort vitreoretinal surgical instrument tracking in three dimensions using deep learning
topic Artificial Intelligence
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9851279/
https://www.ncbi.nlm.nih.gov/pubmed/36648414
http://dx.doi.org/10.1167/tvst.12.1.20
work_keys_str_mv AT baldipierref vitreoretinalsurgicalinstrumenttrackinginthreedimensionsusingdeeplearning
AT abdelkarimsherif vitreoretinalsurgicalinstrumenttrackinginthreedimensionsusingdeeplearning
AT liujunze vitreoretinalsurgicalinstrumenttrackinginthreedimensionsusingdeeplearning
AT tojosiahk vitreoretinalsurgicalinstrumenttrackinginthreedimensionsusingdeeplearning
AT ibarramarialejandradiaz vitreoretinalsurgicalinstrumenttrackinginthreedimensionsusingdeeplearning
AT browneandreww vitreoretinalsurgicalinstrumenttrackinginthreedimensionsusingdeeplearning