Cargando…

Vitreoretinal Surgical Instrument Tracking in Three Dimensions Using Deep Learning

PURPOSE: To evaluate the potential for artificial intelligence-based video analysis to determine surgical instrument characteristics when moving in the three-dimensional vitreous space. METHODS: We designed and manufactured a model eye in which we recorded choreographed videos of many surgical instr...

Descripción completa

Detalles Bibliográficos
Autores principales: Baldi, Pierre F., Abdelkarim, Sherif, Liu, Junze, To, Josiah K., Ibarra, Marialejandra Diaz, Browne, Andrew W.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: The Association for Research in Vision and Ophthalmology 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9851279/
https://www.ncbi.nlm.nih.gov/pubmed/36648414
http://dx.doi.org/10.1167/tvst.12.1.20
Descripción
Sumario:PURPOSE: To evaluate the potential for artificial intelligence-based video analysis to determine surgical instrument characteristics when moving in the three-dimensional vitreous space. METHODS: We designed and manufactured a model eye in which we recorded choreographed videos of many surgical instruments moving throughout the eye. We labeled each frame of the videos to describe the surgical tool characteristics: tool type, location, depth, and insertional laterality. We trained two different deep learning models to predict each of the tool characteristics and evaluated model performances on a subset of images. RESULTS: The accuracy of the classification model on the training set is 84% for the x–y region, 97% for depth, 100% for instrument type, and 100% for laterality of insertion. The accuracy of the classification model on the validation dataset is 83% for the x–y region, 96% for depth, 100% for instrument type, and 100% for laterality of insertion. The close-up detection model performs at 67 frames per second, with precision for most instruments higher than 75%, achieving a mean average precision of 79.3%. CONCLUSIONS: We demonstrated that trained models can track surgical instrument movement in three-dimensional space and determine instrument depth, tip location, instrument insertional laterality, and instrument type. Model performance is nearly instantaneous and justifies further investigation into application to real-world surgical videos. TRANSLATIONAL RELEVANCE: Deep learning offers the potential for software-based safety feedback mechanisms during surgery or the ability to extract metrics of surgical technique that can direct research to optimize surgical outcomes.