Cargando…
Vitreoretinal Surgical Instrument Tracking in Three Dimensions Using Deep Learning
PURPOSE: To evaluate the potential for artificial intelligence-based video analysis to determine surgical instrument characteristics when moving in the three-dimensional vitreous space. METHODS: We designed and manufactured a model eye in which we recorded choreographed videos of many surgical instr...
Autores principales: | , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
The Association for Research in Vision and Ophthalmology
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9851279/ https://www.ncbi.nlm.nih.gov/pubmed/36648414 http://dx.doi.org/10.1167/tvst.12.1.20 |
_version_ | 1784872363841552384 |
---|---|
author | Baldi, Pierre F. Abdelkarim, Sherif Liu, Junze To, Josiah K. Ibarra, Marialejandra Diaz Browne, Andrew W. |
author_facet | Baldi, Pierre F. Abdelkarim, Sherif Liu, Junze To, Josiah K. Ibarra, Marialejandra Diaz Browne, Andrew W. |
author_sort | Baldi, Pierre F. |
collection | PubMed |
description | PURPOSE: To evaluate the potential for artificial intelligence-based video analysis to determine surgical instrument characteristics when moving in the three-dimensional vitreous space. METHODS: We designed and manufactured a model eye in which we recorded choreographed videos of many surgical instruments moving throughout the eye. We labeled each frame of the videos to describe the surgical tool characteristics: tool type, location, depth, and insertional laterality. We trained two different deep learning models to predict each of the tool characteristics and evaluated model performances on a subset of images. RESULTS: The accuracy of the classification model on the training set is 84% for the x–y region, 97% for depth, 100% for instrument type, and 100% for laterality of insertion. The accuracy of the classification model on the validation dataset is 83% for the x–y region, 96% for depth, 100% for instrument type, and 100% for laterality of insertion. The close-up detection model performs at 67 frames per second, with precision for most instruments higher than 75%, achieving a mean average precision of 79.3%. CONCLUSIONS: We demonstrated that trained models can track surgical instrument movement in three-dimensional space and determine instrument depth, tip location, instrument insertional laterality, and instrument type. Model performance is nearly instantaneous and justifies further investigation into application to real-world surgical videos. TRANSLATIONAL RELEVANCE: Deep learning offers the potential for software-based safety feedback mechanisms during surgery or the ability to extract metrics of surgical technique that can direct research to optimize surgical outcomes. |
format | Online Article Text |
id | pubmed-9851279 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | The Association for Research in Vision and Ophthalmology |
record_format | MEDLINE/PubMed |
spelling | pubmed-98512792023-01-20 Vitreoretinal Surgical Instrument Tracking in Three Dimensions Using Deep Learning Baldi, Pierre F. Abdelkarim, Sherif Liu, Junze To, Josiah K. Ibarra, Marialejandra Diaz Browne, Andrew W. Transl Vis Sci Technol Artificial Intelligence PURPOSE: To evaluate the potential for artificial intelligence-based video analysis to determine surgical instrument characteristics when moving in the three-dimensional vitreous space. METHODS: We designed and manufactured a model eye in which we recorded choreographed videos of many surgical instruments moving throughout the eye. We labeled each frame of the videos to describe the surgical tool characteristics: tool type, location, depth, and insertional laterality. We trained two different deep learning models to predict each of the tool characteristics and evaluated model performances on a subset of images. RESULTS: The accuracy of the classification model on the training set is 84% for the x–y region, 97% for depth, 100% for instrument type, and 100% for laterality of insertion. The accuracy of the classification model on the validation dataset is 83% for the x–y region, 96% for depth, 100% for instrument type, and 100% for laterality of insertion. The close-up detection model performs at 67 frames per second, with precision for most instruments higher than 75%, achieving a mean average precision of 79.3%. CONCLUSIONS: We demonstrated that trained models can track surgical instrument movement in three-dimensional space and determine instrument depth, tip location, instrument insertional laterality, and instrument type. Model performance is nearly instantaneous and justifies further investigation into application to real-world surgical videos. TRANSLATIONAL RELEVANCE: Deep learning offers the potential for software-based safety feedback mechanisms during surgery or the ability to extract metrics of surgical technique that can direct research to optimize surgical outcomes. The Association for Research in Vision and Ophthalmology 2023-01-17 /pmc/articles/PMC9851279/ /pubmed/36648414 http://dx.doi.org/10.1167/tvst.12.1.20 Text en Copyright 2023 The Authors https://creativecommons.org/licenses/by-nc-nd/4.0/This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License. |
spellingShingle | Artificial Intelligence Baldi, Pierre F. Abdelkarim, Sherif Liu, Junze To, Josiah K. Ibarra, Marialejandra Diaz Browne, Andrew W. Vitreoretinal Surgical Instrument Tracking in Three Dimensions Using Deep Learning |
title | Vitreoretinal Surgical Instrument Tracking in Three Dimensions Using Deep Learning |
title_full | Vitreoretinal Surgical Instrument Tracking in Three Dimensions Using Deep Learning |
title_fullStr | Vitreoretinal Surgical Instrument Tracking in Three Dimensions Using Deep Learning |
title_full_unstemmed | Vitreoretinal Surgical Instrument Tracking in Three Dimensions Using Deep Learning |
title_short | Vitreoretinal Surgical Instrument Tracking in Three Dimensions Using Deep Learning |
title_sort | vitreoretinal surgical instrument tracking in three dimensions using deep learning |
topic | Artificial Intelligence |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9851279/ https://www.ncbi.nlm.nih.gov/pubmed/36648414 http://dx.doi.org/10.1167/tvst.12.1.20 |
work_keys_str_mv | AT baldipierref vitreoretinalsurgicalinstrumenttrackinginthreedimensionsusingdeeplearning AT abdelkarimsherif vitreoretinalsurgicalinstrumenttrackinginthreedimensionsusingdeeplearning AT liujunze vitreoretinalsurgicalinstrumenttrackinginthreedimensionsusingdeeplearning AT tojosiahk vitreoretinalsurgicalinstrumenttrackinginthreedimensionsusingdeeplearning AT ibarramarialejandradiaz vitreoretinalsurgicalinstrumenttrackinginthreedimensionsusingdeeplearning AT browneandreww vitreoretinalsurgicalinstrumenttrackinginthreedimensionsusingdeeplearning |