Cargando…

In-Person Verification of Deep Learning Algorithm for Diabetic Retinopathy Screening Using Different Techniques Across Fundus Image Devices

PURPOSE: To evaluate the clinical performance of an automated diabetic retinopathy (DR) screening model to detect referable cases at Siriraj Hospital, Bangkok, Thailand. METHODS: A retrospective review of two sets of fundus photographs (Eidon and Nidek) was undertaken. The images were classified by...

Descripción completa

Detalles Bibliográficos
Autores principales: Wongchaisuwat, Nida, Trinavarat, Adisak, Rodanant, Nuttawut, Thoongsuwan, Somanus, Phasukkijwatana, Nopasak, Prakhunhungsit, Supalert, Preechasuk, Lukana, Wongchaisuwat, Papis
Formato: Online Artículo Texto
Lenguaje:English
Publicado: The Association for Research in Vision and Ophthalmology 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8590162/
https://www.ncbi.nlm.nih.gov/pubmed/34767624
http://dx.doi.org/10.1167/tvst.10.13.17
Descripción
Sumario:PURPOSE: To evaluate the clinical performance of an automated diabetic retinopathy (DR) screening model to detect referable cases at Siriraj Hospital, Bangkok, Thailand. METHODS: A retrospective review of two sets of fundus photographs (Eidon and Nidek) was undertaken. The images were classified by DR staging prior to the development of a DR screening model. In a prospective cross-sectional enrollment of patients with diabetes, automated detection of referable DR was compared with the results of the gold standard, a dilated fundus examination. RESULTS: The study analyzed 2533 Nidek fundus images and 1989 Eidon images. The sensitivities calculated for the Nidek and Eidon images were 0.93 and 0.88 and the specificities were 0.91 and 0.85, respectively. In a clinical verification phase using 982 Nidek and 674 Eidon photographs, the calculated sensitivities and specificities were 0.86 and 0.92 for Nidek along with 0.92 and 0.84 for Eidon, respectively. The 60°-field images from the Eidon yielded a more desirable performance in differentiating referable DR than did the corresponding images from the Nidek. CONCLUSIONS: A conventional fundus examination requires intense healthcare resources. It is time consuming and possibly leads to unavoidable human errors. The deep learning algorithm for the detection of referable DR exhibited a favorable performance and is a promising alternative for DR screening. However, variations in the color and pixels of photographs can cause differences in sensitivity and specificity. The image angle and poor quality of fundus photographs were the main limitations of the automated method. TRANSLATIONAL RELEVANCE: The deep learning algorithm, developed from basic research of image processing, was applied to detect referable DR in a real-word clinical care setting.