Cargando…

Multicenter, Head-to-Head, Real-World Validation Study of Seven Automated Artificial Intelligence Diabetic Retinopathy Screening Systems

OBJECTIVE: With rising global prevalence of diabetic retinopathy (DR), automated DR screening is needed for primary care settings. Two automated artificial intelligence (AI)–based DR screening algorithms have U.S. Food and Drug Administration (FDA) approval. Several others are under consideration wh...

Descripción completa

Detalles Bibliográficos
Autores principales: Lee, Aaron Y., Yanagihara, Ryan T., Lee, Cecilia S., Blazes, Marian, Jung, Hoon C., Chee, Yewlin E., Gencarella, Michael D., Gee, Harry, Maa, April Y., Cockerham, Glenn C., Lynch, Mary, Boyko, Edward J.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: American Diabetes Association 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8132324/
https://www.ncbi.nlm.nih.gov/pubmed/33402366
http://dx.doi.org/10.2337/dc20-1877
_version_ 1783694891808718848
author Lee, Aaron Y.
Yanagihara, Ryan T.
Lee, Cecilia S.
Blazes, Marian
Jung, Hoon C.
Chee, Yewlin E.
Gencarella, Michael D.
Gee, Harry
Maa, April Y.
Cockerham, Glenn C.
Lynch, Mary
Boyko, Edward J.
author_facet Lee, Aaron Y.
Yanagihara, Ryan T.
Lee, Cecilia S.
Blazes, Marian
Jung, Hoon C.
Chee, Yewlin E.
Gencarella, Michael D.
Gee, Harry
Maa, April Y.
Cockerham, Glenn C.
Lynch, Mary
Boyko, Edward J.
author_sort Lee, Aaron Y.
collection PubMed
description OBJECTIVE: With rising global prevalence of diabetic retinopathy (DR), automated DR screening is needed for primary care settings. Two automated artificial intelligence (AI)–based DR screening algorithms have U.S. Food and Drug Administration (FDA) approval. Several others are under consideration while in clinical use in other countries, but their real-world performance has not been evaluated systematically. We compared the performance of seven automated AI-based DR screening algorithms (including one FDA-approved algorithm) against human graders when analyzing real-world retinal imaging data. RESEARCH DESIGN AND METHODS: This was a multicenter, noninterventional device validation study evaluating a total of 311,604 retinal images from 23,724 veterans who presented for teleretinal DR screening at the Veterans Affairs (VA) Puget Sound Health Care System (HCS) or Atlanta VA HCS from 2006 to 2018. Five companies provided seven algorithms, including one with FDA approval, that independently analyzed all scans, regardless of image quality. The sensitivity/specificity of each algorithm when classifying images as referable DR or not were compared with original VA teleretinal grades and a regraded arbitrated data set. Value per encounter was estimated. RESULTS: Although high negative predictive values (82.72–93.69%) were observed, sensitivities varied widely (50.98–85.90%). Most algorithms performed no better than humans against the arbitrated data set, but two achieved higher sensitivities, and one yielded comparable sensitivity (80.47%, P = 0.441) and specificity (81.28%, P = 0.195). Notably, one had lower sensitivity (74.42%) for proliferative DR (P = 9.77 × 10(−4)) than the VA teleretinal graders. Value per encounter varied at $15.14–$18.06 for ophthalmologists and $7.74–$9.24 for optometrists. CONCLUSIONS: The DR screening algorithms showed significant performance differences. These results argue for rigorous testing of all such algorithms on real-world data before clinical implementation.
format Online
Article
Text
id pubmed-8132324
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher American Diabetes Association
record_format MEDLINE/PubMed
spelling pubmed-81323242021-07-20 Multicenter, Head-to-Head, Real-World Validation Study of Seven Automated Artificial Intelligence Diabetic Retinopathy Screening Systems Lee, Aaron Y. Yanagihara, Ryan T. Lee, Cecilia S. Blazes, Marian Jung, Hoon C. Chee, Yewlin E. Gencarella, Michael D. Gee, Harry Maa, April Y. Cockerham, Glenn C. Lynch, Mary Boyko, Edward J. Diabetes Care Emerging Technologies: Data Systems and Devices OBJECTIVE: With rising global prevalence of diabetic retinopathy (DR), automated DR screening is needed for primary care settings. Two automated artificial intelligence (AI)–based DR screening algorithms have U.S. Food and Drug Administration (FDA) approval. Several others are under consideration while in clinical use in other countries, but their real-world performance has not been evaluated systematically. We compared the performance of seven automated AI-based DR screening algorithms (including one FDA-approved algorithm) against human graders when analyzing real-world retinal imaging data. RESEARCH DESIGN AND METHODS: This was a multicenter, noninterventional device validation study evaluating a total of 311,604 retinal images from 23,724 veterans who presented for teleretinal DR screening at the Veterans Affairs (VA) Puget Sound Health Care System (HCS) or Atlanta VA HCS from 2006 to 2018. Five companies provided seven algorithms, including one with FDA approval, that independently analyzed all scans, regardless of image quality. The sensitivity/specificity of each algorithm when classifying images as referable DR or not were compared with original VA teleretinal grades and a regraded arbitrated data set. Value per encounter was estimated. RESULTS: Although high negative predictive values (82.72–93.69%) were observed, sensitivities varied widely (50.98–85.90%). Most algorithms performed no better than humans against the arbitrated data set, but two achieved higher sensitivities, and one yielded comparable sensitivity (80.47%, P = 0.441) and specificity (81.28%, P = 0.195). Notably, one had lower sensitivity (74.42%) for proliferative DR (P = 9.77 × 10(−4)) than the VA teleretinal graders. Value per encounter varied at $15.14–$18.06 for ophthalmologists and $7.74–$9.24 for optometrists. CONCLUSIONS: The DR screening algorithms showed significant performance differences. These results argue for rigorous testing of all such algorithms on real-world data before clinical implementation. American Diabetes Association 2021-05 2021-01-05 /pmc/articles/PMC8132324/ /pubmed/33402366 http://dx.doi.org/10.2337/dc20-1877 Text en © 2021 by the American Diabetes Association https://www.diabetesjournals.org/content/licenseReaders may use this article as long as the work is properly cited, the use is educational and not for profit, and the work is not altered. More information is available at https://www.diabetesjournals.org/content/license.
spellingShingle Emerging Technologies: Data Systems and Devices
Lee, Aaron Y.
Yanagihara, Ryan T.
Lee, Cecilia S.
Blazes, Marian
Jung, Hoon C.
Chee, Yewlin E.
Gencarella, Michael D.
Gee, Harry
Maa, April Y.
Cockerham, Glenn C.
Lynch, Mary
Boyko, Edward J.
Multicenter, Head-to-Head, Real-World Validation Study of Seven Automated Artificial Intelligence Diabetic Retinopathy Screening Systems
title Multicenter, Head-to-Head, Real-World Validation Study of Seven Automated Artificial Intelligence Diabetic Retinopathy Screening Systems
title_full Multicenter, Head-to-Head, Real-World Validation Study of Seven Automated Artificial Intelligence Diabetic Retinopathy Screening Systems
title_fullStr Multicenter, Head-to-Head, Real-World Validation Study of Seven Automated Artificial Intelligence Diabetic Retinopathy Screening Systems
title_full_unstemmed Multicenter, Head-to-Head, Real-World Validation Study of Seven Automated Artificial Intelligence Diabetic Retinopathy Screening Systems
title_short Multicenter, Head-to-Head, Real-World Validation Study of Seven Automated Artificial Intelligence Diabetic Retinopathy Screening Systems
title_sort multicenter, head-to-head, real-world validation study of seven automated artificial intelligence diabetic retinopathy screening systems
topic Emerging Technologies: Data Systems and Devices
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8132324/
https://www.ncbi.nlm.nih.gov/pubmed/33402366
http://dx.doi.org/10.2337/dc20-1877
work_keys_str_mv AT leeaarony multicenterheadtoheadrealworldvalidationstudyofsevenautomatedartificialintelligencediabeticretinopathyscreeningsystems
AT yanagihararyant multicenterheadtoheadrealworldvalidationstudyofsevenautomatedartificialintelligencediabeticretinopathyscreeningsystems
AT leececilias multicenterheadtoheadrealworldvalidationstudyofsevenautomatedartificialintelligencediabeticretinopathyscreeningsystems
AT blazesmarian multicenterheadtoheadrealworldvalidationstudyofsevenautomatedartificialintelligencediabeticretinopathyscreeningsystems
AT junghoonc multicenterheadtoheadrealworldvalidationstudyofsevenautomatedartificialintelligencediabeticretinopathyscreeningsystems
AT cheeyewline multicenterheadtoheadrealworldvalidationstudyofsevenautomatedartificialintelligencediabeticretinopathyscreeningsystems
AT gencarellamichaeld multicenterheadtoheadrealworldvalidationstudyofsevenautomatedartificialintelligencediabeticretinopathyscreeningsystems
AT geeharry multicenterheadtoheadrealworldvalidationstudyofsevenautomatedartificialintelligencediabeticretinopathyscreeningsystems
AT maaaprily multicenterheadtoheadrealworldvalidationstudyofsevenautomatedartificialintelligencediabeticretinopathyscreeningsystems
AT cockerhamglennc multicenterheadtoheadrealworldvalidationstudyofsevenautomatedartificialintelligencediabeticretinopathyscreeningsystems
AT lynchmary multicenterheadtoheadrealworldvalidationstudyofsevenautomatedartificialintelligencediabeticretinopathyscreeningsystems
AT boykoedwardj multicenterheadtoheadrealworldvalidationstudyofsevenautomatedartificialintelligencediabeticretinopathyscreeningsystems