Cargando…

From 2 dimensions to 3(rd) dimension: Quantitative prediction of anterior chamber depth from anterior segment photographs via deep-learning

Anterior chamber depth (ACD) is a major risk factor of angle closure disease, and has been used in angle closure screening in various populations. However, ACD is measured from ocular biometer or anterior segment optical coherence tomography (AS-OCT), which are costly and may not be readily availabl...

Descripción completa

Detalles Bibliográficos
Autores principales: Soh, Zhi Da, Jiang, Yixing, S/O Ganesan, Sakthi Selvam, Zhou, Menghan, Nongiur, Monisha, Majithia, Shivani, Tham, Yih Chung, Rim, Tyler Hyungtaek, Qian, Chaoxu, Koh, Victor, Aung, Tin, Wong, Tien Yin, Xu, Xinxing, Liu, Yong, Cheng, Ching-Yu
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Public Library of Science 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9931242/
https://www.ncbi.nlm.nih.gov/pubmed/36812642
http://dx.doi.org/10.1371/journal.pdig.0000193
_version_ 1784889205655076864
author Soh, Zhi Da
Jiang, Yixing
S/O Ganesan, Sakthi Selvam
Zhou, Menghan
Nongiur, Monisha
Majithia, Shivani
Tham, Yih Chung
Rim, Tyler Hyungtaek
Qian, Chaoxu
Koh, Victor
Aung, Tin
Wong, Tien Yin
Xu, Xinxing
Liu, Yong
Cheng, Ching-Yu
author_facet Soh, Zhi Da
Jiang, Yixing
S/O Ganesan, Sakthi Selvam
Zhou, Menghan
Nongiur, Monisha
Majithia, Shivani
Tham, Yih Chung
Rim, Tyler Hyungtaek
Qian, Chaoxu
Koh, Victor
Aung, Tin
Wong, Tien Yin
Xu, Xinxing
Liu, Yong
Cheng, Ching-Yu
author_sort Soh, Zhi Da
collection PubMed
description Anterior chamber depth (ACD) is a major risk factor of angle closure disease, and has been used in angle closure screening in various populations. However, ACD is measured from ocular biometer or anterior segment optical coherence tomography (AS-OCT), which are costly and may not be readily available in primary care and community settings. Thus, this proof-of-concept study aims to predict ACD from low-cost anterior segment photographs (ASPs) using deep-learning (DL). We included 2,311 pairs of ASPs and ACD measurements for algorithm development and validation, and 380 pairs for algorithm testing. We captured ASPs with a digital camera mounted on a slit-lamp biomicroscope. Anterior chamber depth was measured with ocular biometer (IOLMaster700 or Lenstar LS9000) in data used for algorithm development and validation, and with AS-OCT (Visante) in data used for testing. The DL algorithm was modified from the ResNet-50 architecture, and assessed using mean absolute error (MAE), coefficient-of-determination (R(2)), Bland-Altman plot and intraclass correlation coefficients (ICC). In validation, our algorithm predicted ACD with a MAE (standard deviation) of 0.18 (0.14) mm; R(2) = 0.63. The MAE of predicted ACD was 0.18 (0.14) mm in eyes with open angles and 0.19 (0.14) mm in eyes with angle closure. The ICC between actual and predicted ACD measurements was 0.81 (95% CI 0.77, 0.84). In testing, our algorithm predicted ACD with a MAE of 0.23 (0.18) mm; R(2) = 0.37. Saliency maps highlighted the pupil and its margin as the main structures used in ACD prediction. This study demonstrates the possibility of predicting ACD from ASPs via DL. This algorithm mimics an ocular biometer in making its prediction, and provides a foundation to predict other quantitative measurements that are relevant to angle closure screening.
format Online
Article
Text
id pubmed-9931242
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Public Library of Science
record_format MEDLINE/PubMed
spelling pubmed-99312422023-02-16 From 2 dimensions to 3(rd) dimension: Quantitative prediction of anterior chamber depth from anterior segment photographs via deep-learning Soh, Zhi Da Jiang, Yixing S/O Ganesan, Sakthi Selvam Zhou, Menghan Nongiur, Monisha Majithia, Shivani Tham, Yih Chung Rim, Tyler Hyungtaek Qian, Chaoxu Koh, Victor Aung, Tin Wong, Tien Yin Xu, Xinxing Liu, Yong Cheng, Ching-Yu PLOS Digit Health Research Article Anterior chamber depth (ACD) is a major risk factor of angle closure disease, and has been used in angle closure screening in various populations. However, ACD is measured from ocular biometer or anterior segment optical coherence tomography (AS-OCT), which are costly and may not be readily available in primary care and community settings. Thus, this proof-of-concept study aims to predict ACD from low-cost anterior segment photographs (ASPs) using deep-learning (DL). We included 2,311 pairs of ASPs and ACD measurements for algorithm development and validation, and 380 pairs for algorithm testing. We captured ASPs with a digital camera mounted on a slit-lamp biomicroscope. Anterior chamber depth was measured with ocular biometer (IOLMaster700 or Lenstar LS9000) in data used for algorithm development and validation, and with AS-OCT (Visante) in data used for testing. The DL algorithm was modified from the ResNet-50 architecture, and assessed using mean absolute error (MAE), coefficient-of-determination (R(2)), Bland-Altman plot and intraclass correlation coefficients (ICC). In validation, our algorithm predicted ACD with a MAE (standard deviation) of 0.18 (0.14) mm; R(2) = 0.63. The MAE of predicted ACD was 0.18 (0.14) mm in eyes with open angles and 0.19 (0.14) mm in eyes with angle closure. The ICC between actual and predicted ACD measurements was 0.81 (95% CI 0.77, 0.84). In testing, our algorithm predicted ACD with a MAE of 0.23 (0.18) mm; R(2) = 0.37. Saliency maps highlighted the pupil and its margin as the main structures used in ACD prediction. This study demonstrates the possibility of predicting ACD from ASPs via DL. This algorithm mimics an ocular biometer in making its prediction, and provides a foundation to predict other quantitative measurements that are relevant to angle closure screening. Public Library of Science 2023-02-01 /pmc/articles/PMC9931242/ /pubmed/36812642 http://dx.doi.org/10.1371/journal.pdig.0000193 Text en © 2023 Soh et al https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
spellingShingle Research Article
Soh, Zhi Da
Jiang, Yixing
S/O Ganesan, Sakthi Selvam
Zhou, Menghan
Nongiur, Monisha
Majithia, Shivani
Tham, Yih Chung
Rim, Tyler Hyungtaek
Qian, Chaoxu
Koh, Victor
Aung, Tin
Wong, Tien Yin
Xu, Xinxing
Liu, Yong
Cheng, Ching-Yu
From 2 dimensions to 3(rd) dimension: Quantitative prediction of anterior chamber depth from anterior segment photographs via deep-learning
title From 2 dimensions to 3(rd) dimension: Quantitative prediction of anterior chamber depth from anterior segment photographs via deep-learning
title_full From 2 dimensions to 3(rd) dimension: Quantitative prediction of anterior chamber depth from anterior segment photographs via deep-learning
title_fullStr From 2 dimensions to 3(rd) dimension: Quantitative prediction of anterior chamber depth from anterior segment photographs via deep-learning
title_full_unstemmed From 2 dimensions to 3(rd) dimension: Quantitative prediction of anterior chamber depth from anterior segment photographs via deep-learning
title_short From 2 dimensions to 3(rd) dimension: Quantitative prediction of anterior chamber depth from anterior segment photographs via deep-learning
title_sort from 2 dimensions to 3(rd) dimension: quantitative prediction of anterior chamber depth from anterior segment photographs via deep-learning
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9931242/
https://www.ncbi.nlm.nih.gov/pubmed/36812642
http://dx.doi.org/10.1371/journal.pdig.0000193
work_keys_str_mv AT sohzhida from2dimensionsto3rddimensionquantitativepredictionofanteriorchamberdepthfromanteriorsegmentphotographsviadeeplearning
AT jiangyixing from2dimensionsto3rddimensionquantitativepredictionofanteriorchamberdepthfromanteriorsegmentphotographsviadeeplearning
AT soganesansakthiselvam from2dimensionsto3rddimensionquantitativepredictionofanteriorchamberdepthfromanteriorsegmentphotographsviadeeplearning
AT zhoumenghan from2dimensionsto3rddimensionquantitativepredictionofanteriorchamberdepthfromanteriorsegmentphotographsviadeeplearning
AT nongiurmonisha from2dimensionsto3rddimensionquantitativepredictionofanteriorchamberdepthfromanteriorsegmentphotographsviadeeplearning
AT majithiashivani from2dimensionsto3rddimensionquantitativepredictionofanteriorchamberdepthfromanteriorsegmentphotographsviadeeplearning
AT thamyihchung from2dimensionsto3rddimensionquantitativepredictionofanteriorchamberdepthfromanteriorsegmentphotographsviadeeplearning
AT rimtylerhyungtaek from2dimensionsto3rddimensionquantitativepredictionofanteriorchamberdepthfromanteriorsegmentphotographsviadeeplearning
AT qianchaoxu from2dimensionsto3rddimensionquantitativepredictionofanteriorchamberdepthfromanteriorsegmentphotographsviadeeplearning
AT kohvictor from2dimensionsto3rddimensionquantitativepredictionofanteriorchamberdepthfromanteriorsegmentphotographsviadeeplearning
AT aungtin from2dimensionsto3rddimensionquantitativepredictionofanteriorchamberdepthfromanteriorsegmentphotographsviadeeplearning
AT wongtienyin from2dimensionsto3rddimensionquantitativepredictionofanteriorchamberdepthfromanteriorsegmentphotographsviadeeplearning
AT xuxinxing from2dimensionsto3rddimensionquantitativepredictionofanteriorchamberdepthfromanteriorsegmentphotographsviadeeplearning
AT liuyong from2dimensionsto3rddimensionquantitativepredictionofanteriorchamberdepthfromanteriorsegmentphotographsviadeeplearning
AT chengchingyu from2dimensionsto3rddimensionquantitativepredictionofanteriorchamberdepthfromanteriorsegmentphotographsviadeeplearning