Cargando…
The assessment of fundus image quality labeling reliability among graders with different backgrounds
PURPOSE: For the training of machine learning (ML) algorithms, correctly labeled ground truth data are inevitable. In this pilot study, we assessed the performance of graders with different backgrounds in the labeling of retinal fundus image quality. METHODS: Color fundus photographs were labeled us...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Public Library of Science
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9321443/ https://www.ncbi.nlm.nih.gov/pubmed/35881576 http://dx.doi.org/10.1371/journal.pone.0271156 |
_version_ | 1784756046458257408 |
---|---|
author | Laurik-Feuerstein, Kornélia Lenke Sapahia, Rishav Cabrera DeBuc, Delia Somfai, Gábor Márk |
author_facet | Laurik-Feuerstein, Kornélia Lenke Sapahia, Rishav Cabrera DeBuc, Delia Somfai, Gábor Márk |
author_sort | Laurik-Feuerstein, Kornélia Lenke |
collection | PubMed |
description | PURPOSE: For the training of machine learning (ML) algorithms, correctly labeled ground truth data are inevitable. In this pilot study, we assessed the performance of graders with different backgrounds in the labeling of retinal fundus image quality. METHODS: Color fundus photographs were labeled using a Python-based tool using four image categories: excellent (E), good (G), adequate (A) and insufficient for grading (I). We enrolled 8 subjects (4 with and 4 without medical background, groups M and NM, respectively) to whom a tutorial was presented on image quality requirements. We randomly selected 200 images from a pool of 18,145 expert-labeled images (50/E, 50/G, 50/A, 50/I). The performance of the grading was timed and the agreement was assessed. An additional grading round was performed with 14 labels for a more objective analysis. RESULTS: The median time (interquartile range) for the labeling task with 4 categories was 987.8 sec (418.6) for all graders and 872.9 sec (621.0) vs. 1019.8 sec (479.5) in the M vs. NM groups, respectively. Cohen’s weighted kappa showed moderate agreement (0.564) when using four categories that increased to substantial (0.637) when using only three by merging the E and G groups. By the use of 14 labels, the weighted kappa values were 0.594 and 0.667 when assigning four or three categories, respectively. CONCLUSION: Image grading with a Python-based tool seems to be a simple yet possibly efficient solution for the labeling of fundus images according to image quality that does not necessarily require medical background. Such grading can be subject to variability but could still effectively serve the robust identification of images with insufficient quality. This emphasizes the opportunity for the democratization of ML-applications among persons with both medical and non-medical background. However, simplicity of the grading system is key to successful categorization. |
format | Online Article Text |
id | pubmed-9321443 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Public Library of Science |
record_format | MEDLINE/PubMed |
spelling | pubmed-93214432022-07-27 The assessment of fundus image quality labeling reliability among graders with different backgrounds Laurik-Feuerstein, Kornélia Lenke Sapahia, Rishav Cabrera DeBuc, Delia Somfai, Gábor Márk PLoS One Research Article PURPOSE: For the training of machine learning (ML) algorithms, correctly labeled ground truth data are inevitable. In this pilot study, we assessed the performance of graders with different backgrounds in the labeling of retinal fundus image quality. METHODS: Color fundus photographs were labeled using a Python-based tool using four image categories: excellent (E), good (G), adequate (A) and insufficient for grading (I). We enrolled 8 subjects (4 with and 4 without medical background, groups M and NM, respectively) to whom a tutorial was presented on image quality requirements. We randomly selected 200 images from a pool of 18,145 expert-labeled images (50/E, 50/G, 50/A, 50/I). The performance of the grading was timed and the agreement was assessed. An additional grading round was performed with 14 labels for a more objective analysis. RESULTS: The median time (interquartile range) for the labeling task with 4 categories was 987.8 sec (418.6) for all graders and 872.9 sec (621.0) vs. 1019.8 sec (479.5) in the M vs. NM groups, respectively. Cohen’s weighted kappa showed moderate agreement (0.564) when using four categories that increased to substantial (0.637) when using only three by merging the E and G groups. By the use of 14 labels, the weighted kappa values were 0.594 and 0.667 when assigning four or three categories, respectively. CONCLUSION: Image grading with a Python-based tool seems to be a simple yet possibly efficient solution for the labeling of fundus images according to image quality that does not necessarily require medical background. Such grading can be subject to variability but could still effectively serve the robust identification of images with insufficient quality. This emphasizes the opportunity for the democratization of ML-applications among persons with both medical and non-medical background. However, simplicity of the grading system is key to successful categorization. Public Library of Science 2022-07-26 /pmc/articles/PMC9321443/ /pubmed/35881576 http://dx.doi.org/10.1371/journal.pone.0271156 Text en © 2022 Laurik-Feuerstein et al https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. |
spellingShingle | Research Article Laurik-Feuerstein, Kornélia Lenke Sapahia, Rishav Cabrera DeBuc, Delia Somfai, Gábor Márk The assessment of fundus image quality labeling reliability among graders with different backgrounds |
title | The assessment of fundus image quality labeling reliability among graders with different backgrounds |
title_full | The assessment of fundus image quality labeling reliability among graders with different backgrounds |
title_fullStr | The assessment of fundus image quality labeling reliability among graders with different backgrounds |
title_full_unstemmed | The assessment of fundus image quality labeling reliability among graders with different backgrounds |
title_short | The assessment of fundus image quality labeling reliability among graders with different backgrounds |
title_sort | assessment of fundus image quality labeling reliability among graders with different backgrounds |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9321443/ https://www.ncbi.nlm.nih.gov/pubmed/35881576 http://dx.doi.org/10.1371/journal.pone.0271156 |
work_keys_str_mv | AT laurikfeuersteinkornelialenke theassessmentoffundusimagequalitylabelingreliabilityamonggraderswithdifferentbackgrounds AT sapahiarishav theassessmentoffundusimagequalitylabelingreliabilityamonggraderswithdifferentbackgrounds AT cabreradebucdelia theassessmentoffundusimagequalitylabelingreliabilityamonggraderswithdifferentbackgrounds AT somfaigabormark theassessmentoffundusimagequalitylabelingreliabilityamonggraderswithdifferentbackgrounds AT laurikfeuersteinkornelialenke assessmentoffundusimagequalitylabelingreliabilityamonggraderswithdifferentbackgrounds AT sapahiarishav assessmentoffundusimagequalitylabelingreliabilityamonggraderswithdifferentbackgrounds AT cabreradebucdelia assessmentoffundusimagequalitylabelingreliabilityamonggraderswithdifferentbackgrounds AT somfaigabormark assessmentoffundusimagequalitylabelingreliabilityamonggraderswithdifferentbackgrounds |