Cargando…

Scoring reading parameters: An inter-rater reliability study using the MNREAD chart

PURPOSE: First, to evaluate inter-rater reliability when human raters estimate the reading performance of visually impaired individuals using the MNREAD acuity chart. Second, to evaluate the agreement between computer-based scoring algorithms and compare them with human rating. METHODS: Reading perf...

Descripción completa

Detalles Bibliográficos
Autores principales: Baskaran, Karthikeyan, Macedo, Antonio Filipe, He, Yingchen, Hernandez-Moreno, Laura, Queirós, Tatiana, Mansfield, J. Stephen, Calabrèse, Aurélie
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Public Library of Science 2019
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6555504/
https://www.ncbi.nlm.nih.gov/pubmed/31173587
http://dx.doi.org/10.1371/journal.pone.0216775
_version_ 1783425165271498752
author Baskaran, Karthikeyan
Macedo, Antonio Filipe
He, Yingchen
Hernandez-Moreno, Laura
Queirós, Tatiana
Mansfield, J. Stephen
Calabrèse, Aurélie
author_facet Baskaran, Karthikeyan
Macedo, Antonio Filipe
He, Yingchen
Hernandez-Moreno, Laura
Queirós, Tatiana
Mansfield, J. Stephen
Calabrèse, Aurélie
author_sort Baskaran, Karthikeyan
collection PubMed
description PURPOSE: First, to evaluate inter-rater reliability when human raters estimate the reading performance of visually impaired individuals using the MNREAD acuity chart. Second, to evaluate the agreement between computer-based scoring algorithms and compare them with human rating. METHODS: Reading performance was measured for 101 individuals with low vision, using the Portuguese version of the MNREAD test. Seven raters estimated the maximum reading speed (MRS) and critical print size (CPS) of each individual MNREAD curve. MRS and CPS were also calculated automatically for each curve using two different algorithms: the original standard deviation method (SDev) and a non-linear mixed effects (NLME) modeling. Intra-class correlation coefficients (ICC) were used to estimate absolute agreement between raters and/or algorithms. RESULTS: Absolute agreement between raters was ‘excellent’ for MRS (ICC = 0.97; 95%CI [0.96, 0.98]) and ‘moderate’ to ‘good’ for CPS (ICC = 0.77; 95%CI [0.69, 0.83]). For CPS, inter-rater reliability was poorer among less experienced raters (ICC = 0.70; 95%CI [0.57, 0.80]) when compared to experienced ones (ICC = 0.82; 95%CI [0.76, 0.88]). Absolute agreement between the two algorithms was ‘excellent’ for MRS (ICC = 0.96; 95%CI [0.91, 0.98]). For CPS, the best possible agreement was found for CPS defined as the print size sustaining 80% of MRS (ICC = 0.77; 95%CI [0.68, 0.84]). Absolute agreement between raters and automated methods was ‘excellent’ for MRS (ICC = 0.96; 95% CI [0.88, 0.98] for SDev; ICC = 0.97; 95% CI [0.95, 0.98] for NLME). For CPS, absolute agreement between raters and SDev ranged from ‘poor’ to ‘good’ (ICC = 0.66; 95% CI [0.3, 0.80]), while agreement between raters and NLME was ‘good’ (ICC = 0.83; 95% CI [0.76, 0.88]). CONCLUSION: For MRS, inter-rater reliability is excellent, even considering the possibility of noisy and/or incomplete data collected in low-vision individuals. For CPS, inter-rater reliability is lower. This may be problematic, for instance in the context of multisite investigations or follow-up examinations. The NLME method showed better agreement with the raters than the SDev method for both reading parameters. Setting up consensual guidelines to deal with ambiguous curves may help improve reliability. While the exact definition of CPS should be chosen on a case-by-case basis depending on the clinician or researcher’s motivations, evidence suggests that estimating CPS as the smallest print size sustaining about 80% of MRS would increase inter-rater reliability.
format Online
Article
Text
id pubmed-6555504
institution National Center for Biotechnology Information
language English
publishDate 2019
publisher Public Library of Science
record_format MEDLINE/PubMed
spelling pubmed-65555042019-06-17 Scoring reading parameters: An inter-rater reliability study using the MNREAD chart Baskaran, Karthikeyan Macedo, Antonio Filipe He, Yingchen Hernandez-Moreno, Laura Queirós, Tatiana Mansfield, J. Stephen Calabrèse, Aurélie PLoS One Research Article PURPOSE: First, to evaluate inter-rater reliability when human raters estimate the reading performance of visually impaired individuals using the MNREAD acuity chart. Second, to evaluate the agreement between computer-based scoring algorithms and compare them with human rating. METHODS: Reading performance was measured for 101 individuals with low vision, using the Portuguese version of the MNREAD test. Seven raters estimated the maximum reading speed (MRS) and critical print size (CPS) of each individual MNREAD curve. MRS and CPS were also calculated automatically for each curve using two different algorithms: the original standard deviation method (SDev) and a non-linear mixed effects (NLME) modeling. Intra-class correlation coefficients (ICC) were used to estimate absolute agreement between raters and/or algorithms. RESULTS: Absolute agreement between raters was ‘excellent’ for MRS (ICC = 0.97; 95%CI [0.96, 0.98]) and ‘moderate’ to ‘good’ for CPS (ICC = 0.77; 95%CI [0.69, 0.83]). For CPS, inter-rater reliability was poorer among less experienced raters (ICC = 0.70; 95%CI [0.57, 0.80]) when compared to experienced ones (ICC = 0.82; 95%CI [0.76, 0.88]). Absolute agreement between the two algorithms was ‘excellent’ for MRS (ICC = 0.96; 95%CI [0.91, 0.98]). For CPS, the best possible agreement was found for CPS defined as the print size sustaining 80% of MRS (ICC = 0.77; 95%CI [0.68, 0.84]). Absolute agreement between raters and automated methods was ‘excellent’ for MRS (ICC = 0.96; 95% CI [0.88, 0.98] for SDev; ICC = 0.97; 95% CI [0.95, 0.98] for NLME). For CPS, absolute agreement between raters and SDev ranged from ‘poor’ to ‘good’ (ICC = 0.66; 95% CI [0.3, 0.80]), while agreement between raters and NLME was ‘good’ (ICC = 0.83; 95% CI [0.76, 0.88]). CONCLUSION: For MRS, inter-rater reliability is excellent, even considering the possibility of noisy and/or incomplete data collected in low-vision individuals. For CPS, inter-rater reliability is lower. This may be problematic, for instance in the context of multisite investigations or follow-up examinations. The NLME method showed better agreement with the raters than the SDev method for both reading parameters. Setting up consensual guidelines to deal with ambiguous curves may help improve reliability. While the exact definition of CPS should be chosen on a case-by-case basis depending on the clinician or researcher’s motivations, evidence suggests that estimating CPS as the smallest print size sustaining about 80% of MRS would increase inter-rater reliability. Public Library of Science 2019-06-07 /pmc/articles/PMC6555504/ /pubmed/31173587 http://dx.doi.org/10.1371/journal.pone.0216775 Text en © 2019 Baskaran et al http://creativecommons.org/licenses/by/4.0/ This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
spellingShingle Research Article
Baskaran, Karthikeyan
Macedo, Antonio Filipe
He, Yingchen
Hernandez-Moreno, Laura
Queirós, Tatiana
Mansfield, J. Stephen
Calabrèse, Aurélie
Scoring reading parameters: An inter-rater reliability study using the MNREAD chart
title Scoring reading parameters: An inter-rater reliability study using the MNREAD chart
title_full Scoring reading parameters: An inter-rater reliability study using the MNREAD chart
title_fullStr Scoring reading parameters: An inter-rater reliability study using the MNREAD chart
title_full_unstemmed Scoring reading parameters: An inter-rater reliability study using the MNREAD chart
title_short Scoring reading parameters: An inter-rater reliability study using the MNREAD chart
title_sort scoring reading parameters: an inter-rater reliability study using the mnread chart
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6555504/
https://www.ncbi.nlm.nih.gov/pubmed/31173587
http://dx.doi.org/10.1371/journal.pone.0216775
work_keys_str_mv AT baskarankarthikeyan scoringreadingparametersaninterraterreliabilitystudyusingthemnreadchart
AT macedoantoniofilipe scoringreadingparametersaninterraterreliabilitystudyusingthemnreadchart
AT heyingchen scoringreadingparametersaninterraterreliabilitystudyusingthemnreadchart
AT hernandezmorenolaura scoringreadingparametersaninterraterreliabilitystudyusingthemnreadchart
AT queirostatiana scoringreadingparametersaninterraterreliabilitystudyusingthemnreadchart
AT mansfieldjstephen scoringreadingparametersaninterraterreliabilitystudyusingthemnreadchart
AT calabreseaurelie scoringreadingparametersaninterraterreliabilitystudyusingthemnreadchart