Cargando…

A Reliability-Generalization Study of Journal Peer Reviews: A Multilevel Meta-Analysis of Inter-Rater Reliability and Its Determinants

BACKGROUND: This paper presents the first meta-analysis for the inter-rater reliability (IRR) of journal peer reviews. IRR is defined as the extent to which two or more independent reviews of the same scientific document agree. METHODOLOGY/PRINCIPAL FINDINGS: Altogether, 70 reliability coefficients...

Descripción completa

Detalles Bibliográficos
Autores principales: Bornmann, Lutz, Mutz, Rüdiger, Daniel, Hans-Dieter
Formato: Texto
Lenguaje:English
Publicado: Public Library of Science 2010
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3001856/
https://www.ncbi.nlm.nih.gov/pubmed/21179459
http://dx.doi.org/10.1371/journal.pone.0014331
_version_ 1782193662314151936
author Bornmann, Lutz
Mutz, Rüdiger
Daniel, Hans-Dieter
author_facet Bornmann, Lutz
Mutz, Rüdiger
Daniel, Hans-Dieter
author_sort Bornmann, Lutz
collection PubMed
description BACKGROUND: This paper presents the first meta-analysis for the inter-rater reliability (IRR) of journal peer reviews. IRR is defined as the extent to which two or more independent reviews of the same scientific document agree. METHODOLOGY/PRINCIPAL FINDINGS: Altogether, 70 reliability coefficients (Cohen's Kappa, intra-class correlation [ICC], and Pearson product-moment correlation [r]) from 48 studies were taken into account in the meta-analysis. The studies were based on a total of 19,443 manuscripts; on average, each study had a sample size of 311 manuscripts (minimum: 28, maximum: 1983). The results of the meta-analysis confirmed the findings of the narrative literature reviews published to date: The level of IRR (mean ICC/r(2) = .34, mean Cohen's Kappa = .17) was low. To explain the study-to-study variation of the IRR coefficients, meta-regression analyses were calculated using seven covariates. Two covariates that emerged in the meta-regression analyses as statistically significant to gain an approximate homogeneity of the intra-class correlations indicated that, firstly, the more manuscripts that a study is based on, the smaller the reported IRR coefficients are. Secondly, if the information of the rating system for reviewers was reported in a study, then this was associated with a smaller IRR coefficient than if the information was not conveyed. CONCLUSIONS/SIGNIFICANCE: Studies that report a high level of IRR are to be considered less credible than those with a low level of IRR. According to our meta-analysis the IRR of peer assessments is quite limited and needs improvement (e.g., reader system).
format Text
id pubmed-3001856
institution National Center for Biotechnology Information
language English
publishDate 2010
publisher Public Library of Science
record_format MEDLINE/PubMed
spelling pubmed-30018562010-12-21 A Reliability-Generalization Study of Journal Peer Reviews: A Multilevel Meta-Analysis of Inter-Rater Reliability and Its Determinants Bornmann, Lutz Mutz, Rüdiger Daniel, Hans-Dieter PLoS One Research Article BACKGROUND: This paper presents the first meta-analysis for the inter-rater reliability (IRR) of journal peer reviews. IRR is defined as the extent to which two or more independent reviews of the same scientific document agree. METHODOLOGY/PRINCIPAL FINDINGS: Altogether, 70 reliability coefficients (Cohen's Kappa, intra-class correlation [ICC], and Pearson product-moment correlation [r]) from 48 studies were taken into account in the meta-analysis. The studies were based on a total of 19,443 manuscripts; on average, each study had a sample size of 311 manuscripts (minimum: 28, maximum: 1983). The results of the meta-analysis confirmed the findings of the narrative literature reviews published to date: The level of IRR (mean ICC/r(2) = .34, mean Cohen's Kappa = .17) was low. To explain the study-to-study variation of the IRR coefficients, meta-regression analyses were calculated using seven covariates. Two covariates that emerged in the meta-regression analyses as statistically significant to gain an approximate homogeneity of the intra-class correlations indicated that, firstly, the more manuscripts that a study is based on, the smaller the reported IRR coefficients are. Secondly, if the information of the rating system for reviewers was reported in a study, then this was associated with a smaller IRR coefficient than if the information was not conveyed. CONCLUSIONS/SIGNIFICANCE: Studies that report a high level of IRR are to be considered less credible than those with a low level of IRR. According to our meta-analysis the IRR of peer assessments is quite limited and needs improvement (e.g., reader system). Public Library of Science 2010-12-14 /pmc/articles/PMC3001856/ /pubmed/21179459 http://dx.doi.org/10.1371/journal.pone.0014331 Text en Bornmann et al. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are properly credited.
spellingShingle Research Article
Bornmann, Lutz
Mutz, Rüdiger
Daniel, Hans-Dieter
A Reliability-Generalization Study of Journal Peer Reviews: A Multilevel Meta-Analysis of Inter-Rater Reliability and Its Determinants
title A Reliability-Generalization Study of Journal Peer Reviews: A Multilevel Meta-Analysis of Inter-Rater Reliability and Its Determinants
title_full A Reliability-Generalization Study of Journal Peer Reviews: A Multilevel Meta-Analysis of Inter-Rater Reliability and Its Determinants
title_fullStr A Reliability-Generalization Study of Journal Peer Reviews: A Multilevel Meta-Analysis of Inter-Rater Reliability and Its Determinants
title_full_unstemmed A Reliability-Generalization Study of Journal Peer Reviews: A Multilevel Meta-Analysis of Inter-Rater Reliability and Its Determinants
title_short A Reliability-Generalization Study of Journal Peer Reviews: A Multilevel Meta-Analysis of Inter-Rater Reliability and Its Determinants
title_sort reliability-generalization study of journal peer reviews: a multilevel meta-analysis of inter-rater reliability and its determinants
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3001856/
https://www.ncbi.nlm.nih.gov/pubmed/21179459
http://dx.doi.org/10.1371/journal.pone.0014331
work_keys_str_mv AT bornmannlutz areliabilitygeneralizationstudyofjournalpeerreviewsamultilevelmetaanalysisofinterraterreliabilityanditsdeterminants
AT mutzrudiger areliabilitygeneralizationstudyofjournalpeerreviewsamultilevelmetaanalysisofinterraterreliabilityanditsdeterminants
AT danielhansdieter areliabilitygeneralizationstudyofjournalpeerreviewsamultilevelmetaanalysisofinterraterreliabilityanditsdeterminants
AT bornmannlutz reliabilitygeneralizationstudyofjournalpeerreviewsamultilevelmetaanalysisofinterraterreliabilityanditsdeterminants
AT mutzrudiger reliabilitygeneralizationstudyofjournalpeerreviewsamultilevelmetaanalysisofinterraterreliabilityanditsdeterminants
AT danielhansdieter reliabilitygeneralizationstudyofjournalpeerreviewsamultilevelmetaanalysisofinterraterreliabilityanditsdeterminants