Cargando…

Inter-rater reliability of AMSTAR is dependent on the pair of reviewers

BACKGROUND: Inter-rater reliability (IRR) is mainly assessed based on only two reviewers of unknown expertise. The aim of this paper is to examine differences in the IRR of the Assessment of Multiple Systematic Reviews (AMSTAR) and R(evised)-AMSTAR depending on the pair of reviewers. METHODS: Five r...

Descripción completa

Detalles Bibliográficos
Autores principales: Pieper, Dawid, Jacobs, Anja, Weikert, Beate, Fishta, Alba, Wegewitz, Uta
Formato: Online Artículo Texto
Lenguaje:English
Publicado: BioMed Central 2017
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5504630/
https://www.ncbi.nlm.nih.gov/pubmed/28693497
http://dx.doi.org/10.1186/s12874-017-0380-y
_version_ 1783249314066202624
author Pieper, Dawid
Jacobs, Anja
Weikert, Beate
Fishta, Alba
Wegewitz, Uta
author_facet Pieper, Dawid
Jacobs, Anja
Weikert, Beate
Fishta, Alba
Wegewitz, Uta
author_sort Pieper, Dawid
collection PubMed
description BACKGROUND: Inter-rater reliability (IRR) is mainly assessed based on only two reviewers of unknown expertise. The aim of this paper is to examine differences in the IRR of the Assessment of Multiple Systematic Reviews (AMSTAR) and R(evised)-AMSTAR depending on the pair of reviewers. METHODS: Five reviewers independently applied AMSTAR and R-AMSTAR to 16 systematic reviews (eight Cochrane reviews and eight non-Cochrane reviews) from the field of occupational health. Responses were dichotomized and reliability measures were calculated by applying Holsti’s method (r) and Cohen’s kappa (κ) to all potential pairs of reviewers. Given that five reviewers participated in the study, there were ten possible pairs of reviewers. RESULTS: Inter-rater reliability varied for AMSTAR between r = 0.82 and r = 0.98 (median r = 0.88) using Holsti’s method and κ = 0.41 and κ = 0.69 (median κ = 0.52) using Cohen’s kappa and for R-AMSTAR between r = 0.77 and r = 0.89 (median r = 0.82) and κ = 0.32 and κ = 0.67 (median κ = 0.45) depending on the pair of reviewers. The same pair of reviewers yielded the highest IRR for both instruments. Pairwise Cohen’s kappa reliability measures showed a moderate correlation between AMSTAR and R-AMSTAR (Spearman’s ρ =0.50). The mean inter-rater reliability for AMSTAR was highest for item 1 (κ = 1.00) and item 5 (κ = 0.78), while lowest values were found for items 3, 8, 9 and 11, which showed only fair agreement. CONCLUSIONS: Inter-rater reliability varies widely depending on the pair of reviewers. There may be some shortcomings associated with conducting reliability studies with only two reviewers. Further studies should include additional reviewers and should probably also take account of their level of expertise. ELECTRONIC SUPPLEMENTARY MATERIAL: The online version of this article (doi:10.1186/s12874-017-0380-y) contains supplementary material, which is available to authorized users.
format Online
Article
Text
id pubmed-5504630
institution National Center for Biotechnology Information
language English
publishDate 2017
publisher BioMed Central
record_format MEDLINE/PubMed
spelling pubmed-55046302017-07-12 Inter-rater reliability of AMSTAR is dependent on the pair of reviewers Pieper, Dawid Jacobs, Anja Weikert, Beate Fishta, Alba Wegewitz, Uta BMC Med Res Methodol Research Article BACKGROUND: Inter-rater reliability (IRR) is mainly assessed based on only two reviewers of unknown expertise. The aim of this paper is to examine differences in the IRR of the Assessment of Multiple Systematic Reviews (AMSTAR) and R(evised)-AMSTAR depending on the pair of reviewers. METHODS: Five reviewers independently applied AMSTAR and R-AMSTAR to 16 systematic reviews (eight Cochrane reviews and eight non-Cochrane reviews) from the field of occupational health. Responses were dichotomized and reliability measures were calculated by applying Holsti’s method (r) and Cohen’s kappa (κ) to all potential pairs of reviewers. Given that five reviewers participated in the study, there were ten possible pairs of reviewers. RESULTS: Inter-rater reliability varied for AMSTAR between r = 0.82 and r = 0.98 (median r = 0.88) using Holsti’s method and κ = 0.41 and κ = 0.69 (median κ = 0.52) using Cohen’s kappa and for R-AMSTAR between r = 0.77 and r = 0.89 (median r = 0.82) and κ = 0.32 and κ = 0.67 (median κ = 0.45) depending on the pair of reviewers. The same pair of reviewers yielded the highest IRR for both instruments. Pairwise Cohen’s kappa reliability measures showed a moderate correlation between AMSTAR and R-AMSTAR (Spearman’s ρ =0.50). The mean inter-rater reliability for AMSTAR was highest for item 1 (κ = 1.00) and item 5 (κ = 0.78), while lowest values were found for items 3, 8, 9 and 11, which showed only fair agreement. CONCLUSIONS: Inter-rater reliability varies widely depending on the pair of reviewers. There may be some shortcomings associated with conducting reliability studies with only two reviewers. Further studies should include additional reviewers and should probably also take account of their level of expertise. ELECTRONIC SUPPLEMENTARY MATERIAL: The online version of this article (doi:10.1186/s12874-017-0380-y) contains supplementary material, which is available to authorized users. BioMed Central 2017-07-11 /pmc/articles/PMC5504630/ /pubmed/28693497 http://dx.doi.org/10.1186/s12874-017-0380-y Text en © The Author(s). 2017 Open AccessThis article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.
spellingShingle Research Article
Pieper, Dawid
Jacobs, Anja
Weikert, Beate
Fishta, Alba
Wegewitz, Uta
Inter-rater reliability of AMSTAR is dependent on the pair of reviewers
title Inter-rater reliability of AMSTAR is dependent on the pair of reviewers
title_full Inter-rater reliability of AMSTAR is dependent on the pair of reviewers
title_fullStr Inter-rater reliability of AMSTAR is dependent on the pair of reviewers
title_full_unstemmed Inter-rater reliability of AMSTAR is dependent on the pair of reviewers
title_short Inter-rater reliability of AMSTAR is dependent on the pair of reviewers
title_sort inter-rater reliability of amstar is dependent on the pair of reviewers
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5504630/
https://www.ncbi.nlm.nih.gov/pubmed/28693497
http://dx.doi.org/10.1186/s12874-017-0380-y
work_keys_str_mv AT pieperdawid interraterreliabilityofamstarisdependentonthepairofreviewers
AT jacobsanja interraterreliabilityofamstarisdependentonthepairofreviewers
AT weikertbeate interraterreliabilityofamstarisdependentonthepairofreviewers
AT fishtaalba interraterreliabilityofamstarisdependentonthepairofreviewers
AT wegewitzuta interraterreliabilityofamstarisdependentonthepairofreviewers