Cargando…
Inter-rater reliability of AMSTAR is dependent on the pair of reviewers
BACKGROUND: Inter-rater reliability (IRR) is mainly assessed based on only two reviewers of unknown expertise. The aim of this paper is to examine differences in the IRR of the Assessment of Multiple Systematic Reviews (AMSTAR) and R(evised)-AMSTAR depending on the pair of reviewers. METHODS: Five r...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
BioMed Central
2017
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5504630/ https://www.ncbi.nlm.nih.gov/pubmed/28693497 http://dx.doi.org/10.1186/s12874-017-0380-y |
_version_ | 1783249314066202624 |
---|---|
author | Pieper, Dawid Jacobs, Anja Weikert, Beate Fishta, Alba Wegewitz, Uta |
author_facet | Pieper, Dawid Jacobs, Anja Weikert, Beate Fishta, Alba Wegewitz, Uta |
author_sort | Pieper, Dawid |
collection | PubMed |
description | BACKGROUND: Inter-rater reliability (IRR) is mainly assessed based on only two reviewers of unknown expertise. The aim of this paper is to examine differences in the IRR of the Assessment of Multiple Systematic Reviews (AMSTAR) and R(evised)-AMSTAR depending on the pair of reviewers. METHODS: Five reviewers independently applied AMSTAR and R-AMSTAR to 16 systematic reviews (eight Cochrane reviews and eight non-Cochrane reviews) from the field of occupational health. Responses were dichotomized and reliability measures were calculated by applying Holsti’s method (r) and Cohen’s kappa (κ) to all potential pairs of reviewers. Given that five reviewers participated in the study, there were ten possible pairs of reviewers. RESULTS: Inter-rater reliability varied for AMSTAR between r = 0.82 and r = 0.98 (median r = 0.88) using Holsti’s method and κ = 0.41 and κ = 0.69 (median κ = 0.52) using Cohen’s kappa and for R-AMSTAR between r = 0.77 and r = 0.89 (median r = 0.82) and κ = 0.32 and κ = 0.67 (median κ = 0.45) depending on the pair of reviewers. The same pair of reviewers yielded the highest IRR for both instruments. Pairwise Cohen’s kappa reliability measures showed a moderate correlation between AMSTAR and R-AMSTAR (Spearman’s ρ =0.50). The mean inter-rater reliability for AMSTAR was highest for item 1 (κ = 1.00) and item 5 (κ = 0.78), while lowest values were found for items 3, 8, 9 and 11, which showed only fair agreement. CONCLUSIONS: Inter-rater reliability varies widely depending on the pair of reviewers. There may be some shortcomings associated with conducting reliability studies with only two reviewers. Further studies should include additional reviewers and should probably also take account of their level of expertise. ELECTRONIC SUPPLEMENTARY MATERIAL: The online version of this article (doi:10.1186/s12874-017-0380-y) contains supplementary material, which is available to authorized users. |
format | Online Article Text |
id | pubmed-5504630 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2017 |
publisher | BioMed Central |
record_format | MEDLINE/PubMed |
spelling | pubmed-55046302017-07-12 Inter-rater reliability of AMSTAR is dependent on the pair of reviewers Pieper, Dawid Jacobs, Anja Weikert, Beate Fishta, Alba Wegewitz, Uta BMC Med Res Methodol Research Article BACKGROUND: Inter-rater reliability (IRR) is mainly assessed based on only two reviewers of unknown expertise. The aim of this paper is to examine differences in the IRR of the Assessment of Multiple Systematic Reviews (AMSTAR) and R(evised)-AMSTAR depending on the pair of reviewers. METHODS: Five reviewers independently applied AMSTAR and R-AMSTAR to 16 systematic reviews (eight Cochrane reviews and eight non-Cochrane reviews) from the field of occupational health. Responses were dichotomized and reliability measures were calculated by applying Holsti’s method (r) and Cohen’s kappa (κ) to all potential pairs of reviewers. Given that five reviewers participated in the study, there were ten possible pairs of reviewers. RESULTS: Inter-rater reliability varied for AMSTAR between r = 0.82 and r = 0.98 (median r = 0.88) using Holsti’s method and κ = 0.41 and κ = 0.69 (median κ = 0.52) using Cohen’s kappa and for R-AMSTAR between r = 0.77 and r = 0.89 (median r = 0.82) and κ = 0.32 and κ = 0.67 (median κ = 0.45) depending on the pair of reviewers. The same pair of reviewers yielded the highest IRR for both instruments. Pairwise Cohen’s kappa reliability measures showed a moderate correlation between AMSTAR and R-AMSTAR (Spearman’s ρ =0.50). The mean inter-rater reliability for AMSTAR was highest for item 1 (κ = 1.00) and item 5 (κ = 0.78), while lowest values were found for items 3, 8, 9 and 11, which showed only fair agreement. CONCLUSIONS: Inter-rater reliability varies widely depending on the pair of reviewers. There may be some shortcomings associated with conducting reliability studies with only two reviewers. Further studies should include additional reviewers and should probably also take account of their level of expertise. ELECTRONIC SUPPLEMENTARY MATERIAL: The online version of this article (doi:10.1186/s12874-017-0380-y) contains supplementary material, which is available to authorized users. BioMed Central 2017-07-11 /pmc/articles/PMC5504630/ /pubmed/28693497 http://dx.doi.org/10.1186/s12874-017-0380-y Text en © The Author(s). 2017 Open AccessThis article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated. |
spellingShingle | Research Article Pieper, Dawid Jacobs, Anja Weikert, Beate Fishta, Alba Wegewitz, Uta Inter-rater reliability of AMSTAR is dependent on the pair of reviewers |
title | Inter-rater reliability of AMSTAR is dependent on the pair of reviewers |
title_full | Inter-rater reliability of AMSTAR is dependent on the pair of reviewers |
title_fullStr | Inter-rater reliability of AMSTAR is dependent on the pair of reviewers |
title_full_unstemmed | Inter-rater reliability of AMSTAR is dependent on the pair of reviewers |
title_short | Inter-rater reliability of AMSTAR is dependent on the pair of reviewers |
title_sort | inter-rater reliability of amstar is dependent on the pair of reviewers |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5504630/ https://www.ncbi.nlm.nih.gov/pubmed/28693497 http://dx.doi.org/10.1186/s12874-017-0380-y |
work_keys_str_mv | AT pieperdawid interraterreliabilityofamstarisdependentonthepairofreviewers AT jacobsanja interraterreliabilityofamstarisdependentonthepairofreviewers AT weikertbeate interraterreliabilityofamstarisdependentonthepairofreviewers AT fishtaalba interraterreliabilityofamstarisdependentonthepairofreviewers AT wegewitzuta interraterreliabilityofamstarisdependentonthepairofreviewers |