Cargando…

Interrater reliability: the kappa statistic

The kappa statistic is frequently used to test interrater reliability. The importance of rater reliability lies in the fact that it represents the extent to which the data collected in the study are correct representations of the variables measured. Measurement of the extent to which data collectors...

Descripción completa

Detalles Bibliográficos
Autor principal: McHugh, Mary L.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Croatian Society of Medical Biochemistry and Laboratory Medicine 2012
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3900052/
https://www.ncbi.nlm.nih.gov/pubmed/23092060
_version_ 1782300650597515264
author McHugh, Mary L.
author_facet McHugh, Mary L.
author_sort McHugh, Mary L.
collection PubMed
description The kappa statistic is frequently used to test interrater reliability. The importance of rater reliability lies in the fact that it represents the extent to which the data collected in the study are correct representations of the variables measured. Measurement of the extent to which data collectors (raters) assign the same score to the same variable is called interrater reliability. While there have been a variety of methods to measure interrater reliability, traditionally it was measured as percent agreement, calculated as the number of agreement scores divided by the total number of scores. In 1960, Jacob Cohen critiqued use of percent agreement due to its inability to account for chance agreement. He introduced the Cohen’s kappa, developed to account for the possibility that raters actually guess on at least some variables due to uncertainty. Like most correlation statistics, the kappa can range from −1 to +1. While the kappa is one of the most commonly used statistics to test interrater reliability, it has limitations. Judgments about what level of kappa should be acceptable for health research are questioned. Cohen’s suggested interpretation may be too lenient for health related studies because it implies that a score as low as 0.41 might be acceptable. Kappa and percent agreement are compared, and levels for both kappa and percent agreement that should be demanded in healthcare studies are suggested.
format Online
Article
Text
id pubmed-3900052
institution National Center for Biotechnology Information
language English
publishDate 2012
publisher Croatian Society of Medical Biochemistry and Laboratory Medicine
record_format MEDLINE/PubMed
spelling pubmed-39000522014-01-23 Interrater reliability: the kappa statistic McHugh, Mary L. Biochem Med (Zagreb) Lessons in Biostatistics The kappa statistic is frequently used to test interrater reliability. The importance of rater reliability lies in the fact that it represents the extent to which the data collected in the study are correct representations of the variables measured. Measurement of the extent to which data collectors (raters) assign the same score to the same variable is called interrater reliability. While there have been a variety of methods to measure interrater reliability, traditionally it was measured as percent agreement, calculated as the number of agreement scores divided by the total number of scores. In 1960, Jacob Cohen critiqued use of percent agreement due to its inability to account for chance agreement. He introduced the Cohen’s kappa, developed to account for the possibility that raters actually guess on at least some variables due to uncertainty. Like most correlation statistics, the kappa can range from −1 to +1. While the kappa is one of the most commonly used statistics to test interrater reliability, it has limitations. Judgments about what level of kappa should be acceptable for health research are questioned. Cohen’s suggested interpretation may be too lenient for health related studies because it implies that a score as low as 0.41 might be acceptable. Kappa and percent agreement are compared, and levels for both kappa and percent agreement that should be demanded in healthcare studies are suggested. Croatian Society of Medical Biochemistry and Laboratory Medicine 2012-10-15 /pmc/articles/PMC3900052/ /pubmed/23092060 Text en ©Copyright by Croatian Society of Medical Biochemistry and Laboratory Medicine This is an Open Access article distributed under the terms of the Creative Commons Attribution Non-Commercial License (http://creativecommons.org/licenses/by-nc/3.0/) which permits unrestricted non-commercial use, distribution, and reproduction in any medium, provided the original work is properly cited.
spellingShingle Lessons in Biostatistics
McHugh, Mary L.
Interrater reliability: the kappa statistic
title Interrater reliability: the kappa statistic
title_full Interrater reliability: the kappa statistic
title_fullStr Interrater reliability: the kappa statistic
title_full_unstemmed Interrater reliability: the kappa statistic
title_short Interrater reliability: the kappa statistic
title_sort interrater reliability: the kappa statistic
topic Lessons in Biostatistics
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3900052/
https://www.ncbi.nlm.nih.gov/pubmed/23092060
work_keys_str_mv AT mchughmaryl interraterreliabilitythekappastatistic