Cargando…

An instrument for quality assurance in work capacity evaluation: development, evaluation, and inter-rater reliability

BACKGROUND: Employees insured in pension insurance, who are incapable of working due to ill health, are entitled to a disability pension. To assess whether an individual meets the medical requirements to be considered as disabled, a work capacity evaluation is conducted. However, there are no offici...

Descripción completa

Detalles Bibliográficos
Autores principales: Strahl, André, Gerlich, Christian, Alpers, Georg W., Gehrke, Jörg, Müller-Garnn, Annette, Vogel, Heiner
Formato: Online Artículo Texto
Lenguaje:English
Publicado: BioMed Central 2019
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6688267/
https://www.ncbi.nlm.nih.gov/pubmed/31399089
http://dx.doi.org/10.1186/s12913-019-4387-4
_version_ 1783442851794780160
author Strahl, André
Gerlich, Christian
Alpers, Georg W.
Gehrke, Jörg
Müller-Garnn, Annette
Vogel, Heiner
author_facet Strahl, André
Gerlich, Christian
Alpers, Georg W.
Gehrke, Jörg
Müller-Garnn, Annette
Vogel, Heiner
author_sort Strahl, André
collection PubMed
description BACKGROUND: Employees insured in pension insurance, who are incapable of working due to ill health, are entitled to a disability pension. To assess whether an individual meets the medical requirements to be considered as disabled, a work capacity evaluation is conducted. However, there are no official guidelines on how to perform an external quality assurance for this evaluation process. Furthermore, the quality of medical reports in the field of insurance medicine can vary substantially, and systematic evaluations are scarce. Reliability studies using peer review have repeatedly shown insufficient ability to distinguish between high, moderate and low quality. Considering literature recommendations, we developed an instrument to examine the quality of medical experts’ reports. METHODS: The peer review manual developed contains six quality domains (formal structure, clarity, transparency, completeness, medical-scientific principles, and efficiency) comprising 22 items. In addition, a superordinate criterion (survey confirmability) rank the overall quality and usefulness of a report. This criterion evaluates problems of inner logic and reasoning. Development of the manual was assisted by experienced physicians in a pre-test. We examined the observable variance in peer judgements and reliability as the most important outcome criteria. To evaluate inter-rater reliability, 20 anonymous experts’ reports detailing the work capacity evaluation were reviewed by 19 trained raters (peers). Percentage agreement and Kendall’s W, a reliability measure of concordance between two or more peers, were calculated. A total of 325 reviews were conducted. RESULTS: Agreement of peer judgements with respect to the superordinate criterion ranged from 29.2 to 87.5%. Kendall’s W for the quality domain items varied greatly, ranging from 0.09 to 0.88. With respect to the superordinate criterion, Kendall’s W was 0.39, which indicates fair agreement. The results of the percentage agreement revealed systemic peer preferences for certain deficit scale categories. CONCLUSION: The superordinate criterion was not sufficiently reliable. However, in comparison to other reliability studies, this criterion showed an equivalent reliability value. This report aims to encourage further efforts to improve evaluation instruments. To reduce disagreement between peer judgments, we propose the revision of the peer review instrument and the development and implementation of a standardized rater training to improve reliability. ELECTRONIC SUPPLEMENTARY MATERIAL: The online version of this article (10.1186/s12913-019-4387-4) contains supplementary material, which is available to authorized users.
format Online
Article
Text
id pubmed-6688267
institution National Center for Biotechnology Information
language English
publishDate 2019
publisher BioMed Central
record_format MEDLINE/PubMed
spelling pubmed-66882672019-08-14 An instrument for quality assurance in work capacity evaluation: development, evaluation, and inter-rater reliability Strahl, André Gerlich, Christian Alpers, Georg W. Gehrke, Jörg Müller-Garnn, Annette Vogel, Heiner BMC Health Serv Res Research Article BACKGROUND: Employees insured in pension insurance, who are incapable of working due to ill health, are entitled to a disability pension. To assess whether an individual meets the medical requirements to be considered as disabled, a work capacity evaluation is conducted. However, there are no official guidelines on how to perform an external quality assurance for this evaluation process. Furthermore, the quality of medical reports in the field of insurance medicine can vary substantially, and systematic evaluations are scarce. Reliability studies using peer review have repeatedly shown insufficient ability to distinguish between high, moderate and low quality. Considering literature recommendations, we developed an instrument to examine the quality of medical experts’ reports. METHODS: The peer review manual developed contains six quality domains (formal structure, clarity, transparency, completeness, medical-scientific principles, and efficiency) comprising 22 items. In addition, a superordinate criterion (survey confirmability) rank the overall quality and usefulness of a report. This criterion evaluates problems of inner logic and reasoning. Development of the manual was assisted by experienced physicians in a pre-test. We examined the observable variance in peer judgements and reliability as the most important outcome criteria. To evaluate inter-rater reliability, 20 anonymous experts’ reports detailing the work capacity evaluation were reviewed by 19 trained raters (peers). Percentage agreement and Kendall’s W, a reliability measure of concordance between two or more peers, were calculated. A total of 325 reviews were conducted. RESULTS: Agreement of peer judgements with respect to the superordinate criterion ranged from 29.2 to 87.5%. Kendall’s W for the quality domain items varied greatly, ranging from 0.09 to 0.88. With respect to the superordinate criterion, Kendall’s W was 0.39, which indicates fair agreement. The results of the percentage agreement revealed systemic peer preferences for certain deficit scale categories. CONCLUSION: The superordinate criterion was not sufficiently reliable. However, in comparison to other reliability studies, this criterion showed an equivalent reliability value. This report aims to encourage further efforts to improve evaluation instruments. To reduce disagreement between peer judgments, we propose the revision of the peer review instrument and the development and implementation of a standardized rater training to improve reliability. ELECTRONIC SUPPLEMENTARY MATERIAL: The online version of this article (10.1186/s12913-019-4387-4) contains supplementary material, which is available to authorized users. BioMed Central 2019-08-09 /pmc/articles/PMC6688267/ /pubmed/31399089 http://dx.doi.org/10.1186/s12913-019-4387-4 Text en © The Author(s). 2019 Open AccessThis article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.
spellingShingle Research Article
Strahl, André
Gerlich, Christian
Alpers, Georg W.
Gehrke, Jörg
Müller-Garnn, Annette
Vogel, Heiner
An instrument for quality assurance in work capacity evaluation: development, evaluation, and inter-rater reliability
title An instrument for quality assurance in work capacity evaluation: development, evaluation, and inter-rater reliability
title_full An instrument for quality assurance in work capacity evaluation: development, evaluation, and inter-rater reliability
title_fullStr An instrument for quality assurance in work capacity evaluation: development, evaluation, and inter-rater reliability
title_full_unstemmed An instrument for quality assurance in work capacity evaluation: development, evaluation, and inter-rater reliability
title_short An instrument for quality assurance in work capacity evaluation: development, evaluation, and inter-rater reliability
title_sort instrument for quality assurance in work capacity evaluation: development, evaluation, and inter-rater reliability
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6688267/
https://www.ncbi.nlm.nih.gov/pubmed/31399089
http://dx.doi.org/10.1186/s12913-019-4387-4
work_keys_str_mv AT strahlandre aninstrumentforqualityassuranceinworkcapacityevaluationdevelopmentevaluationandinterraterreliability
AT gerlichchristian aninstrumentforqualityassuranceinworkcapacityevaluationdevelopmentevaluationandinterraterreliability
AT alpersgeorgw aninstrumentforqualityassuranceinworkcapacityevaluationdevelopmentevaluationandinterraterreliability
AT gehrkejorg aninstrumentforqualityassuranceinworkcapacityevaluationdevelopmentevaluationandinterraterreliability
AT mullergarnnannette aninstrumentforqualityassuranceinworkcapacityevaluationdevelopmentevaluationandinterraterreliability
AT vogelheiner aninstrumentforqualityassuranceinworkcapacityevaluationdevelopmentevaluationandinterraterreliability
AT strahlandre instrumentforqualityassuranceinworkcapacityevaluationdevelopmentevaluationandinterraterreliability
AT gerlichchristian instrumentforqualityassuranceinworkcapacityevaluationdevelopmentevaluationandinterraterreliability
AT alpersgeorgw instrumentforqualityassuranceinworkcapacityevaluationdevelopmentevaluationandinterraterreliability
AT gehrkejorg instrumentforqualityassuranceinworkcapacityevaluationdevelopmentevaluationandinterraterreliability
AT mullergarnnannette instrumentforqualityassuranceinworkcapacityevaluationdevelopmentevaluationandinterraterreliability
AT vogelheiner instrumentforqualityassuranceinworkcapacityevaluationdevelopmentevaluationandinterraterreliability