Cargando…
Evaluations of statistical methods for outlier detection when benchmarking in clinical registries: a systematic review
OBJECTIVES: Benchmarking is common in clinical registries to support the improvement of health outcomes by identifying underperforming clinician or health service providers. Despite the rise in clinical registries and interest in publicly reporting benchmarking results, appropriate methods for bench...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
BMJ Publishing Group
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10351235/ https://www.ncbi.nlm.nih.gov/pubmed/37451708 http://dx.doi.org/10.1136/bmjopen-2022-069130 |
_version_ | 1785074303520210944 |
---|---|
author | Hansen, Jessy Ahern, Susannah Earnest, Arul |
author_facet | Hansen, Jessy Ahern, Susannah Earnest, Arul |
author_sort | Hansen, Jessy |
collection | PubMed |
description | OBJECTIVES: Benchmarking is common in clinical registries to support the improvement of health outcomes by identifying underperforming clinician or health service providers. Despite the rise in clinical registries and interest in publicly reporting benchmarking results, appropriate methods for benchmarking and outlier detection within clinical registries are not well established, and the current application of methods is inconsistent. The aim of this review was to determine the current statistical methods of outlier detection that have been evaluated in the context of clinical registry benchmarking. DESIGN: A systematic search for studies evaluating the performance of methods to detect outliers when benchmarking in clinical registries was conducted in five databases: EMBASE, ProQuest, Scopus, Web of Science and Google Scholar. A modified healthcare modelling evaluation tool was used to assess quality; data extracted from each study were summarised and presented in a narrative synthesis. RESULTS: Nineteen studies evaluating a variety of statistical methods in 20 clinical registries were included. The majority of studies conducted application studies comparing outliers without statistical performance assessment (79%), while only few studies used simulations to conduct more rigorous evaluations (21%). A common comparison was between random effects and fixed effects regression, which provided mixed results. Registry population coverage, provider case volume minimum and missing data handling were all poorly reported. CONCLUSIONS: The optimal methods for detecting outliers when benchmarking clinical registry data remains unclear, and the use of different models may provide vastly different results. Further research is needed to address the unresolved methodological considerations and evaluate methods across a range of registry conditions. PROSPERO REGISTRATION NUMBER: CRD42022296520. |
format | Online Article Text |
id | pubmed-10351235 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | BMJ Publishing Group |
record_format | MEDLINE/PubMed |
spelling | pubmed-103512352023-07-18 Evaluations of statistical methods for outlier detection when benchmarking in clinical registries: a systematic review Hansen, Jessy Ahern, Susannah Earnest, Arul BMJ Open Epidemiology OBJECTIVES: Benchmarking is common in clinical registries to support the improvement of health outcomes by identifying underperforming clinician or health service providers. Despite the rise in clinical registries and interest in publicly reporting benchmarking results, appropriate methods for benchmarking and outlier detection within clinical registries are not well established, and the current application of methods is inconsistent. The aim of this review was to determine the current statistical methods of outlier detection that have been evaluated in the context of clinical registry benchmarking. DESIGN: A systematic search for studies evaluating the performance of methods to detect outliers when benchmarking in clinical registries was conducted in five databases: EMBASE, ProQuest, Scopus, Web of Science and Google Scholar. A modified healthcare modelling evaluation tool was used to assess quality; data extracted from each study were summarised and presented in a narrative synthesis. RESULTS: Nineteen studies evaluating a variety of statistical methods in 20 clinical registries were included. The majority of studies conducted application studies comparing outliers without statistical performance assessment (79%), while only few studies used simulations to conduct more rigorous evaluations (21%). A common comparison was between random effects and fixed effects regression, which provided mixed results. Registry population coverage, provider case volume minimum and missing data handling were all poorly reported. CONCLUSIONS: The optimal methods for detecting outliers when benchmarking clinical registry data remains unclear, and the use of different models may provide vastly different results. Further research is needed to address the unresolved methodological considerations and evaluate methods across a range of registry conditions. PROSPERO REGISTRATION NUMBER: CRD42022296520. BMJ Publishing Group 2023-07-14 /pmc/articles/PMC10351235/ /pubmed/37451708 http://dx.doi.org/10.1136/bmjopen-2022-069130 Text en © Author(s) (or their employer(s)) 2023. Re-use permitted under CC BY-NC. No commercial re-use. See rights and permissions. Published by BMJ. https://creativecommons.org/licenses/by-nc/4.0/This is an open access article distributed in accordance with the Creative Commons Attribution Non Commercial (CC BY-NC 4.0) license, which permits others to distribute, remix, adapt, build upon this work non-commercially, and license their derivative works on different terms, provided the original work is properly cited, appropriate credit is given, any changes made indicated, and the use is non-commercial. See: http://creativecommons.org/licenses/by-nc/4.0/ (https://creativecommons.org/licenses/by-nc/4.0/) . |
spellingShingle | Epidemiology Hansen, Jessy Ahern, Susannah Earnest, Arul Evaluations of statistical methods for outlier detection when benchmarking in clinical registries: a systematic review |
title | Evaluations of statistical methods for outlier detection when benchmarking in clinical registries: a systematic review |
title_full | Evaluations of statistical methods for outlier detection when benchmarking in clinical registries: a systematic review |
title_fullStr | Evaluations of statistical methods for outlier detection when benchmarking in clinical registries: a systematic review |
title_full_unstemmed | Evaluations of statistical methods for outlier detection when benchmarking in clinical registries: a systematic review |
title_short | Evaluations of statistical methods for outlier detection when benchmarking in clinical registries: a systematic review |
title_sort | evaluations of statistical methods for outlier detection when benchmarking in clinical registries: a systematic review |
topic | Epidemiology |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10351235/ https://www.ncbi.nlm.nih.gov/pubmed/37451708 http://dx.doi.org/10.1136/bmjopen-2022-069130 |
work_keys_str_mv | AT hansenjessy evaluationsofstatisticalmethodsforoutlierdetectionwhenbenchmarkinginclinicalregistriesasystematicreview AT ahernsusannah evaluationsofstatisticalmethodsforoutlierdetectionwhenbenchmarkinginclinicalregistriesasystematicreview AT earnestarul evaluationsofstatisticalmethodsforoutlierdetectionwhenbenchmarkinginclinicalregistriesasystematicreview |