Cargando…
Appraising prediction research: a guide and meta‐review on bias and applicability assessment using the Prediction model Risk Of Bias ASsessment Tool (PROBAST)
Over the past few years, a large number of prediction models have been published, often of poor methodological quality. Seemingly objective and straightforward, prediction models provide a risk estimate for the outcome of interest, usually based on readily available clinical information. Yet, using...
Autores principales: | , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
John Wiley & Sons Australia, Ltd
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9291738/ https://www.ncbi.nlm.nih.gov/pubmed/34138495 http://dx.doi.org/10.1111/nep.13913 |
_version_ | 1784749201466327040 |
---|---|
author | de Jong, Ype Ramspek, Chava L. Zoccali, Carmine Jager, Kitty J. Dekker, Friedo W. van Diepen, Merel |
author_facet | de Jong, Ype Ramspek, Chava L. Zoccali, Carmine Jager, Kitty J. Dekker, Friedo W. van Diepen, Merel |
author_sort | de Jong, Ype |
collection | PubMed |
description | Over the past few years, a large number of prediction models have been published, often of poor methodological quality. Seemingly objective and straightforward, prediction models provide a risk estimate for the outcome of interest, usually based on readily available clinical information. Yet, using models of substandard methodological rigour, especially without external validation, may result in incorrect risk estimates and consequently misclassification. To assess and combat bias in prediction research the prediction model risk of bias assessment tool (PROBAST) was published in 2019. This risk of bias (ROB) tool includes four domains and 20 signalling questions highlighting methodological flaws, and provides guidance in assessing the applicability of the model. In this paper, the PROBAST will be discussed, along with an in‐depth review of two commonly encountered pitfalls in prediction modelling that may induce bias: overfitting and composite endpoints. We illustrate the prevalence of potential bias in prediction models with a meta‐review of 50 systematic reviews that used the PROBAST to appraise their included studies, thus including 1510 different studies on 2104 prediction models. All domains showed an unclear or high ROB; these results were markedly stable over time, highlighting the urgent need for attention on bias in prediction research. This article aims to do just that by providing (1) the clinician with tools to evaluate the (methodological) quality of a clinical prediction model, (2) the researcher working on a review with methods to appraise the included models, and (3) the researcher developing a model with suggestions to improve model quality. |
format | Online Article Text |
id | pubmed-9291738 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | John Wiley & Sons Australia, Ltd |
record_format | MEDLINE/PubMed |
spelling | pubmed-92917382022-07-20 Appraising prediction research: a guide and meta‐review on bias and applicability assessment using the Prediction model Risk Of Bias ASsessment Tool (PROBAST) de Jong, Ype Ramspek, Chava L. Zoccali, Carmine Jager, Kitty J. Dekker, Friedo W. van Diepen, Merel Nephrology (Carlton) Review Articles Over the past few years, a large number of prediction models have been published, often of poor methodological quality. Seemingly objective and straightforward, prediction models provide a risk estimate for the outcome of interest, usually based on readily available clinical information. Yet, using models of substandard methodological rigour, especially without external validation, may result in incorrect risk estimates and consequently misclassification. To assess and combat bias in prediction research the prediction model risk of bias assessment tool (PROBAST) was published in 2019. This risk of bias (ROB) tool includes four domains and 20 signalling questions highlighting methodological flaws, and provides guidance in assessing the applicability of the model. In this paper, the PROBAST will be discussed, along with an in‐depth review of two commonly encountered pitfalls in prediction modelling that may induce bias: overfitting and composite endpoints. We illustrate the prevalence of potential bias in prediction models with a meta‐review of 50 systematic reviews that used the PROBAST to appraise their included studies, thus including 1510 different studies on 2104 prediction models. All domains showed an unclear or high ROB; these results were markedly stable over time, highlighting the urgent need for attention on bias in prediction research. This article aims to do just that by providing (1) the clinician with tools to evaluate the (methodological) quality of a clinical prediction model, (2) the researcher working on a review with methods to appraise the included models, and (3) the researcher developing a model with suggestions to improve model quality. John Wiley & Sons Australia, Ltd 2021-07-08 2021-12 /pmc/articles/PMC9291738/ /pubmed/34138495 http://dx.doi.org/10.1111/nep.13913 Text en © 2021 The Authors. Nephrology published by John Wiley & Sons Australia, Ltd on behalf of Asian Pacific Society of Nephrology. https://creativecommons.org/licenses/by-nc/4.0/This is an open access article under the terms of the http://creativecommons.org/licenses/by-nc/4.0/ (https://creativecommons.org/licenses/by-nc/4.0/) License, which permits use, distribution and reproduction in any medium, provided the original work is properly cited and is not used for commercial purposes. |
spellingShingle | Review Articles de Jong, Ype Ramspek, Chava L. Zoccali, Carmine Jager, Kitty J. Dekker, Friedo W. van Diepen, Merel Appraising prediction research: a guide and meta‐review on bias and applicability assessment using the Prediction model Risk Of Bias ASsessment Tool (PROBAST) |
title | Appraising prediction research: a guide and meta‐review on bias and applicability assessment using the Prediction model Risk Of Bias ASsessment Tool (PROBAST) |
title_full | Appraising prediction research: a guide and meta‐review on bias and applicability assessment using the Prediction model Risk Of Bias ASsessment Tool (PROBAST) |
title_fullStr | Appraising prediction research: a guide and meta‐review on bias and applicability assessment using the Prediction model Risk Of Bias ASsessment Tool (PROBAST) |
title_full_unstemmed | Appraising prediction research: a guide and meta‐review on bias and applicability assessment using the Prediction model Risk Of Bias ASsessment Tool (PROBAST) |
title_short | Appraising prediction research: a guide and meta‐review on bias and applicability assessment using the Prediction model Risk Of Bias ASsessment Tool (PROBAST) |
title_sort | appraising prediction research: a guide and meta‐review on bias and applicability assessment using the prediction model risk of bias assessment tool (probast) |
topic | Review Articles |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9291738/ https://www.ncbi.nlm.nih.gov/pubmed/34138495 http://dx.doi.org/10.1111/nep.13913 |
work_keys_str_mv | AT dejongype appraisingpredictionresearchaguideandmetareviewonbiasandapplicabilityassessmentusingthepredictionmodelriskofbiasassessmenttoolprobast AT ramspekchaval appraisingpredictionresearchaguideandmetareviewonbiasandapplicabilityassessmentusingthepredictionmodelriskofbiasassessmenttoolprobast AT zoccalicarmine appraisingpredictionresearchaguideandmetareviewonbiasandapplicabilityassessmentusingthepredictionmodelriskofbiasassessmenttoolprobast AT jagerkittyj appraisingpredictionresearchaguideandmetareviewonbiasandapplicabilityassessmentusingthepredictionmodelriskofbiasassessmenttoolprobast AT dekkerfriedow appraisingpredictionresearchaguideandmetareviewonbiasandapplicabilityassessmentusingthepredictionmodelriskofbiasassessmenttoolprobast AT vandiepenmerel appraisingpredictionresearchaguideandmetareviewonbiasandapplicabilityassessmentusingthepredictionmodelriskofbiasassessmenttoolprobast |