Cargando…

Appraising prediction research: a guide and meta‐review on bias and applicability assessment using the Prediction model Risk Of Bias ASsessment Tool (PROBAST)

Over the past few years, a large number of prediction models have been published, often of poor methodological quality. Seemingly objective and straightforward, prediction models provide a risk estimate for the outcome of interest, usually based on readily available clinical information. Yet, using...

Descripción completa

Detalles Bibliográficos
Autores principales: de Jong, Ype, Ramspek, Chava L., Zoccali, Carmine, Jager, Kitty J., Dekker, Friedo W., van Diepen, Merel
Formato: Online Artículo Texto
Lenguaje:English
Publicado: John Wiley & Sons Australia, Ltd 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9291738/
https://www.ncbi.nlm.nih.gov/pubmed/34138495
http://dx.doi.org/10.1111/nep.13913
_version_ 1784749201466327040
author de Jong, Ype
Ramspek, Chava L.
Zoccali, Carmine
Jager, Kitty J.
Dekker, Friedo W.
van Diepen, Merel
author_facet de Jong, Ype
Ramspek, Chava L.
Zoccali, Carmine
Jager, Kitty J.
Dekker, Friedo W.
van Diepen, Merel
author_sort de Jong, Ype
collection PubMed
description Over the past few years, a large number of prediction models have been published, often of poor methodological quality. Seemingly objective and straightforward, prediction models provide a risk estimate for the outcome of interest, usually based on readily available clinical information. Yet, using models of substandard methodological rigour, especially without external validation, may result in incorrect risk estimates and consequently misclassification. To assess and combat bias in prediction research the prediction model risk of bias assessment tool (PROBAST) was published in 2019. This risk of bias (ROB) tool includes four domains and 20 signalling questions highlighting methodological flaws, and provides guidance in assessing the applicability of the model. In this paper, the PROBAST will be discussed, along with an in‐depth review of two commonly encountered pitfalls in prediction modelling that may induce bias: overfitting and composite endpoints. We illustrate the prevalence of potential bias in prediction models with a meta‐review of 50 systematic reviews that used the PROBAST to appraise their included studies, thus including 1510 different studies on 2104 prediction models. All domains showed an unclear or high ROB; these results were markedly stable over time, highlighting the urgent need for attention on bias in prediction research. This article aims to do just that by providing (1) the clinician with tools to evaluate the (methodological) quality of a clinical prediction model, (2) the researcher working on a review with methods to appraise the included models, and (3) the researcher developing a model with suggestions to improve model quality.
format Online
Article
Text
id pubmed-9291738
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher John Wiley & Sons Australia, Ltd
record_format MEDLINE/PubMed
spelling pubmed-92917382022-07-20 Appraising prediction research: a guide and meta‐review on bias and applicability assessment using the Prediction model Risk Of Bias ASsessment Tool (PROBAST) de Jong, Ype Ramspek, Chava L. Zoccali, Carmine Jager, Kitty J. Dekker, Friedo W. van Diepen, Merel Nephrology (Carlton) Review Articles Over the past few years, a large number of prediction models have been published, often of poor methodological quality. Seemingly objective and straightforward, prediction models provide a risk estimate for the outcome of interest, usually based on readily available clinical information. Yet, using models of substandard methodological rigour, especially without external validation, may result in incorrect risk estimates and consequently misclassification. To assess and combat bias in prediction research the prediction model risk of bias assessment tool (PROBAST) was published in 2019. This risk of bias (ROB) tool includes four domains and 20 signalling questions highlighting methodological flaws, and provides guidance in assessing the applicability of the model. In this paper, the PROBAST will be discussed, along with an in‐depth review of two commonly encountered pitfalls in prediction modelling that may induce bias: overfitting and composite endpoints. We illustrate the prevalence of potential bias in prediction models with a meta‐review of 50 systematic reviews that used the PROBAST to appraise their included studies, thus including 1510 different studies on 2104 prediction models. All domains showed an unclear or high ROB; these results were markedly stable over time, highlighting the urgent need for attention on bias in prediction research. This article aims to do just that by providing (1) the clinician with tools to evaluate the (methodological) quality of a clinical prediction model, (2) the researcher working on a review with methods to appraise the included models, and (3) the researcher developing a model with suggestions to improve model quality. John Wiley & Sons Australia, Ltd 2021-07-08 2021-12 /pmc/articles/PMC9291738/ /pubmed/34138495 http://dx.doi.org/10.1111/nep.13913 Text en © 2021 The Authors. Nephrology published by John Wiley & Sons Australia, Ltd on behalf of Asian Pacific Society of Nephrology. https://creativecommons.org/licenses/by-nc/4.0/This is an open access article under the terms of the http://creativecommons.org/licenses/by-nc/4.0/ (https://creativecommons.org/licenses/by-nc/4.0/) License, which permits use, distribution and reproduction in any medium, provided the original work is properly cited and is not used for commercial purposes.
spellingShingle Review Articles
de Jong, Ype
Ramspek, Chava L.
Zoccali, Carmine
Jager, Kitty J.
Dekker, Friedo W.
van Diepen, Merel
Appraising prediction research: a guide and meta‐review on bias and applicability assessment using the Prediction model Risk Of Bias ASsessment Tool (PROBAST)
title Appraising prediction research: a guide and meta‐review on bias and applicability assessment using the Prediction model Risk Of Bias ASsessment Tool (PROBAST)
title_full Appraising prediction research: a guide and meta‐review on bias and applicability assessment using the Prediction model Risk Of Bias ASsessment Tool (PROBAST)
title_fullStr Appraising prediction research: a guide and meta‐review on bias and applicability assessment using the Prediction model Risk Of Bias ASsessment Tool (PROBAST)
title_full_unstemmed Appraising prediction research: a guide and meta‐review on bias and applicability assessment using the Prediction model Risk Of Bias ASsessment Tool (PROBAST)
title_short Appraising prediction research: a guide and meta‐review on bias and applicability assessment using the Prediction model Risk Of Bias ASsessment Tool (PROBAST)
title_sort appraising prediction research: a guide and meta‐review on bias and applicability assessment using the prediction model risk of bias assessment tool (probast)
topic Review Articles
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9291738/
https://www.ncbi.nlm.nih.gov/pubmed/34138495
http://dx.doi.org/10.1111/nep.13913
work_keys_str_mv AT dejongype appraisingpredictionresearchaguideandmetareviewonbiasandapplicabilityassessmentusingthepredictionmodelriskofbiasassessmenttoolprobast
AT ramspekchaval appraisingpredictionresearchaguideandmetareviewonbiasandapplicabilityassessmentusingthepredictionmodelriskofbiasassessmenttoolprobast
AT zoccalicarmine appraisingpredictionresearchaguideandmetareviewonbiasandapplicabilityassessmentusingthepredictionmodelriskofbiasassessmenttoolprobast
AT jagerkittyj appraisingpredictionresearchaguideandmetareviewonbiasandapplicabilityassessmentusingthepredictionmodelriskofbiasassessmenttoolprobast
AT dekkerfriedow appraisingpredictionresearchaguideandmetareviewonbiasandapplicabilityassessmentusingthepredictionmodelriskofbiasassessmenttoolprobast
AT vandiepenmerel appraisingpredictionresearchaguideandmetareviewonbiasandapplicabilityassessmentusingthepredictionmodelriskofbiasassessmenttoolprobast