Cargando…
Meta-analysis of prediction model performance across multiple studies: Which scale helps ensure between-study normality for the C-statistic and calibration measures?
If individual participant data are available from multiple studies or clusters, then a prediction model can be externally validated multiple times. This allows the model’s discrimination and calibration performance to be examined across different settings. Random-effects meta-analysis can then be us...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
SAGE Publications
2017
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6193210/ https://www.ncbi.nlm.nih.gov/pubmed/28480827 http://dx.doi.org/10.1177/0962280217705678 |
_version_ | 1783364040126365696 |
---|---|
author | Snell, Kym IE Ensor, Joie Debray, Thomas PA Moons, Karel GM Riley, Richard D |
author_facet | Snell, Kym IE Ensor, Joie Debray, Thomas PA Moons, Karel GM Riley, Richard D |
author_sort | Snell, Kym IE |
collection | PubMed |
description | If individual participant data are available from multiple studies or clusters, then a prediction model can be externally validated multiple times. This allows the model’s discrimination and calibration performance to be examined across different settings. Random-effects meta-analysis can then be used to quantify overall (average) performance and heterogeneity in performance. This typically assumes a normal distribution of ‘true’ performance across studies. We conducted a simulation study to examine this normality assumption for various performance measures relating to a logistic regression prediction model. We simulated data across multiple studies with varying degrees of variability in baseline risk or predictor effects and then evaluated the shape of the between-study distribution in the C-statistic, calibration slope, calibration-in-the-large, and E/O statistic, and possible transformations thereof. We found that a normal between-study distribution was usually reasonable for the calibration slope and calibration-in-the-large; however, the distributions of the C-statistic and E/O were often skewed across studies, particularly in settings with large variability in the predictor effects. Normality was vastly improved when using the logit transformation for the C-statistic and the log transformation for E/O, and therefore we recommend these scales to be used for meta-analysis. An illustrated example is given using a random-effects meta-analysis of the performance of QRISK2 across 25 general practices. |
format | Online Article Text |
id | pubmed-6193210 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2017 |
publisher | SAGE Publications |
record_format | MEDLINE/PubMed |
spelling | pubmed-61932102018-10-24 Meta-analysis of prediction model performance across multiple studies: Which scale helps ensure between-study normality for the C-statistic and calibration measures? Snell, Kym IE Ensor, Joie Debray, Thomas PA Moons, Karel GM Riley, Richard D Stat Methods Med Res Articles If individual participant data are available from multiple studies or clusters, then a prediction model can be externally validated multiple times. This allows the model’s discrimination and calibration performance to be examined across different settings. Random-effects meta-analysis can then be used to quantify overall (average) performance and heterogeneity in performance. This typically assumes a normal distribution of ‘true’ performance across studies. We conducted a simulation study to examine this normality assumption for various performance measures relating to a logistic regression prediction model. We simulated data across multiple studies with varying degrees of variability in baseline risk or predictor effects and then evaluated the shape of the between-study distribution in the C-statistic, calibration slope, calibration-in-the-large, and E/O statistic, and possible transformations thereof. We found that a normal between-study distribution was usually reasonable for the calibration slope and calibration-in-the-large; however, the distributions of the C-statistic and E/O were often skewed across studies, particularly in settings with large variability in the predictor effects. Normality was vastly improved when using the logit transformation for the C-statistic and the log transformation for E/O, and therefore we recommend these scales to be used for meta-analysis. An illustrated example is given using a random-effects meta-analysis of the performance of QRISK2 across 25 general practices. SAGE Publications 2017-05-08 2018-11 /pmc/articles/PMC6193210/ /pubmed/28480827 http://dx.doi.org/10.1177/0962280217705678 Text en © The Author(s) 2017 http://creativecommons.org/licenses/by/4.0/ This article is distributed under the terms of the Creative Commons Attribution 4.0 License (http://www.creativecommons.org/licenses/by/4.0/) which permits any use, reproduction and distribution of the work without further permission provided the original work is attributed as specified on the SAGE and Open Access pages (https://us.sagepub.com/en-us/nam/open-access-at-sage). |
spellingShingle | Articles Snell, Kym IE Ensor, Joie Debray, Thomas PA Moons, Karel GM Riley, Richard D Meta-analysis of prediction model performance across multiple studies: Which scale helps ensure between-study normality for the C-statistic and calibration measures? |
title | Meta-analysis of prediction model performance across multiple
studies: Which scale helps ensure between-study normality for the
C-statistic and calibration measures? |
title_full | Meta-analysis of prediction model performance across multiple
studies: Which scale helps ensure between-study normality for the
C-statistic and calibration measures? |
title_fullStr | Meta-analysis of prediction model performance across multiple
studies: Which scale helps ensure between-study normality for the
C-statistic and calibration measures? |
title_full_unstemmed | Meta-analysis of prediction model performance across multiple
studies: Which scale helps ensure between-study normality for the
C-statistic and calibration measures? |
title_short | Meta-analysis of prediction model performance across multiple
studies: Which scale helps ensure between-study normality for the
C-statistic and calibration measures? |
title_sort | meta-analysis of prediction model performance across multiple
studies: which scale helps ensure between-study normality for the
c-statistic and calibration measures? |
topic | Articles |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6193210/ https://www.ncbi.nlm.nih.gov/pubmed/28480827 http://dx.doi.org/10.1177/0962280217705678 |
work_keys_str_mv | AT snellkymie metaanalysisofpredictionmodelperformanceacrossmultiplestudieswhichscalehelpsensurebetweenstudynormalityforthecstatisticandcalibrationmeasures AT ensorjoie metaanalysisofpredictionmodelperformanceacrossmultiplestudieswhichscalehelpsensurebetweenstudynormalityforthecstatisticandcalibrationmeasures AT debraythomaspa metaanalysisofpredictionmodelperformanceacrossmultiplestudieswhichscalehelpsensurebetweenstudynormalityforthecstatisticandcalibrationmeasures AT moonskarelgm metaanalysisofpredictionmodelperformanceacrossmultiplestudieswhichscalehelpsensurebetweenstudynormalityforthecstatisticandcalibrationmeasures AT rileyrichardd metaanalysisofpredictionmodelperformanceacrossmultiplestudieswhichscalehelpsensurebetweenstudynormalityforthecstatisticandcalibrationmeasures |