Cargando…

Events per variable (EPV) and the relative performance of different strategies for estimating the out-of-sample validity of logistic regression models

We conducted an extensive set of empirical analyses to examine the effect of the number of events per variable (EPV) on the relative performance of three different methods for assessing the predictive accuracy of a logistic regression model: apparent performance in the analysis sample, split-sample...

Descripción completa

Detalles Bibliográficos
Autores principales: Austin, Peter C, Steyerberg, Ewout W
Formato: Online Artículo Texto
Lenguaje:English
Publicado: SAGE Publications 2014
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5394463/
https://www.ncbi.nlm.nih.gov/pubmed/25411322
http://dx.doi.org/10.1177/0962280214558972
_version_ 1783229735261700096
author Austin, Peter C
Steyerberg, Ewout W
author_facet Austin, Peter C
Steyerberg, Ewout W
author_sort Austin, Peter C
collection PubMed
description We conducted an extensive set of empirical analyses to examine the effect of the number of events per variable (EPV) on the relative performance of three different methods for assessing the predictive accuracy of a logistic regression model: apparent performance in the analysis sample, split-sample validation, and optimism correction using bootstrap methods. Using a single dataset of patients hospitalized with heart failure, we compared the estimates of discriminatory performance from these methods to those for a very large independent validation sample arising from the same population. As anticipated, the apparent performance was optimistically biased, with the degree of optimism diminishing as the number of events per variable increased. Differences between the bootstrap-corrected approach and the use of an independent validation sample were minimal once the number of events per variable was at least 20. Split-sample assessment resulted in too pessimistic and highly uncertain estimates of model performance. Apparent performance estimates had lower mean squared error compared to split-sample estimates, but the lowest mean squared error was obtained by bootstrap-corrected optimism estimates. For bias, variance, and mean squared error of the performance estimates, the penalty incurred by using split-sample validation was equivalent to reducing the sample size by a proportion equivalent to the proportion of the sample that was withheld for model validation. In conclusion, split-sample validation is inefficient and apparent performance is too optimistic for internal validation of regression-based prediction models. Modern validation methods, such as bootstrap-based optimism correction, are preferable. While these findings may be unsurprising to many statisticians, the results of the current study reinforce what should be considered good statistical practice in the development and validation of clinical prediction models.
format Online
Article
Text
id pubmed-5394463
institution National Center for Biotechnology Information
language English
publishDate 2014
publisher SAGE Publications
record_format MEDLINE/PubMed
spelling pubmed-53944632017-04-26 Events per variable (EPV) and the relative performance of different strategies for estimating the out-of-sample validity of logistic regression models Austin, Peter C Steyerberg, Ewout W Stat Methods Med Res Articles We conducted an extensive set of empirical analyses to examine the effect of the number of events per variable (EPV) on the relative performance of three different methods for assessing the predictive accuracy of a logistic regression model: apparent performance in the analysis sample, split-sample validation, and optimism correction using bootstrap methods. Using a single dataset of patients hospitalized with heart failure, we compared the estimates of discriminatory performance from these methods to those for a very large independent validation sample arising from the same population. As anticipated, the apparent performance was optimistically biased, with the degree of optimism diminishing as the number of events per variable increased. Differences between the bootstrap-corrected approach and the use of an independent validation sample were minimal once the number of events per variable was at least 20. Split-sample assessment resulted in too pessimistic and highly uncertain estimates of model performance. Apparent performance estimates had lower mean squared error compared to split-sample estimates, but the lowest mean squared error was obtained by bootstrap-corrected optimism estimates. For bias, variance, and mean squared error of the performance estimates, the penalty incurred by using split-sample validation was equivalent to reducing the sample size by a proportion equivalent to the proportion of the sample that was withheld for model validation. In conclusion, split-sample validation is inefficient and apparent performance is too optimistic for internal validation of regression-based prediction models. Modern validation methods, such as bootstrap-based optimism correction, are preferable. While these findings may be unsurprising to many statisticians, the results of the current study reinforce what should be considered good statistical practice in the development and validation of clinical prediction models. SAGE Publications 2014-11-19 2017-04 /pmc/articles/PMC5394463/ /pubmed/25411322 http://dx.doi.org/10.1177/0962280214558972 Text en © The Author(s) 2014 http://creativecommons.org/licenses/by-nc/3.0/ This article is distributed under the terms of the Creative Commons Attribution-NonCommercial 3.0 License (http://www.creativecommons.org/licenses/by-nc/3.0/) which permits non-commercial use, reproduction and distribution of the work without further permission provided the original work is attributed as specified on the SAGE and Open Access page(https://us.sagepub.com/en-us/nam/open-access-at-sage).
spellingShingle Articles
Austin, Peter C
Steyerberg, Ewout W
Events per variable (EPV) and the relative performance of different strategies for estimating the out-of-sample validity of logistic regression models
title Events per variable (EPV) and the relative performance of different strategies for estimating the out-of-sample validity of logistic regression models
title_full Events per variable (EPV) and the relative performance of different strategies for estimating the out-of-sample validity of logistic regression models
title_fullStr Events per variable (EPV) and the relative performance of different strategies for estimating the out-of-sample validity of logistic regression models
title_full_unstemmed Events per variable (EPV) and the relative performance of different strategies for estimating the out-of-sample validity of logistic regression models
title_short Events per variable (EPV) and the relative performance of different strategies for estimating the out-of-sample validity of logistic regression models
title_sort events per variable (epv) and the relative performance of different strategies for estimating the out-of-sample validity of logistic regression models
topic Articles
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5394463/
https://www.ncbi.nlm.nih.gov/pubmed/25411322
http://dx.doi.org/10.1177/0962280214558972
work_keys_str_mv AT austinpeterc eventspervariableepvandtherelativeperformanceofdifferentstrategiesforestimatingtheoutofsamplevalidityoflogisticregressionmodels
AT steyerbergewoutw eventspervariableepvandtherelativeperformanceofdifferentstrategiesforestimatingtheoutofsamplevalidityoflogisticregressionmodels