Cargando…
Cross-validation pitfalls when selecting and assessing regression and classification models
BACKGROUND: We address the problem of selecting and assessing classification and regression models using cross-validation. Current state-of-the-art methods can yield models with high variance, rendering them unsuitable for a number of practical applications including QSAR. In this paper we describe...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Springer International Publishing
2014
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3994246/ https://www.ncbi.nlm.nih.gov/pubmed/24678909 http://dx.doi.org/10.1186/1758-2946-6-10 |
_version_ | 1782312694514188288 |
---|---|
author | Krstajic, Damjan Buturovic, Ljubomir J Leahy, David E Thomas, Simon |
author_facet | Krstajic, Damjan Buturovic, Ljubomir J Leahy, David E Thomas, Simon |
author_sort | Krstajic, Damjan |
collection | PubMed |
description | BACKGROUND: We address the problem of selecting and assessing classification and regression models using cross-validation. Current state-of-the-art methods can yield models with high variance, rendering them unsuitable for a number of practical applications including QSAR. In this paper we describe and evaluate best practices which improve reliability and increase confidence in selected models. A key operational component of the proposed methods is cloud computing which enables routine use of previously infeasible approaches. METHODS: We describe in detail an algorithm for repeated grid-search V-fold cross-validation for parameter tuning in classification and regression, and we define a repeated nested cross-validation algorithm for model assessment. As regards variable selection and parameter tuning we define two algorithms (repeated grid-search cross-validation and double cross-validation), and provide arguments for using the repeated grid-search in the general case. RESULTS: We show results of our algorithms on seven QSAR datasets. The variation of the prediction performance, which is the result of choosing different splits of the dataset in V-fold cross-validation, needs to be taken into account when selecting and assessing classification and regression models. CONCLUSIONS: We demonstrate the importance of repeating cross-validation when selecting an optimal model, as well as the importance of repeating nested cross-validation when assessing a prediction error. ELECTRONIC SUPPLEMENTARY MATERIAL: The online version of this article (doi:10.1186/1758-2946-6-10) contains supplementary material, which is available to authorized users. |
format | Online Article Text |
id | pubmed-3994246 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2014 |
publisher | Springer International Publishing |
record_format | MEDLINE/PubMed |
spelling | pubmed-39942462014-04-23 Cross-validation pitfalls when selecting and assessing regression and classification models Krstajic, Damjan Buturovic, Ljubomir J Leahy, David E Thomas, Simon J Cheminform Methodology BACKGROUND: We address the problem of selecting and assessing classification and regression models using cross-validation. Current state-of-the-art methods can yield models with high variance, rendering them unsuitable for a number of practical applications including QSAR. In this paper we describe and evaluate best practices which improve reliability and increase confidence in selected models. A key operational component of the proposed methods is cloud computing which enables routine use of previously infeasible approaches. METHODS: We describe in detail an algorithm for repeated grid-search V-fold cross-validation for parameter tuning in classification and regression, and we define a repeated nested cross-validation algorithm for model assessment. As regards variable selection and parameter tuning we define two algorithms (repeated grid-search cross-validation and double cross-validation), and provide arguments for using the repeated grid-search in the general case. RESULTS: We show results of our algorithms on seven QSAR datasets. The variation of the prediction performance, which is the result of choosing different splits of the dataset in V-fold cross-validation, needs to be taken into account when selecting and assessing classification and regression models. CONCLUSIONS: We demonstrate the importance of repeating cross-validation when selecting an optimal model, as well as the importance of repeating nested cross-validation when assessing a prediction error. ELECTRONIC SUPPLEMENTARY MATERIAL: The online version of this article (doi:10.1186/1758-2946-6-10) contains supplementary material, which is available to authorized users. Springer International Publishing 2014-03-29 /pmc/articles/PMC3994246/ /pubmed/24678909 http://dx.doi.org/10.1186/1758-2946-6-10 Text en © Krstajic et al.; licensee Chemistry Central Ltd. 2014 This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly credited. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated. |
spellingShingle | Methodology Krstajic, Damjan Buturovic, Ljubomir J Leahy, David E Thomas, Simon Cross-validation pitfalls when selecting and assessing regression and classification models |
title | Cross-validation pitfalls when selecting and assessing regression and classification models |
title_full | Cross-validation pitfalls when selecting and assessing regression and classification models |
title_fullStr | Cross-validation pitfalls when selecting and assessing regression and classification models |
title_full_unstemmed | Cross-validation pitfalls when selecting and assessing regression and classification models |
title_short | Cross-validation pitfalls when selecting and assessing regression and classification models |
title_sort | cross-validation pitfalls when selecting and assessing regression and classification models |
topic | Methodology |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3994246/ https://www.ncbi.nlm.nih.gov/pubmed/24678909 http://dx.doi.org/10.1186/1758-2946-6-10 |
work_keys_str_mv | AT krstajicdamjan crossvalidationpitfallswhenselectingandassessingregressionandclassificationmodels AT buturovicljubomirj crossvalidationpitfallswhenselectingandassessingregressionandclassificationmodels AT leahydavide crossvalidationpitfallswhenselectingandassessingregressionandclassificationmodels AT thomassimon crossvalidationpitfallswhenselectingandassessingregressionandclassificationmodels |