Cargando…

Cross-validation pitfalls when selecting and assessing regression and classification models

BACKGROUND: We address the problem of selecting and assessing classification and regression models using cross-validation. Current state-of-the-art methods can yield models with high variance, rendering them unsuitable for a number of practical applications including QSAR. In this paper we describe...

Descripción completa

Detalles Bibliográficos
Autores principales: Krstajic, Damjan, Buturovic, Ljubomir J, Leahy, David E, Thomas, Simon
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Springer International Publishing 2014
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3994246/
https://www.ncbi.nlm.nih.gov/pubmed/24678909
http://dx.doi.org/10.1186/1758-2946-6-10
_version_ 1782312694514188288
author Krstajic, Damjan
Buturovic, Ljubomir J
Leahy, David E
Thomas, Simon
author_facet Krstajic, Damjan
Buturovic, Ljubomir J
Leahy, David E
Thomas, Simon
author_sort Krstajic, Damjan
collection PubMed
description BACKGROUND: We address the problem of selecting and assessing classification and regression models using cross-validation. Current state-of-the-art methods can yield models with high variance, rendering them unsuitable for a number of practical applications including QSAR. In this paper we describe and evaluate best practices which improve reliability and increase confidence in selected models. A key operational component of the proposed methods is cloud computing which enables routine use of previously infeasible approaches. METHODS: We describe in detail an algorithm for repeated grid-search V-fold cross-validation for parameter tuning in classification and regression, and we define a repeated nested cross-validation algorithm for model assessment. As regards variable selection and parameter tuning we define two algorithms (repeated grid-search cross-validation and double cross-validation), and provide arguments for using the repeated grid-search in the general case. RESULTS: We show results of our algorithms on seven QSAR datasets. The variation of the prediction performance, which is the result of choosing different splits of the dataset in V-fold cross-validation, needs to be taken into account when selecting and assessing classification and regression models. CONCLUSIONS: We demonstrate the importance of repeating cross-validation when selecting an optimal model, as well as the importance of repeating nested cross-validation when assessing a prediction error. ELECTRONIC SUPPLEMENTARY MATERIAL: The online version of this article (doi:10.1186/1758-2946-6-10) contains supplementary material, which is available to authorized users.
format Online
Article
Text
id pubmed-3994246
institution National Center for Biotechnology Information
language English
publishDate 2014
publisher Springer International Publishing
record_format MEDLINE/PubMed
spelling pubmed-39942462014-04-23 Cross-validation pitfalls when selecting and assessing regression and classification models Krstajic, Damjan Buturovic, Ljubomir J Leahy, David E Thomas, Simon J Cheminform Methodology BACKGROUND: We address the problem of selecting and assessing classification and regression models using cross-validation. Current state-of-the-art methods can yield models with high variance, rendering them unsuitable for a number of practical applications including QSAR. In this paper we describe and evaluate best practices which improve reliability and increase confidence in selected models. A key operational component of the proposed methods is cloud computing which enables routine use of previously infeasible approaches. METHODS: We describe in detail an algorithm for repeated grid-search V-fold cross-validation for parameter tuning in classification and regression, and we define a repeated nested cross-validation algorithm for model assessment. As regards variable selection and parameter tuning we define two algorithms (repeated grid-search cross-validation and double cross-validation), and provide arguments for using the repeated grid-search in the general case. RESULTS: We show results of our algorithms on seven QSAR datasets. The variation of the prediction performance, which is the result of choosing different splits of the dataset in V-fold cross-validation, needs to be taken into account when selecting and assessing classification and regression models. CONCLUSIONS: We demonstrate the importance of repeating cross-validation when selecting an optimal model, as well as the importance of repeating nested cross-validation when assessing a prediction error. ELECTRONIC SUPPLEMENTARY MATERIAL: The online version of this article (doi:10.1186/1758-2946-6-10) contains supplementary material, which is available to authorized users. Springer International Publishing 2014-03-29 /pmc/articles/PMC3994246/ /pubmed/24678909 http://dx.doi.org/10.1186/1758-2946-6-10 Text en © Krstajic et al.; licensee Chemistry Central Ltd. 2014 This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly credited. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.
spellingShingle Methodology
Krstajic, Damjan
Buturovic, Ljubomir J
Leahy, David E
Thomas, Simon
Cross-validation pitfalls when selecting and assessing regression and classification models
title Cross-validation pitfalls when selecting and assessing regression and classification models
title_full Cross-validation pitfalls when selecting and assessing regression and classification models
title_fullStr Cross-validation pitfalls when selecting and assessing regression and classification models
title_full_unstemmed Cross-validation pitfalls when selecting and assessing regression and classification models
title_short Cross-validation pitfalls when selecting and assessing regression and classification models
title_sort cross-validation pitfalls when selecting and assessing regression and classification models
topic Methodology
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3994246/
https://www.ncbi.nlm.nih.gov/pubmed/24678909
http://dx.doi.org/10.1186/1758-2946-6-10
work_keys_str_mv AT krstajicdamjan crossvalidationpitfallswhenselectingandassessingregressionandclassificationmodels
AT buturovicljubomirj crossvalidationpitfallswhenselectingandassessingregressionandclassificationmodels
AT leahydavide crossvalidationpitfallswhenselectingandassessingregressionandclassificationmodels
AT thomassimon crossvalidationpitfallswhenselectingandassessingregressionandclassificationmodels