Cargando…

Evaluation of multiple prediction models: A novel view on model selection and performance assessment

Model selection and performance assessment for prediction models are important tasks in machine learning, e.g. for the development of medical diagnosis or prognosis rules based on complex data. A common approach is to select the best model via cross-validation and to evaluate this final model on an...

Descripción completa

Detalles Bibliográficos
Autores principales: Westphal, Max, Brannath, Werner
Formato: Online Artículo Texto
Lenguaje:English
Publicado: SAGE Publications 2019
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7270727/
https://www.ncbi.nlm.nih.gov/pubmed/31510862
http://dx.doi.org/10.1177/0962280219854487
_version_ 1783541946766065664
author Westphal, Max
Brannath, Werner
author_facet Westphal, Max
Brannath, Werner
author_sort Westphal, Max
collection PubMed
description Model selection and performance assessment for prediction models are important tasks in machine learning, e.g. for the development of medical diagnosis or prognosis rules based on complex data. A common approach is to select the best model via cross-validation and to evaluate this final model on an independent dataset. In this work, we propose to instead evaluate several models simultaneously. These may result from varied hyperparameters or completely different learning algorithms. Our main goal is to increase the probability to correctly identify a model that performs sufficiently well. In this case, adjusting for multiplicity is necessary in the evaluation stage to avoid an inflation of the family wise error rate. We apply the so-called maxT-approach which is based on the joint distribution of test statistics and suitable to (approximately) control the family-wise error rate for a wide variety of performance measures. We conclude that evaluating only a single final model is suboptimal. Instead, several promising models should be evaluated simultaneously, e.g. all models within one standard error of the best validation model. This strategy has proven to increase the probability to correctly identify a good model as well as the final model performance in extensive simulation studies.
format Online
Article
Text
id pubmed-7270727
institution National Center for Biotechnology Information
language English
publishDate 2019
publisher SAGE Publications
record_format MEDLINE/PubMed
spelling pubmed-72707272020-06-23 Evaluation of multiple prediction models: A novel view on model selection and performance assessment Westphal, Max Brannath, Werner Stat Methods Med Res Articles Model selection and performance assessment for prediction models are important tasks in machine learning, e.g. for the development of medical diagnosis or prognosis rules based on complex data. A common approach is to select the best model via cross-validation and to evaluate this final model on an independent dataset. In this work, we propose to instead evaluate several models simultaneously. These may result from varied hyperparameters or completely different learning algorithms. Our main goal is to increase the probability to correctly identify a model that performs sufficiently well. In this case, adjusting for multiplicity is necessary in the evaluation stage to avoid an inflation of the family wise error rate. We apply the so-called maxT-approach which is based on the joint distribution of test statistics and suitable to (approximately) control the family-wise error rate for a wide variety of performance measures. We conclude that evaluating only a single final model is suboptimal. Instead, several promising models should be evaluated simultaneously, e.g. all models within one standard error of the best validation model. This strategy has proven to increase the probability to correctly identify a good model as well as the final model performance in extensive simulation studies. SAGE Publications 2019-09-12 2020-06 /pmc/articles/PMC7270727/ /pubmed/31510862 http://dx.doi.org/10.1177/0962280219854487 Text en © The Author(s) 2019 http://creativecommons.org/licenses/by-nc/4.0/ This article is distributed under the terms of the Creative Commons Attribution-NonCommercial 4.0 License (http://www.creativecommons.org/licenses/by-nc/4.0/) which permits non-commercial use, reproduction and distribution of the work without further permission provided the original work is attributed as specified on the SAGE and Open Access pages (https://us.sagepub.com/en-us/nam/open-access-at-sage).
spellingShingle Articles
Westphal, Max
Brannath, Werner
Evaluation of multiple prediction models: A novel view on model selection and performance assessment
title Evaluation of multiple prediction models: A novel view on model selection and performance assessment
title_full Evaluation of multiple prediction models: A novel view on model selection and performance assessment
title_fullStr Evaluation of multiple prediction models: A novel view on model selection and performance assessment
title_full_unstemmed Evaluation of multiple prediction models: A novel view on model selection and performance assessment
title_short Evaluation of multiple prediction models: A novel view on model selection and performance assessment
title_sort evaluation of multiple prediction models: a novel view on model selection and performance assessment
topic Articles
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7270727/
https://www.ncbi.nlm.nih.gov/pubmed/31510862
http://dx.doi.org/10.1177/0962280219854487
work_keys_str_mv AT westphalmax evaluationofmultiplepredictionmodelsanovelviewonmodelselectionandperformanceassessment
AT brannathwerner evaluationofmultiplepredictionmodelsanovelviewonmodelselectionandperformanceassessment