Cargando…

Evaluation of Prediction-Oriented Model Selection Metrics for Extended Redundancy Analysis

Extended redundancy analysis (ERA) is a statistical method that relates multiple sets of predictors to response variables. In ERA, the conventional approach of model evaluation tends to overestimate the performance of a model since the performance is assessed using the same sample used for model dev...

Descripción completa

Detalles Bibliográficos
Autores principales: Kim, Sunmee, Hwang, Heungsun
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9035639/
https://www.ncbi.nlm.nih.gov/pubmed/35478763
http://dx.doi.org/10.3389/fpsyg.2022.821897
_version_ 1784693339248918528
author Kim, Sunmee
Hwang, Heungsun
author_facet Kim, Sunmee
Hwang, Heungsun
author_sort Kim, Sunmee
collection PubMed
description Extended redundancy analysis (ERA) is a statistical method that relates multiple sets of predictors to response variables. In ERA, the conventional approach of model evaluation tends to overestimate the performance of a model since the performance is assessed using the same sample used for model development. To avoid the overly optimistic assessment, we introduce a new model evaluation approach for ERA, which utilizes computer-intensive resampling methods to assess how well a model performs on unseen data. Specifically, we suggest several new model evaluation metrics for ERA that compute a model’s performance on out-of-sample data, i.e., data not used for model development. Although considerable work has been done in machine learning and statistics to examine the utility of cross-validation and bootstrap variants for assessing such out-of-sample predictive performance, to date, no research has been carried out in the context of ERA. We use simulated and real data examples to compare the proposed model evaluation approach with the conventional one. Results show the conventional approach always favor more complex ERA models, thereby failing to prevent the problem of overfitting in model selection. Conversely, the proposed approach can select the true ERA model among many mis-specified (i.e., underfitted and overfitted) models.
format Online
Article
Text
id pubmed-9035639
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-90356392022-04-26 Evaluation of Prediction-Oriented Model Selection Metrics for Extended Redundancy Analysis Kim, Sunmee Hwang, Heungsun Front Psychol Psychology Extended redundancy analysis (ERA) is a statistical method that relates multiple sets of predictors to response variables. In ERA, the conventional approach of model evaluation tends to overestimate the performance of a model since the performance is assessed using the same sample used for model development. To avoid the overly optimistic assessment, we introduce a new model evaluation approach for ERA, which utilizes computer-intensive resampling methods to assess how well a model performs on unseen data. Specifically, we suggest several new model evaluation metrics for ERA that compute a model’s performance on out-of-sample data, i.e., data not used for model development. Although considerable work has been done in machine learning and statistics to examine the utility of cross-validation and bootstrap variants for assessing such out-of-sample predictive performance, to date, no research has been carried out in the context of ERA. We use simulated and real data examples to compare the proposed model evaluation approach with the conventional one. Results show the conventional approach always favor more complex ERA models, thereby failing to prevent the problem of overfitting in model selection. Conversely, the proposed approach can select the true ERA model among many mis-specified (i.e., underfitted and overfitted) models. Frontiers Media S.A. 2022-04-11 /pmc/articles/PMC9035639/ /pubmed/35478763 http://dx.doi.org/10.3389/fpsyg.2022.821897 Text en Copyright © 2022 Kim and Hwang. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Psychology
Kim, Sunmee
Hwang, Heungsun
Evaluation of Prediction-Oriented Model Selection Metrics for Extended Redundancy Analysis
title Evaluation of Prediction-Oriented Model Selection Metrics for Extended Redundancy Analysis
title_full Evaluation of Prediction-Oriented Model Selection Metrics for Extended Redundancy Analysis
title_fullStr Evaluation of Prediction-Oriented Model Selection Metrics for Extended Redundancy Analysis
title_full_unstemmed Evaluation of Prediction-Oriented Model Selection Metrics for Extended Redundancy Analysis
title_short Evaluation of Prediction-Oriented Model Selection Metrics for Extended Redundancy Analysis
title_sort evaluation of prediction-oriented model selection metrics for extended redundancy analysis
topic Psychology
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9035639/
https://www.ncbi.nlm.nih.gov/pubmed/35478763
http://dx.doi.org/10.3389/fpsyg.2022.821897
work_keys_str_mv AT kimsunmee evaluationofpredictionorientedmodelselectionmetricsforextendedredundancyanalysis
AT hwangheungsun evaluationofpredictionorientedmodelselectionmetricsforextendedredundancyanalysis