Cargando…
Evaluation of Prediction-Oriented Model Selection Metrics for Extended Redundancy Analysis
Extended redundancy analysis (ERA) is a statistical method that relates multiple sets of predictors to response variables. In ERA, the conventional approach of model evaluation tends to overestimate the performance of a model since the performance is assessed using the same sample used for model dev...
Autores principales: | , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Frontiers Media S.A.
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9035639/ https://www.ncbi.nlm.nih.gov/pubmed/35478763 http://dx.doi.org/10.3389/fpsyg.2022.821897 |
_version_ | 1784693339248918528 |
---|---|
author | Kim, Sunmee Hwang, Heungsun |
author_facet | Kim, Sunmee Hwang, Heungsun |
author_sort | Kim, Sunmee |
collection | PubMed |
description | Extended redundancy analysis (ERA) is a statistical method that relates multiple sets of predictors to response variables. In ERA, the conventional approach of model evaluation tends to overestimate the performance of a model since the performance is assessed using the same sample used for model development. To avoid the overly optimistic assessment, we introduce a new model evaluation approach for ERA, which utilizes computer-intensive resampling methods to assess how well a model performs on unseen data. Specifically, we suggest several new model evaluation metrics for ERA that compute a model’s performance on out-of-sample data, i.e., data not used for model development. Although considerable work has been done in machine learning and statistics to examine the utility of cross-validation and bootstrap variants for assessing such out-of-sample predictive performance, to date, no research has been carried out in the context of ERA. We use simulated and real data examples to compare the proposed model evaluation approach with the conventional one. Results show the conventional approach always favor more complex ERA models, thereby failing to prevent the problem of overfitting in model selection. Conversely, the proposed approach can select the true ERA model among many mis-specified (i.e., underfitted and overfitted) models. |
format | Online Article Text |
id | pubmed-9035639 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Frontiers Media S.A. |
record_format | MEDLINE/PubMed |
spelling | pubmed-90356392022-04-26 Evaluation of Prediction-Oriented Model Selection Metrics for Extended Redundancy Analysis Kim, Sunmee Hwang, Heungsun Front Psychol Psychology Extended redundancy analysis (ERA) is a statistical method that relates multiple sets of predictors to response variables. In ERA, the conventional approach of model evaluation tends to overestimate the performance of a model since the performance is assessed using the same sample used for model development. To avoid the overly optimistic assessment, we introduce a new model evaluation approach for ERA, which utilizes computer-intensive resampling methods to assess how well a model performs on unseen data. Specifically, we suggest several new model evaluation metrics for ERA that compute a model’s performance on out-of-sample data, i.e., data not used for model development. Although considerable work has been done in machine learning and statistics to examine the utility of cross-validation and bootstrap variants for assessing such out-of-sample predictive performance, to date, no research has been carried out in the context of ERA. We use simulated and real data examples to compare the proposed model evaluation approach with the conventional one. Results show the conventional approach always favor more complex ERA models, thereby failing to prevent the problem of overfitting in model selection. Conversely, the proposed approach can select the true ERA model among many mis-specified (i.e., underfitted and overfitted) models. Frontiers Media S.A. 2022-04-11 /pmc/articles/PMC9035639/ /pubmed/35478763 http://dx.doi.org/10.3389/fpsyg.2022.821897 Text en Copyright © 2022 Kim and Hwang. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. |
spellingShingle | Psychology Kim, Sunmee Hwang, Heungsun Evaluation of Prediction-Oriented Model Selection Metrics for Extended Redundancy Analysis |
title | Evaluation of Prediction-Oriented Model Selection Metrics for Extended Redundancy Analysis |
title_full | Evaluation of Prediction-Oriented Model Selection Metrics for Extended Redundancy Analysis |
title_fullStr | Evaluation of Prediction-Oriented Model Selection Metrics for Extended Redundancy Analysis |
title_full_unstemmed | Evaluation of Prediction-Oriented Model Selection Metrics for Extended Redundancy Analysis |
title_short | Evaluation of Prediction-Oriented Model Selection Metrics for Extended Redundancy Analysis |
title_sort | evaluation of prediction-oriented model selection metrics for extended redundancy analysis |
topic | Psychology |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9035639/ https://www.ncbi.nlm.nih.gov/pubmed/35478763 http://dx.doi.org/10.3389/fpsyg.2022.821897 |
work_keys_str_mv | AT kimsunmee evaluationofpredictionorientedmodelselectionmetricsforextendedredundancyanalysis AT hwangheungsun evaluationofpredictionorientedmodelselectionmetricsforextendedredundancyanalysis |