Cargando…
Variable Selection and Regularization in Quantile Regression via Minimum Covariance Determinant Based Weights
The importance of variable selection and regularization procedures in multiple regression analysis cannot be overemphasized. These procedures are adversely affected by predictor space data aberrations as well as outliers in the response space. To counter the latter, robust statistical procedures suc...
Autores principales: | , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2020
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7823782/ https://www.ncbi.nlm.nih.gov/pubmed/33383623 http://dx.doi.org/10.3390/e23010033 |
_version_ | 1783639918009909248 |
---|---|
author | Ranganai, Edmore Mudhombo, Innocent |
author_facet | Ranganai, Edmore Mudhombo, Innocent |
author_sort | Ranganai, Edmore |
collection | PubMed |
description | The importance of variable selection and regularization procedures in multiple regression analysis cannot be overemphasized. These procedures are adversely affected by predictor space data aberrations as well as outliers in the response space. To counter the latter, robust statistical procedures such as quantile regression which generalizes the well-known least absolute deviation procedure to all quantile levels have been proposed in the literature. Quantile regression is robust to response variable outliers but very susceptible to outliers in the predictor space (high leverage points) which may alter the eigen-structure of the predictor matrix. High leverage points that alter the eigen-structure of the predictor matrix by creating or hiding collinearity are referred to as collinearity influential points. In this paper, we suggest generalizing the penalized weighted least absolute deviation to all quantile levels, i.e., to penalized weighted quantile regression using the RIDGE, LASSO, and elastic net penalties as a remedy against collinearity influential points and high leverage points in general. To maintain robustness, we make use of very robust weights based on the computationally intensive high breakdown minimum covariance determinant. Simulations and applications to well-known data sets from the literature show an improvement in variable selection and regularization due to the robust weighting formulation. |
format | Online Article Text |
id | pubmed-7823782 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2020 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-78237822021-02-24 Variable Selection and Regularization in Quantile Regression via Minimum Covariance Determinant Based Weights Ranganai, Edmore Mudhombo, Innocent Entropy (Basel) Article The importance of variable selection and regularization procedures in multiple regression analysis cannot be overemphasized. These procedures are adversely affected by predictor space data aberrations as well as outliers in the response space. To counter the latter, robust statistical procedures such as quantile regression which generalizes the well-known least absolute deviation procedure to all quantile levels have been proposed in the literature. Quantile regression is robust to response variable outliers but very susceptible to outliers in the predictor space (high leverage points) which may alter the eigen-structure of the predictor matrix. High leverage points that alter the eigen-structure of the predictor matrix by creating or hiding collinearity are referred to as collinearity influential points. In this paper, we suggest generalizing the penalized weighted least absolute deviation to all quantile levels, i.e., to penalized weighted quantile regression using the RIDGE, LASSO, and elastic net penalties as a remedy against collinearity influential points and high leverage points in general. To maintain robustness, we make use of very robust weights based on the computationally intensive high breakdown minimum covariance determinant. Simulations and applications to well-known data sets from the literature show an improvement in variable selection and regularization due to the robust weighting formulation. MDPI 2020-12-29 /pmc/articles/PMC7823782/ /pubmed/33383623 http://dx.doi.org/10.3390/e23010033 Text en © 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Ranganai, Edmore Mudhombo, Innocent Variable Selection and Regularization in Quantile Regression via Minimum Covariance Determinant Based Weights |
title | Variable Selection and Regularization in Quantile Regression via Minimum Covariance Determinant Based Weights |
title_full | Variable Selection and Regularization in Quantile Regression via Minimum Covariance Determinant Based Weights |
title_fullStr | Variable Selection and Regularization in Quantile Regression via Minimum Covariance Determinant Based Weights |
title_full_unstemmed | Variable Selection and Regularization in Quantile Regression via Minimum Covariance Determinant Based Weights |
title_short | Variable Selection and Regularization in Quantile Regression via Minimum Covariance Determinant Based Weights |
title_sort | variable selection and regularization in quantile regression via minimum covariance determinant based weights |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7823782/ https://www.ncbi.nlm.nih.gov/pubmed/33383623 http://dx.doi.org/10.3390/e23010033 |
work_keys_str_mv | AT ranganaiedmore variableselectionandregularizationinquantileregressionviaminimumcovariancedeterminantbasedweights AT mudhomboinnocent variableselectionandregularizationinquantileregressionviaminimumcovariancedeterminantbasedweights |