Cargando…

Regularization Methods Based on the L(q)-Likelihood for Linear Models with Heavy-Tailed Errors

We propose regularization methods for linear models based on the [Formula: see text]-likelihood, which is a generalization of the log-likelihood using a power function. Regularization methods are popular for the estimation in the normal linear model. However, heavy-tailed errors are also important i...

Descripción completa

Detalles Bibliográficos
Autor principal: Hirose, Yoshihiro
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7597096/
https://www.ncbi.nlm.nih.gov/pubmed/33286805
http://dx.doi.org/10.3390/e22091036
Descripción
Sumario:We propose regularization methods for linear models based on the [Formula: see text]-likelihood, which is a generalization of the log-likelihood using a power function. Regularization methods are popular for the estimation in the normal linear model. However, heavy-tailed errors are also important in statistics and machine learning. We assume q-normal distributions as the errors in linear models. A q-normal distribution is heavy-tailed, which is defined using a power function, not the exponential function. We find that the proposed methods for linear models with q-normal errors coincide with the ordinary regularization methods that are applied to the normal linear model. The proposed methods can be computed using existing packages because they are penalized least squares methods. We examine the proposed methods using numerical experiments, showing that the methods perform well, even when the error is heavy-tailed. The numerical experiments also illustrate that our methods work well in model selection and generalization, especially when the error is slightly heavy-tailed.