Cargando…

Information-Corrected Estimation: A Generalization Error Reducing Parameter Estimation Method

Modern computational models in supervised machine learning are often highly parameterized universal approximators. As such, the value of the parameters is unimportant, and only the out of sample performance is considered. On the other hand much of the literature on model estimation assumes that the...

Descripción completa

Detalles Bibliográficos
Autores principales: Dixon, Matthew, Ward, Tyler
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8621511/
https://www.ncbi.nlm.nih.gov/pubmed/34828117
http://dx.doi.org/10.3390/e23111419
_version_ 1784605475558391808
author Dixon, Matthew
Ward, Tyler
author_facet Dixon, Matthew
Ward, Tyler
author_sort Dixon, Matthew
collection PubMed
description Modern computational models in supervised machine learning are often highly parameterized universal approximators. As such, the value of the parameters is unimportant, and only the out of sample performance is considered. On the other hand much of the literature on model estimation assumes that the parameters themselves have intrinsic value, and thus is concerned with bias and variance of parameter estimates, which may not have any simple relationship to out of sample model performance. Therefore, within supervised machine learning, heavy use is made of ridge regression (i.e., L2 regularization), which requires the the estimation of hyperparameters and can be rendered ineffective by certain model parameterizations. We introduce an objective function which we refer to as Information-Corrected Estimation (ICE) that reduces KL divergence based generalization error for supervised machine learning. ICE attempts to directly maximize a corrected likelihood function as an estimator of the KL divergence. Such an approach is proven, theoretically, to be effective for a wide class of models, with only mild regularity restrictions. Under finite sample sizes, this corrected estimation procedure is shown experimentally to lead to significant reduction in generalization error compared to maximum likelihood estimation and L2 regularization.
format Online
Article
Text
id pubmed-8621511
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-86215112021-11-27 Information-Corrected Estimation: A Generalization Error Reducing Parameter Estimation Method Dixon, Matthew Ward, Tyler Entropy (Basel) Article Modern computational models in supervised machine learning are often highly parameterized universal approximators. As such, the value of the parameters is unimportant, and only the out of sample performance is considered. On the other hand much of the literature on model estimation assumes that the parameters themselves have intrinsic value, and thus is concerned with bias and variance of parameter estimates, which may not have any simple relationship to out of sample model performance. Therefore, within supervised machine learning, heavy use is made of ridge regression (i.e., L2 regularization), which requires the the estimation of hyperparameters and can be rendered ineffective by certain model parameterizations. We introduce an objective function which we refer to as Information-Corrected Estimation (ICE) that reduces KL divergence based generalization error for supervised machine learning. ICE attempts to directly maximize a corrected likelihood function as an estimator of the KL divergence. Such an approach is proven, theoretically, to be effective for a wide class of models, with only mild regularity restrictions. Under finite sample sizes, this corrected estimation procedure is shown experimentally to lead to significant reduction in generalization error compared to maximum likelihood estimation and L2 regularization. MDPI 2021-10-28 /pmc/articles/PMC8621511/ /pubmed/34828117 http://dx.doi.org/10.3390/e23111419 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Dixon, Matthew
Ward, Tyler
Information-Corrected Estimation: A Generalization Error Reducing Parameter Estimation Method
title Information-Corrected Estimation: A Generalization Error Reducing Parameter Estimation Method
title_full Information-Corrected Estimation: A Generalization Error Reducing Parameter Estimation Method
title_fullStr Information-Corrected Estimation: A Generalization Error Reducing Parameter Estimation Method
title_full_unstemmed Information-Corrected Estimation: A Generalization Error Reducing Parameter Estimation Method
title_short Information-Corrected Estimation: A Generalization Error Reducing Parameter Estimation Method
title_sort information-corrected estimation: a generalization error reducing parameter estimation method
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8621511/
https://www.ncbi.nlm.nih.gov/pubmed/34828117
http://dx.doi.org/10.3390/e23111419
work_keys_str_mv AT dixonmatthew informationcorrectedestimationageneralizationerrorreducingparameterestimationmethod
AT wardtyler informationcorrectedestimationageneralizationerrorreducingparameterestimationmethod