Cargando…
Gradient Regularization as Approximate Variational Inference
We developed Variational Laplace for Bayesian neural networks (BNNs), which exploits a local approximation of the curvature of the likelihood to estimate the ELBO without the need for stochastic sampling of the neural-network weights. The Variational Laplace objective is simple to evaluate, as it is...
Autores principales: | , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8700595/ https://www.ncbi.nlm.nih.gov/pubmed/34945935 http://dx.doi.org/10.3390/e23121629 |
_version_ | 1784620794979024896 |
---|---|
author | Unlu, Ali Aitchison, Laurence |
author_facet | Unlu, Ali Aitchison, Laurence |
author_sort | Unlu, Ali |
collection | PubMed |
description | We developed Variational Laplace for Bayesian neural networks (BNNs), which exploits a local approximation of the curvature of the likelihood to estimate the ELBO without the need for stochastic sampling of the neural-network weights. The Variational Laplace objective is simple to evaluate, as it is the log-likelihood plus weight-decay, plus a squared-gradient regularizer. Variational Laplace gave better test performance and expected calibration errors than maximum a posteriori inference and standard sampling-based variational inference, despite using the same variational approximate posterior. Finally, we emphasize the care needed in benchmarking standard VI, as there is a risk of stopping before the variance parameters have converged. We show that early-stopping can be avoided by increasing the learning rate for the variance parameters. |
format | Online Article Text |
id | pubmed-8700595 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-87005952021-12-24 Gradient Regularization as Approximate Variational Inference Unlu, Ali Aitchison, Laurence Entropy (Basel) Article We developed Variational Laplace for Bayesian neural networks (BNNs), which exploits a local approximation of the curvature of the likelihood to estimate the ELBO without the need for stochastic sampling of the neural-network weights. The Variational Laplace objective is simple to evaluate, as it is the log-likelihood plus weight-decay, plus a squared-gradient regularizer. Variational Laplace gave better test performance and expected calibration errors than maximum a posteriori inference and standard sampling-based variational inference, despite using the same variational approximate posterior. Finally, we emphasize the care needed in benchmarking standard VI, as there is a risk of stopping before the variance parameters have converged. We show that early-stopping can be avoided by increasing the learning rate for the variance parameters. MDPI 2021-12-03 /pmc/articles/PMC8700595/ /pubmed/34945935 http://dx.doi.org/10.3390/e23121629 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Unlu, Ali Aitchison, Laurence Gradient Regularization as Approximate Variational Inference |
title | Gradient Regularization as Approximate Variational Inference |
title_full | Gradient Regularization as Approximate Variational Inference |
title_fullStr | Gradient Regularization as Approximate Variational Inference |
title_full_unstemmed | Gradient Regularization as Approximate Variational Inference |
title_short | Gradient Regularization as Approximate Variational Inference |
title_sort | gradient regularization as approximate variational inference |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8700595/ https://www.ncbi.nlm.nih.gov/pubmed/34945935 http://dx.doi.org/10.3390/e23121629 |
work_keys_str_mv | AT unluali gradientregularizationasapproximatevariationalinference AT aitchisonlaurence gradientregularizationasapproximatevariationalinference |