Cargando…

Gradient Regularization as Approximate Variational Inference

We developed Variational Laplace for Bayesian neural networks (BNNs), which exploits a local approximation of the curvature of the likelihood to estimate the ELBO without the need for stochastic sampling of the neural-network weights. The Variational Laplace objective is simple to evaluate, as it is...

Descripción completa

Detalles Bibliográficos
Autores principales: Unlu, Ali, Aitchison, Laurence
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8700595/
https://www.ncbi.nlm.nih.gov/pubmed/34945935
http://dx.doi.org/10.3390/e23121629
_version_ 1784620794979024896
author Unlu, Ali
Aitchison, Laurence
author_facet Unlu, Ali
Aitchison, Laurence
author_sort Unlu, Ali
collection PubMed
description We developed Variational Laplace for Bayesian neural networks (BNNs), which exploits a local approximation of the curvature of the likelihood to estimate the ELBO without the need for stochastic sampling of the neural-network weights. The Variational Laplace objective is simple to evaluate, as it is the log-likelihood plus weight-decay, plus a squared-gradient regularizer. Variational Laplace gave better test performance and expected calibration errors than maximum a posteriori inference and standard sampling-based variational inference, despite using the same variational approximate posterior. Finally, we emphasize the care needed in benchmarking standard VI, as there is a risk of stopping before the variance parameters have converged. We show that early-stopping can be avoided by increasing the learning rate for the variance parameters.
format Online
Article
Text
id pubmed-8700595
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-87005952021-12-24 Gradient Regularization as Approximate Variational Inference Unlu, Ali Aitchison, Laurence Entropy (Basel) Article We developed Variational Laplace for Bayesian neural networks (BNNs), which exploits a local approximation of the curvature of the likelihood to estimate the ELBO without the need for stochastic sampling of the neural-network weights. The Variational Laplace objective is simple to evaluate, as it is the log-likelihood plus weight-decay, plus a squared-gradient regularizer. Variational Laplace gave better test performance and expected calibration errors than maximum a posteriori inference and standard sampling-based variational inference, despite using the same variational approximate posterior. Finally, we emphasize the care needed in benchmarking standard VI, as there is a risk of stopping before the variance parameters have converged. We show that early-stopping can be avoided by increasing the learning rate for the variance parameters. MDPI 2021-12-03 /pmc/articles/PMC8700595/ /pubmed/34945935 http://dx.doi.org/10.3390/e23121629 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Unlu, Ali
Aitchison, Laurence
Gradient Regularization as Approximate Variational Inference
title Gradient Regularization as Approximate Variational Inference
title_full Gradient Regularization as Approximate Variational Inference
title_fullStr Gradient Regularization as Approximate Variational Inference
title_full_unstemmed Gradient Regularization as Approximate Variational Inference
title_short Gradient Regularization as Approximate Variational Inference
title_sort gradient regularization as approximate variational inference
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8700595/
https://www.ncbi.nlm.nih.gov/pubmed/34945935
http://dx.doi.org/10.3390/e23121629
work_keys_str_mv AT unluali gradientregularizationasapproximatevariationalinference
AT aitchisonlaurence gradientregularizationasapproximatevariationalinference