Cargando…

A Scalable Bayesian Sampling Method Based on Stochastic Gradient Descent Isotropization

Stochastic gradient sg-based algorithms for Markov chain Monte Carlo sampling (sgmcmc) tackle large-scale Bayesian modeling problems by operating on mini-batches and injecting noise on sgsteps. The sampling properties of these algorithms are determined by user choices, such as the covariance of the...

Descripción completa

Detalles Bibliográficos
Autores principales: Franzese, Giulio, Milios, Dimitrios, Filippone, Maurizio, Michiardi, Pietro
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8626050/
https://www.ncbi.nlm.nih.gov/pubmed/34828123
http://dx.doi.org/10.3390/e23111426
_version_ 1784606571633836032
author Franzese, Giulio
Milios, Dimitrios
Filippone, Maurizio
Michiardi, Pietro
author_facet Franzese, Giulio
Milios, Dimitrios
Filippone, Maurizio
Michiardi, Pietro
author_sort Franzese, Giulio
collection PubMed
description Stochastic gradient sg-based algorithms for Markov chain Monte Carlo sampling (sgmcmc) tackle large-scale Bayesian modeling problems by operating on mini-batches and injecting noise on sgsteps. The sampling properties of these algorithms are determined by user choices, such as the covariance of the injected noise and the learning rate, and by problem-specific factors, such as assumptions on the loss landscape and the covariance of sg noise. However, current sgmcmc algorithms applied to popular complex models such as Deep Nets cannot simultaneously satisfy the assumptions on loss landscapes and on the behavior of the covariance of the sg noise, while operating with the practical requirement of non-vanishing learning rates. In this work we propose a novel practical method, which makes the sg noise isotropic, using a fixed learning rate that we determine analytically. Extensive experimental validations indicate that our proposal is competitive with the state of the art on sgmcmc.
format Online
Article
Text
id pubmed-8626050
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-86260502021-11-27 A Scalable Bayesian Sampling Method Based on Stochastic Gradient Descent Isotropization Franzese, Giulio Milios, Dimitrios Filippone, Maurizio Michiardi, Pietro Entropy (Basel) Article Stochastic gradient sg-based algorithms for Markov chain Monte Carlo sampling (sgmcmc) tackle large-scale Bayesian modeling problems by operating on mini-batches and injecting noise on sgsteps. The sampling properties of these algorithms are determined by user choices, such as the covariance of the injected noise and the learning rate, and by problem-specific factors, such as assumptions on the loss landscape and the covariance of sg noise. However, current sgmcmc algorithms applied to popular complex models such as Deep Nets cannot simultaneously satisfy the assumptions on loss landscapes and on the behavior of the covariance of the sg noise, while operating with the practical requirement of non-vanishing learning rates. In this work we propose a novel practical method, which makes the sg noise isotropic, using a fixed learning rate that we determine analytically. Extensive experimental validations indicate that our proposal is competitive with the state of the art on sgmcmc. MDPI 2021-10-28 /pmc/articles/PMC8626050/ /pubmed/34828123 http://dx.doi.org/10.3390/e23111426 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Franzese, Giulio
Milios, Dimitrios
Filippone, Maurizio
Michiardi, Pietro
A Scalable Bayesian Sampling Method Based on Stochastic Gradient Descent Isotropization
title A Scalable Bayesian Sampling Method Based on Stochastic Gradient Descent Isotropization
title_full A Scalable Bayesian Sampling Method Based on Stochastic Gradient Descent Isotropization
title_fullStr A Scalable Bayesian Sampling Method Based on Stochastic Gradient Descent Isotropization
title_full_unstemmed A Scalable Bayesian Sampling Method Based on Stochastic Gradient Descent Isotropization
title_short A Scalable Bayesian Sampling Method Based on Stochastic Gradient Descent Isotropization
title_sort scalable bayesian sampling method based on stochastic gradient descent isotropization
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8626050/
https://www.ncbi.nlm.nih.gov/pubmed/34828123
http://dx.doi.org/10.3390/e23111426
work_keys_str_mv AT franzesegiulio ascalablebayesiansamplingmethodbasedonstochasticgradientdescentisotropization
AT miliosdimitrios ascalablebayesiansamplingmethodbasedonstochasticgradientdescentisotropization
AT filipponemaurizio ascalablebayesiansamplingmethodbasedonstochasticgradientdescentisotropization
AT michiardipietro ascalablebayesiansamplingmethodbasedonstochasticgradientdescentisotropization
AT franzesegiulio scalablebayesiansamplingmethodbasedonstochasticgradientdescentisotropization
AT miliosdimitrios scalablebayesiansamplingmethodbasedonstochasticgradientdescentisotropization
AT filipponemaurizio scalablebayesiansamplingmethodbasedonstochasticgradientdescentisotropization
AT michiardipietro scalablebayesiansamplingmethodbasedonstochasticgradientdescentisotropization