Cargando…

Differentiable PAC–Bayes Objectives with Partially Aggregated Neural Networks

We make two related contributions motivated by the challenge of training stochastic neural networks, particularly in a PAC–Bayesian setting: (1) we show how averaging over an ensemble of stochastic neural networks enables a new class of partially-aggregated estimators, proving that these lead to unb...

Descripción completa

Detalles Bibliográficos
Autores principales: Biggs, Felix, Guedj, Benjamin
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8535105/
https://www.ncbi.nlm.nih.gov/pubmed/34682004
http://dx.doi.org/10.3390/e23101280
_version_ 1784587698189631488
author Biggs, Felix
Guedj, Benjamin
author_facet Biggs, Felix
Guedj, Benjamin
author_sort Biggs, Felix
collection PubMed
description We make two related contributions motivated by the challenge of training stochastic neural networks, particularly in a PAC–Bayesian setting: (1) we show how averaging over an ensemble of stochastic neural networks enables a new class of partially-aggregated estimators, proving that these lead to unbiased lower-variance output and gradient estimators; (2) we reformulate a PAC–Bayesian bound for signed-output networks to derive in combination with the above a directly optimisable, differentiable objective and a generalisation guarantee, without using a surrogate loss or loosening the bound. We show empirically that this leads to competitive generalisation guarantees and compares favourably to other methods for training such networks. Finally, we note that the above leads to a simpler PAC–Bayesian training scheme for sign-activation networks than previous work.
format Online
Article
Text
id pubmed-8535105
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-85351052021-10-23 Differentiable PAC–Bayes Objectives with Partially Aggregated Neural Networks Biggs, Felix Guedj, Benjamin Entropy (Basel) Article We make two related contributions motivated by the challenge of training stochastic neural networks, particularly in a PAC–Bayesian setting: (1) we show how averaging over an ensemble of stochastic neural networks enables a new class of partially-aggregated estimators, proving that these lead to unbiased lower-variance output and gradient estimators; (2) we reformulate a PAC–Bayesian bound for signed-output networks to derive in combination with the above a directly optimisable, differentiable objective and a generalisation guarantee, without using a surrogate loss or loosening the bound. We show empirically that this leads to competitive generalisation guarantees and compares favourably to other methods for training such networks. Finally, we note that the above leads to a simpler PAC–Bayesian training scheme for sign-activation networks than previous work. MDPI 2021-09-29 /pmc/articles/PMC8535105/ /pubmed/34682004 http://dx.doi.org/10.3390/e23101280 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Biggs, Felix
Guedj, Benjamin
Differentiable PAC–Bayes Objectives with Partially Aggregated Neural Networks
title Differentiable PAC–Bayes Objectives with Partially Aggregated Neural Networks
title_full Differentiable PAC–Bayes Objectives with Partially Aggregated Neural Networks
title_fullStr Differentiable PAC–Bayes Objectives with Partially Aggregated Neural Networks
title_full_unstemmed Differentiable PAC–Bayes Objectives with Partially Aggregated Neural Networks
title_short Differentiable PAC–Bayes Objectives with Partially Aggregated Neural Networks
title_sort differentiable pac–bayes objectives with partially aggregated neural networks
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8535105/
https://www.ncbi.nlm.nih.gov/pubmed/34682004
http://dx.doi.org/10.3390/e23101280
work_keys_str_mv AT biggsfelix differentiablepacbayesobjectiveswithpartiallyaggregatedneuralnetworks
AT guedjbenjamin differentiablepacbayesobjectiveswithpartiallyaggregatedneuralnetworks