Cargando…
Differentiable PAC–Bayes Objectives with Partially Aggregated Neural Networks
We make two related contributions motivated by the challenge of training stochastic neural networks, particularly in a PAC–Bayesian setting: (1) we show how averaging over an ensemble of stochastic neural networks enables a new class of partially-aggregated estimators, proving that these lead to unb...
Autores principales: | , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8535105/ https://www.ncbi.nlm.nih.gov/pubmed/34682004 http://dx.doi.org/10.3390/e23101280 |
_version_ | 1784587698189631488 |
---|---|
author | Biggs, Felix Guedj, Benjamin |
author_facet | Biggs, Felix Guedj, Benjamin |
author_sort | Biggs, Felix |
collection | PubMed |
description | We make two related contributions motivated by the challenge of training stochastic neural networks, particularly in a PAC–Bayesian setting: (1) we show how averaging over an ensemble of stochastic neural networks enables a new class of partially-aggregated estimators, proving that these lead to unbiased lower-variance output and gradient estimators; (2) we reformulate a PAC–Bayesian bound for signed-output networks to derive in combination with the above a directly optimisable, differentiable objective and a generalisation guarantee, without using a surrogate loss or loosening the bound. We show empirically that this leads to competitive generalisation guarantees and compares favourably to other methods for training such networks. Finally, we note that the above leads to a simpler PAC–Bayesian training scheme for sign-activation networks than previous work. |
format | Online Article Text |
id | pubmed-8535105 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-85351052021-10-23 Differentiable PAC–Bayes Objectives with Partially Aggregated Neural Networks Biggs, Felix Guedj, Benjamin Entropy (Basel) Article We make two related contributions motivated by the challenge of training stochastic neural networks, particularly in a PAC–Bayesian setting: (1) we show how averaging over an ensemble of stochastic neural networks enables a new class of partially-aggregated estimators, proving that these lead to unbiased lower-variance output and gradient estimators; (2) we reformulate a PAC–Bayesian bound for signed-output networks to derive in combination with the above a directly optimisable, differentiable objective and a generalisation guarantee, without using a surrogate loss or loosening the bound. We show empirically that this leads to competitive generalisation guarantees and compares favourably to other methods for training such networks. Finally, we note that the above leads to a simpler PAC–Bayesian training scheme for sign-activation networks than previous work. MDPI 2021-09-29 /pmc/articles/PMC8535105/ /pubmed/34682004 http://dx.doi.org/10.3390/e23101280 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Biggs, Felix Guedj, Benjamin Differentiable PAC–Bayes Objectives with Partially Aggregated Neural Networks |
title | Differentiable PAC–Bayes Objectives with Partially Aggregated Neural Networks |
title_full | Differentiable PAC–Bayes Objectives with Partially Aggregated Neural Networks |
title_fullStr | Differentiable PAC–Bayes Objectives with Partially Aggregated Neural Networks |
title_full_unstemmed | Differentiable PAC–Bayes Objectives with Partially Aggregated Neural Networks |
title_short | Differentiable PAC–Bayes Objectives with Partially Aggregated Neural Networks |
title_sort | differentiable pac–bayes objectives with partially aggregated neural networks |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8535105/ https://www.ncbi.nlm.nih.gov/pubmed/34682004 http://dx.doi.org/10.3390/e23101280 |
work_keys_str_mv | AT biggsfelix differentiablepacbayesobjectiveswithpartiallyaggregatedneuralnetworks AT guedjbenjamin differentiablepacbayesobjectiveswithpartiallyaggregatedneuralnetworks |