Cargando…
Bayesian interpolation with deep linear networks
Characterizing how neural network depth, width, and dataset size jointly impact model quality is a central problem in deep learning theory. We give here a complete solution in the special case of linear networks with output dimension one trained using zero noise Bayesian inference with Gaussian weig...
Autores principales: | , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
National Academy of Sciences
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10266010/ https://www.ncbi.nlm.nih.gov/pubmed/37252994 http://dx.doi.org/10.1073/pnas.2301345120 |
_version_ | 1785145915499085824 |
---|---|
author | Hanin, Boris Zlokapa, Alexander |
author_facet | Hanin, Boris Zlokapa, Alexander |
author_sort | Hanin, Boris |
collection | PubMed |
description | Characterizing how neural network depth, width, and dataset size jointly impact model quality is a central problem in deep learning theory. We give here a complete solution in the special case of linear networks with output dimension one trained using zero noise Bayesian inference with Gaussian weight priors and mean squared error as a negative log-likelihood. For any training dataset, network depth, and hidden layer widths, we find non-asymptotic expressions for the predictive posterior and Bayesian model evidence in terms of Meijer-G functions, a class of meromorphic special functions of a single complex variable. Through novel asymptotic expansions of these Meijer-G functions, a rich new picture of the joint role of depth, width, and dataset size emerges. We show that linear networks make provably optimal predictions at infinite depth: the posterior of infinitely deep linear networks with data-agnostic priors is the same as that of shallow networks with evidence-maximizing data-dependent priors. This yields a principled reason to prefer deeper networks when priors are forced to be data-agnostic. Moreover, we show that with data-agnostic priors, Bayesian model evidence in wide linear networks is maximized at infinite depth, elucidating the salutary role of increased depth for model selection. Underpinning our results is a novel emergent notion of effective depth, given by the number of hidden layers times the number of data points divided by the network width; this determines the structure of the posterior in the large-data limit. |
format | Online Article Text |
id | pubmed-10266010 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | National Academy of Sciences |
record_format | MEDLINE/PubMed |
spelling | pubmed-102660102023-11-30 Bayesian interpolation with deep linear networks Hanin, Boris Zlokapa, Alexander Proc Natl Acad Sci U S A Physical Sciences Characterizing how neural network depth, width, and dataset size jointly impact model quality is a central problem in deep learning theory. We give here a complete solution in the special case of linear networks with output dimension one trained using zero noise Bayesian inference with Gaussian weight priors and mean squared error as a negative log-likelihood. For any training dataset, network depth, and hidden layer widths, we find non-asymptotic expressions for the predictive posterior and Bayesian model evidence in terms of Meijer-G functions, a class of meromorphic special functions of a single complex variable. Through novel asymptotic expansions of these Meijer-G functions, a rich new picture of the joint role of depth, width, and dataset size emerges. We show that linear networks make provably optimal predictions at infinite depth: the posterior of infinitely deep linear networks with data-agnostic priors is the same as that of shallow networks with evidence-maximizing data-dependent priors. This yields a principled reason to prefer deeper networks when priors are forced to be data-agnostic. Moreover, we show that with data-agnostic priors, Bayesian model evidence in wide linear networks is maximized at infinite depth, elucidating the salutary role of increased depth for model selection. Underpinning our results is a novel emergent notion of effective depth, given by the number of hidden layers times the number of data points divided by the network width; this determines the structure of the posterior in the large-data limit. National Academy of Sciences 2023-05-30 2023-06-06 /pmc/articles/PMC10266010/ /pubmed/37252994 http://dx.doi.org/10.1073/pnas.2301345120 Text en Copyright © 2023 the Author(s). Published by PNAS. https://creativecommons.org/licenses/by-nc-nd/4.0/This article is distributed under Creative Commons Attribution-NonCommercial-NoDerivatives License 4.0 (CC BY-NC-ND) (https://creativecommons.org/licenses/by-nc-nd/4.0/) . |
spellingShingle | Physical Sciences Hanin, Boris Zlokapa, Alexander Bayesian interpolation with deep linear networks |
title | Bayesian interpolation with deep linear networks |
title_full | Bayesian interpolation with deep linear networks |
title_fullStr | Bayesian interpolation with deep linear networks |
title_full_unstemmed | Bayesian interpolation with deep linear networks |
title_short | Bayesian interpolation with deep linear networks |
title_sort | bayesian interpolation with deep linear networks |
topic | Physical Sciences |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10266010/ https://www.ncbi.nlm.nih.gov/pubmed/37252994 http://dx.doi.org/10.1073/pnas.2301345120 |
work_keys_str_mv | AT haninboris bayesianinterpolationwithdeeplinearnetworks AT zlokapaalexander bayesianinterpolationwithdeeplinearnetworks |