Cargando…
Shallow Univariate ReLU Networks as Splines: Initialization, Loss Surface, Hessian, and Gradient Flow Dynamics
Understanding the learning dynamics and inductive bias of neural networks (NNs) is hindered by the opacity of the relationship between NN parameters and the function represented. Partially, this is due to symmetries inherent within the NN parameterization, allowing multiple different parameter setti...
Autores principales: | , , , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Frontiers Media S.A.
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9131019/ https://www.ncbi.nlm.nih.gov/pubmed/35647529 http://dx.doi.org/10.3389/frai.2022.889981 |
_version_ | 1784713097455337472 |
---|---|
author | Sahs, Justin Pyle, Ryan Damaraju, Aneel Caro, Josue Ortega Tavaslioglu, Onur Lu, Andy Anselmi, Fabio Patel, Ankit B. |
author_facet | Sahs, Justin Pyle, Ryan Damaraju, Aneel Caro, Josue Ortega Tavaslioglu, Onur Lu, Andy Anselmi, Fabio Patel, Ankit B. |
author_sort | Sahs, Justin |
collection | PubMed |
description | Understanding the learning dynamics and inductive bias of neural networks (NNs) is hindered by the opacity of the relationship between NN parameters and the function represented. Partially, this is due to symmetries inherent within the NN parameterization, allowing multiple different parameter settings to result in an identical output function, resulting in both an unclear relationship and redundant degrees of freedom. The NN parameterization is invariant under two symmetries: permutation of the neurons and a continuous family of transformations of the scale of weight and bias parameters. We propose taking a quotient with respect to the second symmetry group and reparametrizing ReLU NNs as continuous piecewise linear splines. Using this spline lens, we study learning dynamics in shallow univariate ReLU NNs, finding unexpected insights and explanations for several perplexing phenomena. We develop a surprisingly simple and transparent view of the structure of the loss surface, including its critical and fixed points, Hessian, and Hessian spectrum. We also show that standard weight initializations yield very flat initial functions, and that this flatness, together with overparametrization and the initial weight scale, is responsible for the strength and type of implicit regularization, consistent with previous work. Our implicit regularization results are complementary to recent work, showing that initialization scale critically controls implicit regularization via a kernel-based argument. Overall, removing the weight scale symmetry enables us to prove these results more simply and enables us to prove new results and gain new insights while offering a far more transparent and intuitive picture. Looking forward, our quotiented spline-based approach will extend naturally to the multivariate and deep settings, and alongside the kernel-based view, we believe it will play a foundational role in efforts to understand neural networks. Videos of learning dynamics using a spline-based visualization are available at http://shorturl.at/tFWZ2. |
format | Online Article Text |
id | pubmed-9131019 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Frontiers Media S.A. |
record_format | MEDLINE/PubMed |
spelling | pubmed-91310192022-05-26 Shallow Univariate ReLU Networks as Splines: Initialization, Loss Surface, Hessian, and Gradient Flow Dynamics Sahs, Justin Pyle, Ryan Damaraju, Aneel Caro, Josue Ortega Tavaslioglu, Onur Lu, Andy Anselmi, Fabio Patel, Ankit B. Front Artif Intell Artificial Intelligence Understanding the learning dynamics and inductive bias of neural networks (NNs) is hindered by the opacity of the relationship between NN parameters and the function represented. Partially, this is due to symmetries inherent within the NN parameterization, allowing multiple different parameter settings to result in an identical output function, resulting in both an unclear relationship and redundant degrees of freedom. The NN parameterization is invariant under two symmetries: permutation of the neurons and a continuous family of transformations of the scale of weight and bias parameters. We propose taking a quotient with respect to the second symmetry group and reparametrizing ReLU NNs as continuous piecewise linear splines. Using this spline lens, we study learning dynamics in shallow univariate ReLU NNs, finding unexpected insights and explanations for several perplexing phenomena. We develop a surprisingly simple and transparent view of the structure of the loss surface, including its critical and fixed points, Hessian, and Hessian spectrum. We also show that standard weight initializations yield very flat initial functions, and that this flatness, together with overparametrization and the initial weight scale, is responsible for the strength and type of implicit regularization, consistent with previous work. Our implicit regularization results are complementary to recent work, showing that initialization scale critically controls implicit regularization via a kernel-based argument. Overall, removing the weight scale symmetry enables us to prove these results more simply and enables us to prove new results and gain new insights while offering a far more transparent and intuitive picture. Looking forward, our quotiented spline-based approach will extend naturally to the multivariate and deep settings, and alongside the kernel-based view, we believe it will play a foundational role in efforts to understand neural networks. Videos of learning dynamics using a spline-based visualization are available at http://shorturl.at/tFWZ2. Frontiers Media S.A. 2022-05-11 /pmc/articles/PMC9131019/ /pubmed/35647529 http://dx.doi.org/10.3389/frai.2022.889981 Text en Copyright © 2022 Sahs, Pyle, Damaraju, Caro, Tavaslioglu, Lu, Anselmi and Patel. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. |
spellingShingle | Artificial Intelligence Sahs, Justin Pyle, Ryan Damaraju, Aneel Caro, Josue Ortega Tavaslioglu, Onur Lu, Andy Anselmi, Fabio Patel, Ankit B. Shallow Univariate ReLU Networks as Splines: Initialization, Loss Surface, Hessian, and Gradient Flow Dynamics |
title | Shallow Univariate ReLU Networks as Splines: Initialization, Loss Surface, Hessian, and Gradient Flow Dynamics |
title_full | Shallow Univariate ReLU Networks as Splines: Initialization, Loss Surface, Hessian, and Gradient Flow Dynamics |
title_fullStr | Shallow Univariate ReLU Networks as Splines: Initialization, Loss Surface, Hessian, and Gradient Flow Dynamics |
title_full_unstemmed | Shallow Univariate ReLU Networks as Splines: Initialization, Loss Surface, Hessian, and Gradient Flow Dynamics |
title_short | Shallow Univariate ReLU Networks as Splines: Initialization, Loss Surface, Hessian, and Gradient Flow Dynamics |
title_sort | shallow univariate relu networks as splines: initialization, loss surface, hessian, and gradient flow dynamics |
topic | Artificial Intelligence |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9131019/ https://www.ncbi.nlm.nih.gov/pubmed/35647529 http://dx.doi.org/10.3389/frai.2022.889981 |
work_keys_str_mv | AT sahsjustin shallowunivariaterelunetworksassplinesinitializationlosssurfacehessianandgradientflowdynamics AT pyleryan shallowunivariaterelunetworksassplinesinitializationlosssurfacehessianandgradientflowdynamics AT damarajuaneel shallowunivariaterelunetworksassplinesinitializationlosssurfacehessianandgradientflowdynamics AT carojosueortega shallowunivariaterelunetworksassplinesinitializationlosssurfacehessianandgradientflowdynamics AT tavasliogluonur shallowunivariaterelunetworksassplinesinitializationlosssurfacehessianandgradientflowdynamics AT luandy shallowunivariaterelunetworksassplinesinitializationlosssurfacehessianandgradientflowdynamics AT anselmifabio shallowunivariaterelunetworksassplinesinitializationlosssurfacehessianandgradientflowdynamics AT patelankitb shallowunivariaterelunetworksassplinesinitializationlosssurfacehessianandgradientflowdynamics |