Cargando…

A Correspondence Between Normalization Strategies in Artificial and Biological Neural Networks

A fundamental challenge at the interface of machine learning and neuroscience is to uncover computational principles that are shared between artificial and biological neural networks. In deep learning, normalization methods such as batch normalization, weight normalization, and their many variants h...

Descripción completa

Detalles Bibliográficos
Autores principales: Shen, Yang, Wang, Julia, Navlakha, Saket
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MIT Press 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8662716/
https://www.ncbi.nlm.nih.gov/pubmed/34474484
http://dx.doi.org/10.1162/neco_a_01439
_version_ 1784613497183666176
author Shen, Yang
Wang, Julia
Navlakha, Saket
author_facet Shen, Yang
Wang, Julia
Navlakha, Saket
author_sort Shen, Yang
collection PubMed
description A fundamental challenge at the interface of machine learning and neuroscience is to uncover computational principles that are shared between artificial and biological neural networks. In deep learning, normalization methods such as batch normalization, weight normalization, and their many variants help to stabilize hidden unit activity and accelerate network training, and these methods have been called one of the most important recent innovations for optimizing deep networks. In the brain, homeostatic plasticity represents a set of mechanisms that also stabilize and normalize network activity to lie within certain ranges, and these mechanisms are critical for maintaining normal brain function. In this article, we discuss parallels between artificial and biological normalization methods at four spatial scales: normalization of a single neuron's activity, normalization of synaptic weights of a neuron, normalization of a layer of neurons, and normalization of a network of neurons. We argue that both types of methods are functionally equivalent—that is, both push activation patterns of hidden units toward a homeostatic state, where all neurons are equally used—and we argue that such representations can improve coding capacity, discrimination, and regularization. As a proof of concept, we develop an algorithm, inspired by a neural normalization technique called synaptic scaling, and show that this algorithm performs competitively against existing normalization methods on several data sets. Overall, we hope this bidirectional connection will inspire neuroscientists and machine learners in three ways: to uncover new normalization algorithms based on established neurobiological principles; to help quantify the trade-offs of different homeostatic plasticity mechanisms used in the brain; and to offer insights about how stability may not hinder, but may actually promote, plasticity.
format Online
Article
Text
id pubmed-8662716
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MIT Press
record_format MEDLINE/PubMed
spelling pubmed-86627162022-02-12 A Correspondence Between Normalization Strategies in Artificial and Biological Neural Networks Shen, Yang Wang, Julia Navlakha, Saket Neural Comput Research Article A fundamental challenge at the interface of machine learning and neuroscience is to uncover computational principles that are shared between artificial and biological neural networks. In deep learning, normalization methods such as batch normalization, weight normalization, and their many variants help to stabilize hidden unit activity and accelerate network training, and these methods have been called one of the most important recent innovations for optimizing deep networks. In the brain, homeostatic plasticity represents a set of mechanisms that also stabilize and normalize network activity to lie within certain ranges, and these mechanisms are critical for maintaining normal brain function. In this article, we discuss parallels between artificial and biological normalization methods at four spatial scales: normalization of a single neuron's activity, normalization of synaptic weights of a neuron, normalization of a layer of neurons, and normalization of a network of neurons. We argue that both types of methods are functionally equivalent—that is, both push activation patterns of hidden units toward a homeostatic state, where all neurons are equally used—and we argue that such representations can improve coding capacity, discrimination, and regularization. As a proof of concept, we develop an algorithm, inspired by a neural normalization technique called synaptic scaling, and show that this algorithm performs competitively against existing normalization methods on several data sets. Overall, we hope this bidirectional connection will inspire neuroscientists and machine learners in three ways: to uncover new normalization algorithms based on established neurobiological principles; to help quantify the trade-offs of different homeostatic plasticity mechanisms used in the brain; and to offer insights about how stability may not hinder, but may actually promote, plasticity. MIT Press 2021-11-12 /pmc/articles/PMC8662716/ /pubmed/34474484 http://dx.doi.org/10.1162/neco_a_01439 Text en © 2021 Massachusetts Institute of Technology https://creativecommons.org/licenses/by-nc/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution-NonCommercial 4.0 International (CC BY-NC 4.0) license, which permits copying and redistributing the material in any medium or format for noncommercial purposes only. For a full description of the license, please visit https://creativecommons.org/licenses/by-nc/4.0 (https://creativecommons.org/licenses/by-nc/4.0/) .
spellingShingle Research Article
Shen, Yang
Wang, Julia
Navlakha, Saket
A Correspondence Between Normalization Strategies in Artificial and Biological Neural Networks
title A Correspondence Between Normalization Strategies in Artificial and Biological Neural Networks
title_full A Correspondence Between Normalization Strategies in Artificial and Biological Neural Networks
title_fullStr A Correspondence Between Normalization Strategies in Artificial and Biological Neural Networks
title_full_unstemmed A Correspondence Between Normalization Strategies in Artificial and Biological Neural Networks
title_short A Correspondence Between Normalization Strategies in Artificial and Biological Neural Networks
title_sort correspondence between normalization strategies in artificial and biological neural networks
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8662716/
https://www.ncbi.nlm.nih.gov/pubmed/34474484
http://dx.doi.org/10.1162/neco_a_01439
work_keys_str_mv AT shenyang acorrespondencebetweennormalizationstrategiesinartificialandbiologicalneuralnetworks
AT wangjulia acorrespondencebetweennormalizationstrategiesinartificialandbiologicalneuralnetworks
AT navlakhasaket acorrespondencebetweennormalizationstrategiesinartificialandbiologicalneuralnetworks
AT shenyang correspondencebetweennormalizationstrategiesinartificialandbiologicalneuralnetworks
AT wangjulia correspondencebetweennormalizationstrategiesinartificialandbiologicalneuralnetworks
AT navlakhasaket correspondencebetweennormalizationstrategiesinartificialandbiologicalneuralnetworks