Cargando…
Regularity Normalization: Neuroscience-Inspired Unsupervised Attention across Neural Network Layers †
Inspired by the adaptation phenomenon of neuronal firing, we propose the regularity normalization (RN) as an unsupervised attention mechanism (UAM) which computes the statistical regularity in the implicit space of neural networks under the Minimum Description Length (MDL) principle. Treating the ne...
Autor principal: | |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8774926/ https://www.ncbi.nlm.nih.gov/pubmed/35052085 http://dx.doi.org/10.3390/e24010059 |
_version_ | 1784636460474826752 |
---|---|
author | Lin, Baihan |
author_facet | Lin, Baihan |
author_sort | Lin, Baihan |
collection | PubMed |
description | Inspired by the adaptation phenomenon of neuronal firing, we propose the regularity normalization (RN) as an unsupervised attention mechanism (UAM) which computes the statistical regularity in the implicit space of neural networks under the Minimum Description Length (MDL) principle. Treating the neural network optimization process as a partially observable model selection problem, the regularity normalization constrains the implicit space by a normalization factor, the universal code length. We compute this universal code incrementally across neural network layers and demonstrate the flexibility to include data priors such as top-down attention and other oracle information. Empirically, our approach outperforms existing normalization methods in tackling limited, imbalanced and non-stationary input distribution in image classification, classic control, procedurally-generated reinforcement learning, generative modeling, handwriting generation and question answering tasks with various neural network architectures. Lastly, the unsupervised attention mechanisms is a useful probing tool for neural networks by tracking the dependency and critical learning stages across layers and recurrent time steps of deep networks. |
format | Online Article Text |
id | pubmed-8774926 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-87749262022-01-21 Regularity Normalization: Neuroscience-Inspired Unsupervised Attention across Neural Network Layers † Lin, Baihan Entropy (Basel) Article Inspired by the adaptation phenomenon of neuronal firing, we propose the regularity normalization (RN) as an unsupervised attention mechanism (UAM) which computes the statistical regularity in the implicit space of neural networks under the Minimum Description Length (MDL) principle. Treating the neural network optimization process as a partially observable model selection problem, the regularity normalization constrains the implicit space by a normalization factor, the universal code length. We compute this universal code incrementally across neural network layers and demonstrate the flexibility to include data priors such as top-down attention and other oracle information. Empirically, our approach outperforms existing normalization methods in tackling limited, imbalanced and non-stationary input distribution in image classification, classic control, procedurally-generated reinforcement learning, generative modeling, handwriting generation and question answering tasks with various neural network architectures. Lastly, the unsupervised attention mechanisms is a useful probing tool for neural networks by tracking the dependency and critical learning stages across layers and recurrent time steps of deep networks. MDPI 2021-12-28 /pmc/articles/PMC8774926/ /pubmed/35052085 http://dx.doi.org/10.3390/e24010059 Text en © 2021 by the author. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Lin, Baihan Regularity Normalization: Neuroscience-Inspired Unsupervised Attention across Neural Network Layers † |
title | Regularity Normalization: Neuroscience-Inspired Unsupervised Attention across Neural Network Layers † |
title_full | Regularity Normalization: Neuroscience-Inspired Unsupervised Attention across Neural Network Layers † |
title_fullStr | Regularity Normalization: Neuroscience-Inspired Unsupervised Attention across Neural Network Layers † |
title_full_unstemmed | Regularity Normalization: Neuroscience-Inspired Unsupervised Attention across Neural Network Layers † |
title_short | Regularity Normalization: Neuroscience-Inspired Unsupervised Attention across Neural Network Layers † |
title_sort | regularity normalization: neuroscience-inspired unsupervised attention across neural network layers † |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8774926/ https://www.ncbi.nlm.nih.gov/pubmed/35052085 http://dx.doi.org/10.3390/e24010059 |
work_keys_str_mv | AT linbaihan regularitynormalizationneuroscienceinspiredunsupervisedattentionacrossneuralnetworklayers |