Cargando…

Regularity Normalization: Neuroscience-Inspired Unsupervised Attention across Neural Network Layers †

Inspired by the adaptation phenomenon of neuronal firing, we propose the regularity normalization (RN) as an unsupervised attention mechanism (UAM) which computes the statistical regularity in the implicit space of neural networks under the Minimum Description Length (MDL) principle. Treating the ne...

Descripción completa

Detalles Bibliográficos
Autor principal: Lin, Baihan
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8774926/
https://www.ncbi.nlm.nih.gov/pubmed/35052085
http://dx.doi.org/10.3390/e24010059
_version_ 1784636460474826752
author Lin, Baihan
author_facet Lin, Baihan
author_sort Lin, Baihan
collection PubMed
description Inspired by the adaptation phenomenon of neuronal firing, we propose the regularity normalization (RN) as an unsupervised attention mechanism (UAM) which computes the statistical regularity in the implicit space of neural networks under the Minimum Description Length (MDL) principle. Treating the neural network optimization process as a partially observable model selection problem, the regularity normalization constrains the implicit space by a normalization factor, the universal code length. We compute this universal code incrementally across neural network layers and demonstrate the flexibility to include data priors such as top-down attention and other oracle information. Empirically, our approach outperforms existing normalization methods in tackling limited, imbalanced and non-stationary input distribution in image classification, classic control, procedurally-generated reinforcement learning, generative modeling, handwriting generation and question answering tasks with various neural network architectures. Lastly, the unsupervised attention mechanisms is a useful probing tool for neural networks by tracking the dependency and critical learning stages across layers and recurrent time steps of deep networks.
format Online
Article
Text
id pubmed-8774926
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-87749262022-01-21 Regularity Normalization: Neuroscience-Inspired Unsupervised Attention across Neural Network Layers † Lin, Baihan Entropy (Basel) Article Inspired by the adaptation phenomenon of neuronal firing, we propose the regularity normalization (RN) as an unsupervised attention mechanism (UAM) which computes the statistical regularity in the implicit space of neural networks under the Minimum Description Length (MDL) principle. Treating the neural network optimization process as a partially observable model selection problem, the regularity normalization constrains the implicit space by a normalization factor, the universal code length. We compute this universal code incrementally across neural network layers and demonstrate the flexibility to include data priors such as top-down attention and other oracle information. Empirically, our approach outperforms existing normalization methods in tackling limited, imbalanced and non-stationary input distribution in image classification, classic control, procedurally-generated reinforcement learning, generative modeling, handwriting generation and question answering tasks with various neural network architectures. Lastly, the unsupervised attention mechanisms is a useful probing tool for neural networks by tracking the dependency and critical learning stages across layers and recurrent time steps of deep networks. MDPI 2021-12-28 /pmc/articles/PMC8774926/ /pubmed/35052085 http://dx.doi.org/10.3390/e24010059 Text en © 2021 by the author. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Lin, Baihan
Regularity Normalization: Neuroscience-Inspired Unsupervised Attention across Neural Network Layers †
title Regularity Normalization: Neuroscience-Inspired Unsupervised Attention across Neural Network Layers †
title_full Regularity Normalization: Neuroscience-Inspired Unsupervised Attention across Neural Network Layers †
title_fullStr Regularity Normalization: Neuroscience-Inspired Unsupervised Attention across Neural Network Layers †
title_full_unstemmed Regularity Normalization: Neuroscience-Inspired Unsupervised Attention across Neural Network Layers †
title_short Regularity Normalization: Neuroscience-Inspired Unsupervised Attention across Neural Network Layers †
title_sort regularity normalization: neuroscience-inspired unsupervised attention across neural network layers †
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8774926/
https://www.ncbi.nlm.nih.gov/pubmed/35052085
http://dx.doi.org/10.3390/e24010059
work_keys_str_mv AT linbaihan regularitynormalizationneuroscienceinspiredunsupervisedattentionacrossneuralnetworklayers