Cargando…

Theory of Gating in Recurrent Neural Networks

Recurrent neural networks (RNNs) are powerful dynamical models, widely used in machine learning (ML) and neuroscience. Prior theoretical work has focused on RNNs with additive interactions. However gating i.e., multiplicative interactions are ubiquitous in real neurons and also the central feature o...

Descripción completa

Detalles Bibliográficos
Autores principales: Krishnamurthy, Kamesh, Can, Tankut, Schwab, David J.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9762509/
https://www.ncbi.nlm.nih.gov/pubmed/36545030
http://dx.doi.org/10.1103/physrevx.12.011011
_version_ 1784852877300203520
author Krishnamurthy, Kamesh
Can, Tankut
Schwab, David J.
author_facet Krishnamurthy, Kamesh
Can, Tankut
Schwab, David J.
author_sort Krishnamurthy, Kamesh
collection PubMed
description Recurrent neural networks (RNNs) are powerful dynamical models, widely used in machine learning (ML) and neuroscience. Prior theoretical work has focused on RNNs with additive interactions. However gating i.e., multiplicative interactions are ubiquitous in real neurons and also the central feature of the best-performing RNNs in ML. Here, we show that gating offers flexible control of two salient features of the collective dynamics: (i) timescales and (ii) dimensionality. The gate controlling timescales leads to a novel marginally stable state, where the network functions as a flexible integrator. Unlike previous approaches, gating permits this important function without parameter fine-tuning or special symmetries. Gates also provide a flexible, context-dependent mechanism to reset the memory trace, thus complementing the memory function. The gate modulating the dimensionality can induce a novel, discontinuous chaotic transition, where inputs push a stable system to strong chaotic activity, in contrast to the typically stabilizing effect of inputs. At this transition, unlike additive RNNs, the proliferation of critical points (topological complexity) is decoupled from the appearance of chaotic dynamics (dynamical complexity). The rich dynamics are summarized in phase diagrams, thus providing a map for principled parameter initialization choices to ML practitioners.
format Online
Article
Text
id pubmed-9762509
institution National Center for Biotechnology Information
language English
publishDate 2022
record_format MEDLINE/PubMed
spelling pubmed-97625092023-01-18 Theory of Gating in Recurrent Neural Networks Krishnamurthy, Kamesh Can, Tankut Schwab, David J. Phys Rev X Article Recurrent neural networks (RNNs) are powerful dynamical models, widely used in machine learning (ML) and neuroscience. Prior theoretical work has focused on RNNs with additive interactions. However gating i.e., multiplicative interactions are ubiquitous in real neurons and also the central feature of the best-performing RNNs in ML. Here, we show that gating offers flexible control of two salient features of the collective dynamics: (i) timescales and (ii) dimensionality. The gate controlling timescales leads to a novel marginally stable state, where the network functions as a flexible integrator. Unlike previous approaches, gating permits this important function without parameter fine-tuning or special symmetries. Gates also provide a flexible, context-dependent mechanism to reset the memory trace, thus complementing the memory function. The gate modulating the dimensionality can induce a novel, discontinuous chaotic transition, where inputs push a stable system to strong chaotic activity, in contrast to the typically stabilizing effect of inputs. At this transition, unlike additive RNNs, the proliferation of critical points (topological complexity) is decoupled from the appearance of chaotic dynamics (dynamical complexity). The rich dynamics are summarized in phase diagrams, thus providing a map for principled parameter initialization choices to ML practitioners. 2022 2022-01-18 /pmc/articles/PMC9762509/ /pubmed/36545030 http://dx.doi.org/10.1103/physrevx.12.011011 Text en https://creativecommons.org/licenses/by/4.0/Published by the American Physical Society under the terms of the Creative Commons Attribution 4.0 International (https://creativecommons.org/licenses/by/4.0/) license. Further distribution of this work must maintain attribution to the author(s) and the published article’s title, journal citation, and DOI.
spellingShingle Article
Krishnamurthy, Kamesh
Can, Tankut
Schwab, David J.
Theory of Gating in Recurrent Neural Networks
title Theory of Gating in Recurrent Neural Networks
title_full Theory of Gating in Recurrent Neural Networks
title_fullStr Theory of Gating in Recurrent Neural Networks
title_full_unstemmed Theory of Gating in Recurrent Neural Networks
title_short Theory of Gating in Recurrent Neural Networks
title_sort theory of gating in recurrent neural networks
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9762509/
https://www.ncbi.nlm.nih.gov/pubmed/36545030
http://dx.doi.org/10.1103/physrevx.12.011011
work_keys_str_mv AT krishnamurthykamesh theoryofgatinginrecurrentneuralnetworks
AT cantankut theoryofgatinginrecurrentneuralnetworks
AT schwabdavidj theoryofgatinginrecurrentneuralnetworks