Cargando…
Theory of Gating in Recurrent Neural Networks
Recurrent neural networks (RNNs) are powerful dynamical models, widely used in machine learning (ML) and neuroscience. Prior theoretical work has focused on RNNs with additive interactions. However gating i.e., multiplicative interactions are ubiquitous in real neurons and also the central feature o...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9762509/ https://www.ncbi.nlm.nih.gov/pubmed/36545030 http://dx.doi.org/10.1103/physrevx.12.011011 |
_version_ | 1784852877300203520 |
---|---|
author | Krishnamurthy, Kamesh Can, Tankut Schwab, David J. |
author_facet | Krishnamurthy, Kamesh Can, Tankut Schwab, David J. |
author_sort | Krishnamurthy, Kamesh |
collection | PubMed |
description | Recurrent neural networks (RNNs) are powerful dynamical models, widely used in machine learning (ML) and neuroscience. Prior theoretical work has focused on RNNs with additive interactions. However gating i.e., multiplicative interactions are ubiquitous in real neurons and also the central feature of the best-performing RNNs in ML. Here, we show that gating offers flexible control of two salient features of the collective dynamics: (i) timescales and (ii) dimensionality. The gate controlling timescales leads to a novel marginally stable state, where the network functions as a flexible integrator. Unlike previous approaches, gating permits this important function without parameter fine-tuning or special symmetries. Gates also provide a flexible, context-dependent mechanism to reset the memory trace, thus complementing the memory function. The gate modulating the dimensionality can induce a novel, discontinuous chaotic transition, where inputs push a stable system to strong chaotic activity, in contrast to the typically stabilizing effect of inputs. At this transition, unlike additive RNNs, the proliferation of critical points (topological complexity) is decoupled from the appearance of chaotic dynamics (dynamical complexity). The rich dynamics are summarized in phase diagrams, thus providing a map for principled parameter initialization choices to ML practitioners. |
format | Online Article Text |
id | pubmed-9762509 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
record_format | MEDLINE/PubMed |
spelling | pubmed-97625092023-01-18 Theory of Gating in Recurrent Neural Networks Krishnamurthy, Kamesh Can, Tankut Schwab, David J. Phys Rev X Article Recurrent neural networks (RNNs) are powerful dynamical models, widely used in machine learning (ML) and neuroscience. Prior theoretical work has focused on RNNs with additive interactions. However gating i.e., multiplicative interactions are ubiquitous in real neurons and also the central feature of the best-performing RNNs in ML. Here, we show that gating offers flexible control of two salient features of the collective dynamics: (i) timescales and (ii) dimensionality. The gate controlling timescales leads to a novel marginally stable state, where the network functions as a flexible integrator. Unlike previous approaches, gating permits this important function without parameter fine-tuning or special symmetries. Gates also provide a flexible, context-dependent mechanism to reset the memory trace, thus complementing the memory function. The gate modulating the dimensionality can induce a novel, discontinuous chaotic transition, where inputs push a stable system to strong chaotic activity, in contrast to the typically stabilizing effect of inputs. At this transition, unlike additive RNNs, the proliferation of critical points (topological complexity) is decoupled from the appearance of chaotic dynamics (dynamical complexity). The rich dynamics are summarized in phase diagrams, thus providing a map for principled parameter initialization choices to ML practitioners. 2022 2022-01-18 /pmc/articles/PMC9762509/ /pubmed/36545030 http://dx.doi.org/10.1103/physrevx.12.011011 Text en https://creativecommons.org/licenses/by/4.0/Published by the American Physical Society under the terms of the Creative Commons Attribution 4.0 International (https://creativecommons.org/licenses/by/4.0/) license. Further distribution of this work must maintain attribution to the author(s) and the published article’s title, journal citation, and DOI. |
spellingShingle | Article Krishnamurthy, Kamesh Can, Tankut Schwab, David J. Theory of Gating in Recurrent Neural Networks |
title | Theory of Gating in Recurrent Neural Networks |
title_full | Theory of Gating in Recurrent Neural Networks |
title_fullStr | Theory of Gating in Recurrent Neural Networks |
title_full_unstemmed | Theory of Gating in Recurrent Neural Networks |
title_short | Theory of Gating in Recurrent Neural Networks |
title_sort | theory of gating in recurrent neural networks |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9762509/ https://www.ncbi.nlm.nih.gov/pubmed/36545030 http://dx.doi.org/10.1103/physrevx.12.011011 |
work_keys_str_mv | AT krishnamurthykamesh theoryofgatinginrecurrentneuralnetworks AT cantankut theoryofgatinginrecurrentneuralnetworks AT schwabdavidj theoryofgatinginrecurrentneuralnetworks |