Cargando…

Weight statistics controls dynamics in recurrent neural networks

Recurrent neural networks are complex non-linear systems, capable of ongoing activity in the absence of driving inputs. The dynamical properties of these systems, in particular their long-time attractor states, are determined on the microscopic level by the connection strengths w(ij) between the ind...

Descripción completa

Detalles Bibliográficos
Autores principales: Krauss, Patrick, Schuster, Marc, Dietrich, Verena, Schilling, Achim, Schulze, Holger, Metzner, Claus
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Public Library of Science 2019
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6456246/
https://www.ncbi.nlm.nih.gov/pubmed/30964879
http://dx.doi.org/10.1371/journal.pone.0214541
_version_ 1783409739904843776
author Krauss, Patrick
Schuster, Marc
Dietrich, Verena
Schilling, Achim
Schulze, Holger
Metzner, Claus
author_facet Krauss, Patrick
Schuster, Marc
Dietrich, Verena
Schilling, Achim
Schulze, Holger
Metzner, Claus
author_sort Krauss, Patrick
collection PubMed
description Recurrent neural networks are complex non-linear systems, capable of ongoing activity in the absence of driving inputs. The dynamical properties of these systems, in particular their long-time attractor states, are determined on the microscopic level by the connection strengths w(ij) between the individual neurons. However, little is known to which extent network dynamics is tunable on a more coarse-grained level by the statistical features of the weight matrix. In this work, we investigate the dynamics of recurrent networks of Boltzmann neurons. In particular we study the impact of three statistical parameters: density (the fraction of non-zero connections), balance (the ratio of excitatory to inhibitory connections), and symmetry (the fraction of neuron pairs with w(ij) = w(ji)). By computing a ‘phase diagram’ of network dynamics, we find that balance is the essential control parameter: Its gradual increase from negative to positive values drives the system from oscillatory behavior into a chaotic regime, and eventually into stationary fixed points. Only directly at the border of the chaotic regime do the neural networks display rich but regular dynamics, thus enabling actual information processing. These results suggest that the brain, too, is fine-tuned to the ‘edge of chaos’ by assuring a proper balance between excitatory and inhibitory neural connections.
format Online
Article
Text
id pubmed-6456246
institution National Center for Biotechnology Information
language English
publishDate 2019
publisher Public Library of Science
record_format MEDLINE/PubMed
spelling pubmed-64562462019-05-03 Weight statistics controls dynamics in recurrent neural networks Krauss, Patrick Schuster, Marc Dietrich, Verena Schilling, Achim Schulze, Holger Metzner, Claus PLoS One Research Article Recurrent neural networks are complex non-linear systems, capable of ongoing activity in the absence of driving inputs. The dynamical properties of these systems, in particular their long-time attractor states, are determined on the microscopic level by the connection strengths w(ij) between the individual neurons. However, little is known to which extent network dynamics is tunable on a more coarse-grained level by the statistical features of the weight matrix. In this work, we investigate the dynamics of recurrent networks of Boltzmann neurons. In particular we study the impact of three statistical parameters: density (the fraction of non-zero connections), balance (the ratio of excitatory to inhibitory connections), and symmetry (the fraction of neuron pairs with w(ij) = w(ji)). By computing a ‘phase diagram’ of network dynamics, we find that balance is the essential control parameter: Its gradual increase from negative to positive values drives the system from oscillatory behavior into a chaotic regime, and eventually into stationary fixed points. Only directly at the border of the chaotic regime do the neural networks display rich but regular dynamics, thus enabling actual information processing. These results suggest that the brain, too, is fine-tuned to the ‘edge of chaos’ by assuring a proper balance between excitatory and inhibitory neural connections. Public Library of Science 2019-04-09 /pmc/articles/PMC6456246/ /pubmed/30964879 http://dx.doi.org/10.1371/journal.pone.0214541 Text en © 2019 Krauss et al http://creativecommons.org/licenses/by/4.0/ This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
spellingShingle Research Article
Krauss, Patrick
Schuster, Marc
Dietrich, Verena
Schilling, Achim
Schulze, Holger
Metzner, Claus
Weight statistics controls dynamics in recurrent neural networks
title Weight statistics controls dynamics in recurrent neural networks
title_full Weight statistics controls dynamics in recurrent neural networks
title_fullStr Weight statistics controls dynamics in recurrent neural networks
title_full_unstemmed Weight statistics controls dynamics in recurrent neural networks
title_short Weight statistics controls dynamics in recurrent neural networks
title_sort weight statistics controls dynamics in recurrent neural networks
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6456246/
https://www.ncbi.nlm.nih.gov/pubmed/30964879
http://dx.doi.org/10.1371/journal.pone.0214541
work_keys_str_mv AT krausspatrick weightstatisticscontrolsdynamicsinrecurrentneuralnetworks
AT schustermarc weightstatisticscontrolsdynamicsinrecurrentneuralnetworks
AT dietrichverena weightstatisticscontrolsdynamicsinrecurrentneuralnetworks
AT schillingachim weightstatisticscontrolsdynamicsinrecurrentneuralnetworks
AT schulzeholger weightstatisticscontrolsdynamicsinrecurrentneuralnetworks
AT metznerclaus weightstatisticscontrolsdynamicsinrecurrentneuralnetworks