Cargando…

Inhomogeneous Sparseness Leads to Dynamic Instability During Sequence Memory Recall in a Recurrent Neural Network Model

Theoretical models of associative memory generally assume most of their parameters to be homogeneous across the network. Conversely, biological neural networks exhibit high variability of structural as well as activity parameters. In this paper, we extend the classical clipped learning rule by Wills...

Descripción completa

Detalles Bibliográficos
Autores principales: Medina, Daniel, Leibold, Christian
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Springer 2013
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3844438/
https://www.ncbi.nlm.nih.gov/pubmed/23876197
http://dx.doi.org/10.1186/2190-8567-3-8
_version_ 1782293182497685504
author Medina, Daniel
Leibold, Christian
author_facet Medina, Daniel
Leibold, Christian
author_sort Medina, Daniel
collection PubMed
description Theoretical models of associative memory generally assume most of their parameters to be homogeneous across the network. Conversely, biological neural networks exhibit high variability of structural as well as activity parameters. In this paper, we extend the classical clipped learning rule by Willshaw to networks with inhomogeneous sparseness, i.e., the number of active neurons may vary across memory items. We evaluate this learning rule for sequence memory networks with instantaneous feedback inhibition and show that little surprisingly, memory capacity degrades with increased variability in sparseness. The loss of capacity, however, is very small for short sequences of less than about 10 associations. Most interestingly, we further show that, due to feedback inhibition, too large patterns are much less detrimental for memory capacity than too small patterns.
format Online
Article
Text
id pubmed-3844438
institution National Center for Biotechnology Information
language English
publishDate 2013
publisher Springer
record_format MEDLINE/PubMed
spelling pubmed-38444382013-12-03 Inhomogeneous Sparseness Leads to Dynamic Instability During Sequence Memory Recall in a Recurrent Neural Network Model Medina, Daniel Leibold, Christian J Math Neurosci Research Theoretical models of associative memory generally assume most of their parameters to be homogeneous across the network. Conversely, biological neural networks exhibit high variability of structural as well as activity parameters. In this paper, we extend the classical clipped learning rule by Willshaw to networks with inhomogeneous sparseness, i.e., the number of active neurons may vary across memory items. We evaluate this learning rule for sequence memory networks with instantaneous feedback inhibition and show that little surprisingly, memory capacity degrades with increased variability in sparseness. The loss of capacity, however, is very small for short sequences of less than about 10 associations. Most interestingly, we further show that, due to feedback inhibition, too large patterns are much less detrimental for memory capacity than too small patterns. Springer 2013-07-22 /pmc/articles/PMC3844438/ /pubmed/23876197 http://dx.doi.org/10.1186/2190-8567-3-8 Text en Copyright © 2013 D. Medina, C. Leibold; licensee Springer http://creativecommons.org/licenses/by/2.0 This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
spellingShingle Research
Medina, Daniel
Leibold, Christian
Inhomogeneous Sparseness Leads to Dynamic Instability During Sequence Memory Recall in a Recurrent Neural Network Model
title Inhomogeneous Sparseness Leads to Dynamic Instability During Sequence Memory Recall in a Recurrent Neural Network Model
title_full Inhomogeneous Sparseness Leads to Dynamic Instability During Sequence Memory Recall in a Recurrent Neural Network Model
title_fullStr Inhomogeneous Sparseness Leads to Dynamic Instability During Sequence Memory Recall in a Recurrent Neural Network Model
title_full_unstemmed Inhomogeneous Sparseness Leads to Dynamic Instability During Sequence Memory Recall in a Recurrent Neural Network Model
title_short Inhomogeneous Sparseness Leads to Dynamic Instability During Sequence Memory Recall in a Recurrent Neural Network Model
title_sort inhomogeneous sparseness leads to dynamic instability during sequence memory recall in a recurrent neural network model
topic Research
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3844438/
https://www.ncbi.nlm.nih.gov/pubmed/23876197
http://dx.doi.org/10.1186/2190-8567-3-8
work_keys_str_mv AT medinadaniel inhomogeneoussparsenessleadstodynamicinstabilityduringsequencememoryrecallinarecurrentneuralnetworkmodel
AT leiboldchristian inhomogeneoussparsenessleadstodynamicinstabilityduringsequencememoryrecallinarecurrentneuralnetworkmodel