Cargando…
Multiscale analysis of slow-fast neuronal learning models with noise
This paper deals with the application of temporal averaging methods to recurrent networks of noisy neurons undergoing a slow and unsupervised modification of their connectivity matrix called learning. Three time-scales arise for these models: (i) the fast neuronal dynamics, (ii) the intermediate ext...
Autores principales: | , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Springer
2012
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3571918/ https://www.ncbi.nlm.nih.gov/pubmed/23174307 http://dx.doi.org/10.1186/2190-8567-2-13 |
_version_ | 1782259234578104320 |
---|---|
author | Galtier, Mathieu Wainrib, Gilles |
author_facet | Galtier, Mathieu Wainrib, Gilles |
author_sort | Galtier, Mathieu |
collection | PubMed |
description | This paper deals with the application of temporal averaging methods to recurrent networks of noisy neurons undergoing a slow and unsupervised modification of their connectivity matrix called learning. Three time-scales arise for these models: (i) the fast neuronal dynamics, (ii) the intermediate external input to the system, and (iii) the slow learning mechanisms. Based on this time-scale separation, we apply an extension of the mathematical theory of stochastic averaging with periodic forcing in order to derive a reduced deterministic model for the connectivity dynamics. We focus on a class of models where the activity is linear to understand the specificity of several learning rules (Hebbian, trace or anti-symmetric learning). In a weakly connected regime, we study the equilibrium connectivity which gathers the entire ‘knowledge’ of the network about the inputs. We develop an asymptotic method to approximate this equilibrium. We show that the symmetric part of the connectivity post-learning encodes the correlation structure of the inputs, whereas the anti-symmetric part corresponds to the cross correlation between the inputs and their time derivative. Moreover, the time-scales ratio appears as an important parameter revealing temporal correlations. |
format | Online Article Text |
id | pubmed-3571918 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2012 |
publisher | Springer |
record_format | MEDLINE/PubMed |
spelling | pubmed-35719182013-02-15 Multiscale analysis of slow-fast neuronal learning models with noise Galtier, Mathieu Wainrib, Gilles J Math Neurosci Research This paper deals with the application of temporal averaging methods to recurrent networks of noisy neurons undergoing a slow and unsupervised modification of their connectivity matrix called learning. Three time-scales arise for these models: (i) the fast neuronal dynamics, (ii) the intermediate external input to the system, and (iii) the slow learning mechanisms. Based on this time-scale separation, we apply an extension of the mathematical theory of stochastic averaging with periodic forcing in order to derive a reduced deterministic model for the connectivity dynamics. We focus on a class of models where the activity is linear to understand the specificity of several learning rules (Hebbian, trace or anti-symmetric learning). In a weakly connected regime, we study the equilibrium connectivity which gathers the entire ‘knowledge’ of the network about the inputs. We develop an asymptotic method to approximate this equilibrium. We show that the symmetric part of the connectivity post-learning encodes the correlation structure of the inputs, whereas the anti-symmetric part corresponds to the cross correlation between the inputs and their time derivative. Moreover, the time-scales ratio appears as an important parameter revealing temporal correlations. Springer 2012-11-22 /pmc/articles/PMC3571918/ /pubmed/23174307 http://dx.doi.org/10.1186/2190-8567-2-13 Text en Copyright ©2012 M. Galtier, G. Wainrib; licensee Springer http://creativecommons.org/licenses/by/2.0 This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. |
spellingShingle | Research Galtier, Mathieu Wainrib, Gilles Multiscale analysis of slow-fast neuronal learning models with noise |
title | Multiscale analysis of slow-fast neuronal learning models with noise |
title_full | Multiscale analysis of slow-fast neuronal learning models with noise |
title_fullStr | Multiscale analysis of slow-fast neuronal learning models with noise |
title_full_unstemmed | Multiscale analysis of slow-fast neuronal learning models with noise |
title_short | Multiscale analysis of slow-fast neuronal learning models with noise |
title_sort | multiscale analysis of slow-fast neuronal learning models with noise |
topic | Research |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3571918/ https://www.ncbi.nlm.nih.gov/pubmed/23174307 http://dx.doi.org/10.1186/2190-8567-2-13 |
work_keys_str_mv | AT galtiermathieu multiscaleanalysisofslowfastneuronallearningmodelswithnoise AT wainribgilles multiscaleanalysisofslowfastneuronallearningmodelswithnoise |