Cargando…

Multiscale analysis of slow-fast neuronal learning models with noise

This paper deals with the application of temporal averaging methods to recurrent networks of noisy neurons undergoing a slow and unsupervised modification of their connectivity matrix called learning. Three time-scales arise for these models: (i) the fast neuronal dynamics, (ii) the intermediate ext...

Descripción completa

Detalles Bibliográficos
Autores principales: Galtier, Mathieu, Wainrib, Gilles
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Springer 2012
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3571918/
https://www.ncbi.nlm.nih.gov/pubmed/23174307
http://dx.doi.org/10.1186/2190-8567-2-13
_version_ 1782259234578104320
author Galtier, Mathieu
Wainrib, Gilles
author_facet Galtier, Mathieu
Wainrib, Gilles
author_sort Galtier, Mathieu
collection PubMed
description This paper deals with the application of temporal averaging methods to recurrent networks of noisy neurons undergoing a slow and unsupervised modification of their connectivity matrix called learning. Three time-scales arise for these models: (i) the fast neuronal dynamics, (ii) the intermediate external input to the system, and (iii) the slow learning mechanisms. Based on this time-scale separation, we apply an extension of the mathematical theory of stochastic averaging with periodic forcing in order to derive a reduced deterministic model for the connectivity dynamics. We focus on a class of models where the activity is linear to understand the specificity of several learning rules (Hebbian, trace or anti-symmetric learning). In a weakly connected regime, we study the equilibrium connectivity which gathers the entire ‘knowledge’ of the network about the inputs. We develop an asymptotic method to approximate this equilibrium. We show that the symmetric part of the connectivity post-learning encodes the correlation structure of the inputs, whereas the anti-symmetric part corresponds to the cross correlation between the inputs and their time derivative. Moreover, the time-scales ratio appears as an important parameter revealing temporal correlations.
format Online
Article
Text
id pubmed-3571918
institution National Center for Biotechnology Information
language English
publishDate 2012
publisher Springer
record_format MEDLINE/PubMed
spelling pubmed-35719182013-02-15 Multiscale analysis of slow-fast neuronal learning models with noise Galtier, Mathieu Wainrib, Gilles J Math Neurosci Research This paper deals with the application of temporal averaging methods to recurrent networks of noisy neurons undergoing a slow and unsupervised modification of their connectivity matrix called learning. Three time-scales arise for these models: (i) the fast neuronal dynamics, (ii) the intermediate external input to the system, and (iii) the slow learning mechanisms. Based on this time-scale separation, we apply an extension of the mathematical theory of stochastic averaging with periodic forcing in order to derive a reduced deterministic model for the connectivity dynamics. We focus on a class of models where the activity is linear to understand the specificity of several learning rules (Hebbian, trace or anti-symmetric learning). In a weakly connected regime, we study the equilibrium connectivity which gathers the entire ‘knowledge’ of the network about the inputs. We develop an asymptotic method to approximate this equilibrium. We show that the symmetric part of the connectivity post-learning encodes the correlation structure of the inputs, whereas the anti-symmetric part corresponds to the cross correlation between the inputs and their time derivative. Moreover, the time-scales ratio appears as an important parameter revealing temporal correlations. Springer 2012-11-22 /pmc/articles/PMC3571918/ /pubmed/23174307 http://dx.doi.org/10.1186/2190-8567-2-13 Text en Copyright ©2012 M. Galtier, G. Wainrib; licensee Springer http://creativecommons.org/licenses/by/2.0 This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
spellingShingle Research
Galtier, Mathieu
Wainrib, Gilles
Multiscale analysis of slow-fast neuronal learning models with noise
title Multiscale analysis of slow-fast neuronal learning models with noise
title_full Multiscale analysis of slow-fast neuronal learning models with noise
title_fullStr Multiscale analysis of slow-fast neuronal learning models with noise
title_full_unstemmed Multiscale analysis of slow-fast neuronal learning models with noise
title_short Multiscale analysis of slow-fast neuronal learning models with noise
title_sort multiscale analysis of slow-fast neuronal learning models with noise
topic Research
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3571918/
https://www.ncbi.nlm.nih.gov/pubmed/23174307
http://dx.doi.org/10.1186/2190-8567-2-13
work_keys_str_mv AT galtiermathieu multiscaleanalysisofslowfastneuronallearningmodelswithnoise
AT wainribgilles multiscaleanalysisofslowfastneuronallearningmodelswithnoise