Cargando…

The covariance perceptron: A new paradigm for classification and processing of time series in recurrent neuronal networks

Learning in neuronal networks has developed in many directions, in particular to reproduce cognitive tasks like image recognition and speech processing. Implementations have been inspired by stereotypical neuronal responses like tuning curves in the visual system, where, for example, ON/OFF cells fi...

Descripción completa

Detalles Bibliográficos
Autores principales: Gilson, Matthieu, Dahmen, David, Moreno-Bote, Rubén, Insabato, Andrea, Helias, Moritz
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Public Library of Science 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7595646/
https://www.ncbi.nlm.nih.gov/pubmed/33044953
http://dx.doi.org/10.1371/journal.pcbi.1008127
_version_ 1783601922683437056
author Gilson, Matthieu
Dahmen, David
Moreno-Bote, Rubén
Insabato, Andrea
Helias, Moritz
author_facet Gilson, Matthieu
Dahmen, David
Moreno-Bote, Rubén
Insabato, Andrea
Helias, Moritz
author_sort Gilson, Matthieu
collection PubMed
description Learning in neuronal networks has developed in many directions, in particular to reproduce cognitive tasks like image recognition and speech processing. Implementations have been inspired by stereotypical neuronal responses like tuning curves in the visual system, where, for example, ON/OFF cells fire or not depending on the contrast in their receptive fields. Classical models of neuronal networks therefore map a set of input signals to a set of activity levels in the output of the network. Each category of inputs is thereby predominantly characterized by its mean. In the case of time series, fluctuations around this mean constitute noise in this view. For this paradigm, the high variability exhibited by the cortical activity may thus imply limitations or constraints, which have been discussed for many years. For example, the need for averaging neuronal activity over long periods or large groups of cells to assess a robust mean and to diminish the effect of noise correlations. To reconcile robust computations with variable neuronal activity, we here propose a conceptual change of perspective by employing variability of activity as the basis for stimulus-related information to be learned by neurons, rather than merely being the noise that corrupts the mean signal. In this new paradigm both afferent and recurrent weights in a network are tuned to shape the input-output mapping for covariances, the second-order statistics of the fluctuating activity. When including time lags, covariance patterns define a natural metric for time series that capture their propagating nature. We develop the theory for classification of time series based on their spatio-temporal covariances, which reflect dynamical properties. We demonstrate that recurrent connectivity is able to transform information contained in the temporal structure of the signal into spatial covariances. Finally, we use the MNIST database to show how the covariance perceptron can capture specific second-order statistical patterns generated by moving digits.
format Online
Article
Text
id pubmed-7595646
institution National Center for Biotechnology Information
language English
publishDate 2020
publisher Public Library of Science
record_format MEDLINE/PubMed
spelling pubmed-75956462020-11-03 The covariance perceptron: A new paradigm for classification and processing of time series in recurrent neuronal networks Gilson, Matthieu Dahmen, David Moreno-Bote, Rubén Insabato, Andrea Helias, Moritz PLoS Comput Biol Research Article Learning in neuronal networks has developed in many directions, in particular to reproduce cognitive tasks like image recognition and speech processing. Implementations have been inspired by stereotypical neuronal responses like tuning curves in the visual system, where, for example, ON/OFF cells fire or not depending on the contrast in their receptive fields. Classical models of neuronal networks therefore map a set of input signals to a set of activity levels in the output of the network. Each category of inputs is thereby predominantly characterized by its mean. In the case of time series, fluctuations around this mean constitute noise in this view. For this paradigm, the high variability exhibited by the cortical activity may thus imply limitations or constraints, which have been discussed for many years. For example, the need for averaging neuronal activity over long periods or large groups of cells to assess a robust mean and to diminish the effect of noise correlations. To reconcile robust computations with variable neuronal activity, we here propose a conceptual change of perspective by employing variability of activity as the basis for stimulus-related information to be learned by neurons, rather than merely being the noise that corrupts the mean signal. In this new paradigm both afferent and recurrent weights in a network are tuned to shape the input-output mapping for covariances, the second-order statistics of the fluctuating activity. When including time lags, covariance patterns define a natural metric for time series that capture their propagating nature. We develop the theory for classification of time series based on their spatio-temporal covariances, which reflect dynamical properties. We demonstrate that recurrent connectivity is able to transform information contained in the temporal structure of the signal into spatial covariances. Finally, we use the MNIST database to show how the covariance perceptron can capture specific second-order statistical patterns generated by moving digits. Public Library of Science 2020-10-12 /pmc/articles/PMC7595646/ /pubmed/33044953 http://dx.doi.org/10.1371/journal.pcbi.1008127 Text en © 2020 Gilson et al http://creativecommons.org/licenses/by/4.0/ This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
spellingShingle Research Article
Gilson, Matthieu
Dahmen, David
Moreno-Bote, Rubén
Insabato, Andrea
Helias, Moritz
The covariance perceptron: A new paradigm for classification and processing of time series in recurrent neuronal networks
title The covariance perceptron: A new paradigm for classification and processing of time series in recurrent neuronal networks
title_full The covariance perceptron: A new paradigm for classification and processing of time series in recurrent neuronal networks
title_fullStr The covariance perceptron: A new paradigm for classification and processing of time series in recurrent neuronal networks
title_full_unstemmed The covariance perceptron: A new paradigm for classification and processing of time series in recurrent neuronal networks
title_short The covariance perceptron: A new paradigm for classification and processing of time series in recurrent neuronal networks
title_sort covariance perceptron: a new paradigm for classification and processing of time series in recurrent neuronal networks
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7595646/
https://www.ncbi.nlm.nih.gov/pubmed/33044953
http://dx.doi.org/10.1371/journal.pcbi.1008127
work_keys_str_mv AT gilsonmatthieu thecovarianceperceptronanewparadigmforclassificationandprocessingoftimeseriesinrecurrentneuronalnetworks
AT dahmendavid thecovarianceperceptronanewparadigmforclassificationandprocessingoftimeseriesinrecurrentneuronalnetworks
AT morenoboteruben thecovarianceperceptronanewparadigmforclassificationandprocessingoftimeseriesinrecurrentneuronalnetworks
AT insabatoandrea thecovarianceperceptronanewparadigmforclassificationandprocessingoftimeseriesinrecurrentneuronalnetworks
AT heliasmoritz thecovarianceperceptronanewparadigmforclassificationandprocessingoftimeseriesinrecurrentneuronalnetworks
AT gilsonmatthieu covarianceperceptronanewparadigmforclassificationandprocessingoftimeseriesinrecurrentneuronalnetworks
AT dahmendavid covarianceperceptronanewparadigmforclassificationandprocessingoftimeseriesinrecurrentneuronalnetworks
AT morenoboteruben covarianceperceptronanewparadigmforclassificationandprocessingoftimeseriesinrecurrentneuronalnetworks
AT insabatoandrea covarianceperceptronanewparadigmforclassificationandprocessingoftimeseriesinrecurrentneuronalnetworks
AT heliasmoritz covarianceperceptronanewparadigmforclassificationandprocessingoftimeseriesinrecurrentneuronalnetworks