Cargando…

Pre-Synaptic Pool Modification (PSPM): A supervised learning procedure for recurrent spiking neural networks

Learning synaptic weights of spiking neural network (SNN) models that can reproduce target spike trains from provided neural firing data is a central problem in computational neuroscience and spike-based computing. The discovery of the optimal weight values can be posed as a supervised learning task...

Descripción completa

Detalles Bibliográficos
Autores principales: Bagley, Bryce Allen, Bordelon, Blake, Moseley, Benjamin, Wessel, Ralf
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Public Library of Science 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7039446/
https://www.ncbi.nlm.nih.gov/pubmed/32092107
http://dx.doi.org/10.1371/journal.pone.0229083
_version_ 1783500816234053632
author Bagley, Bryce Allen
Bordelon, Blake
Moseley, Benjamin
Wessel, Ralf
author_facet Bagley, Bryce Allen
Bordelon, Blake
Moseley, Benjamin
Wessel, Ralf
author_sort Bagley, Bryce Allen
collection PubMed
description Learning synaptic weights of spiking neural network (SNN) models that can reproduce target spike trains from provided neural firing data is a central problem in computational neuroscience and spike-based computing. The discovery of the optimal weight values can be posed as a supervised learning task wherein the weights of the model network are chosen to maximize the similarity between the target spike trains and the model outputs. It is still largely unknown whether optimizing spike train similarity of highly recurrent SNNs produces weight matrices similar to those of the ground truth model. To this end, we propose flexible heuristic supervised learning rules, termed Pre-Synaptic Pool Modification (PSPM), that rely on stochastic weight updates in order to produce spikes within a short window of the desired times and eliminate spikes outside of this window. PSPM improves spike train similarity for all-to-all SNNs and makes no assumption about the post-synaptic potential of the neurons or the structure of the network since no gradients are required. We test whether optimizing for spike train similarity entails the discovery of accurate weights and explore the relative contributions of local and homeostatic weight updates. Although PSPM improves similarity between spike trains, the learned weights often differ from the weights of the ground truth model, implying that connectome inference from spike data may require additional constraints on connectivity statistics. We also find that spike train similarity is sensitive to local updates, but other measures of network activity such as avalanche distributions, can be learned through synaptic homeostasis.
format Online
Article
Text
id pubmed-7039446
institution National Center for Biotechnology Information
language English
publishDate 2020
publisher Public Library of Science
record_format MEDLINE/PubMed
spelling pubmed-70394462020-03-06 Pre-Synaptic Pool Modification (PSPM): A supervised learning procedure for recurrent spiking neural networks Bagley, Bryce Allen Bordelon, Blake Moseley, Benjamin Wessel, Ralf PLoS One Research Article Learning synaptic weights of spiking neural network (SNN) models that can reproduce target spike trains from provided neural firing data is a central problem in computational neuroscience and spike-based computing. The discovery of the optimal weight values can be posed as a supervised learning task wherein the weights of the model network are chosen to maximize the similarity between the target spike trains and the model outputs. It is still largely unknown whether optimizing spike train similarity of highly recurrent SNNs produces weight matrices similar to those of the ground truth model. To this end, we propose flexible heuristic supervised learning rules, termed Pre-Synaptic Pool Modification (PSPM), that rely on stochastic weight updates in order to produce spikes within a short window of the desired times and eliminate spikes outside of this window. PSPM improves spike train similarity for all-to-all SNNs and makes no assumption about the post-synaptic potential of the neurons or the structure of the network since no gradients are required. We test whether optimizing for spike train similarity entails the discovery of accurate weights and explore the relative contributions of local and homeostatic weight updates. Although PSPM improves similarity between spike trains, the learned weights often differ from the weights of the ground truth model, implying that connectome inference from spike data may require additional constraints on connectivity statistics. We also find that spike train similarity is sensitive to local updates, but other measures of network activity such as avalanche distributions, can be learned through synaptic homeostasis. Public Library of Science 2020-02-24 /pmc/articles/PMC7039446/ /pubmed/32092107 http://dx.doi.org/10.1371/journal.pone.0229083 Text en © 2020 Bagley et al http://creativecommons.org/licenses/by/4.0/ This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
spellingShingle Research Article
Bagley, Bryce Allen
Bordelon, Blake
Moseley, Benjamin
Wessel, Ralf
Pre-Synaptic Pool Modification (PSPM): A supervised learning procedure for recurrent spiking neural networks
title Pre-Synaptic Pool Modification (PSPM): A supervised learning procedure for recurrent spiking neural networks
title_full Pre-Synaptic Pool Modification (PSPM): A supervised learning procedure for recurrent spiking neural networks
title_fullStr Pre-Synaptic Pool Modification (PSPM): A supervised learning procedure for recurrent spiking neural networks
title_full_unstemmed Pre-Synaptic Pool Modification (PSPM): A supervised learning procedure for recurrent spiking neural networks
title_short Pre-Synaptic Pool Modification (PSPM): A supervised learning procedure for recurrent spiking neural networks
title_sort pre-synaptic pool modification (pspm): a supervised learning procedure for recurrent spiking neural networks
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7039446/
https://www.ncbi.nlm.nih.gov/pubmed/32092107
http://dx.doi.org/10.1371/journal.pone.0229083
work_keys_str_mv AT bagleybryceallen presynapticpoolmodificationpspmasupervisedlearningprocedureforrecurrentspikingneuralnetworks
AT bordelonblake presynapticpoolmodificationpspmasupervisedlearningprocedureforrecurrentspikingneuralnetworks
AT moseleybenjamin presynapticpoolmodificationpspmasupervisedlearningprocedureforrecurrentspikingneuralnetworks
AT wesselralf presynapticpoolmodificationpspmasupervisedlearningprocedureforrecurrentspikingneuralnetworks