Cargando…

Recurrent Spiking Networks Solve Planning Tasks

A recurrent spiking neural network is proposed that implements planning as probabilistic inference for finite and infinite horizon tasks. The architecture splits this problem into two parts: The stochastic transient firing of the network embodies the dynamics of the planning task. With appropriate i...

Descripción completa

Detalles Bibliográficos
Autores principales: Rueckert, Elmar, Kappel, David, Tanneberg, Daniel, Pecevski, Dejan, Peters, Jan
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Nature Publishing Group 2016
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4758071/
https://www.ncbi.nlm.nih.gov/pubmed/26888174
http://dx.doi.org/10.1038/srep21142
_version_ 1782416556480790528
author Rueckert, Elmar
Kappel, David
Tanneberg, Daniel
Pecevski, Dejan
Peters, Jan
author_facet Rueckert, Elmar
Kappel, David
Tanneberg, Daniel
Pecevski, Dejan
Peters, Jan
author_sort Rueckert, Elmar
collection PubMed
description A recurrent spiking neural network is proposed that implements planning as probabilistic inference for finite and infinite horizon tasks. The architecture splits this problem into two parts: The stochastic transient firing of the network embodies the dynamics of the planning task. With appropriate injected input this dynamics is shaped to generate high-reward state trajectories. A general class of reward-modulated plasticity rules for these afferent synapses is presented. The updates optimize the likelihood of getting a reward through a variant of an Expectation Maximization algorithm and learning is guaranteed to convergence to a local maximum. We find that the network dynamics are qualitatively similar to transient firing patterns during planning and foraging in the hippocampus of awake behaving rats. The model extends classical attractor models and provides a testable prediction on identifying modulating contextual information. In a real robot arm reaching and obstacle avoidance task the ability to represent multiple task solutions is investigated. The neural planning method with its local update rules provides the basis for future neuromorphic hardware implementations with promising potentials like large data processing abilities and early initiation of strategies to avoid dangerous situations in robot co-worker scenarios.
format Online
Article
Text
id pubmed-4758071
institution National Center for Biotechnology Information
language English
publishDate 2016
publisher Nature Publishing Group
record_format MEDLINE/PubMed
spelling pubmed-47580712016-02-26 Recurrent Spiking Networks Solve Planning Tasks Rueckert, Elmar Kappel, David Tanneberg, Daniel Pecevski, Dejan Peters, Jan Sci Rep Article A recurrent spiking neural network is proposed that implements planning as probabilistic inference for finite and infinite horizon tasks. The architecture splits this problem into two parts: The stochastic transient firing of the network embodies the dynamics of the planning task. With appropriate injected input this dynamics is shaped to generate high-reward state trajectories. A general class of reward-modulated plasticity rules for these afferent synapses is presented. The updates optimize the likelihood of getting a reward through a variant of an Expectation Maximization algorithm and learning is guaranteed to convergence to a local maximum. We find that the network dynamics are qualitatively similar to transient firing patterns during planning and foraging in the hippocampus of awake behaving rats. The model extends classical attractor models and provides a testable prediction on identifying modulating contextual information. In a real robot arm reaching and obstacle avoidance task the ability to represent multiple task solutions is investigated. The neural planning method with its local update rules provides the basis for future neuromorphic hardware implementations with promising potentials like large data processing abilities and early initiation of strategies to avoid dangerous situations in robot co-worker scenarios. Nature Publishing Group 2016-02-18 /pmc/articles/PMC4758071/ /pubmed/26888174 http://dx.doi.org/10.1038/srep21142 Text en Copyright © 2016, Macmillan Publishers Limited http://creativecommons.org/licenses/by/4.0/ This work is licensed under a Creative Commons Attribution 4.0 International License. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in the credit line; if the material is not included under the Creative Commons license, users will need to obtain permission from the license holder to reproduce the material. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/
spellingShingle Article
Rueckert, Elmar
Kappel, David
Tanneberg, Daniel
Pecevski, Dejan
Peters, Jan
Recurrent Spiking Networks Solve Planning Tasks
title Recurrent Spiking Networks Solve Planning Tasks
title_full Recurrent Spiking Networks Solve Planning Tasks
title_fullStr Recurrent Spiking Networks Solve Planning Tasks
title_full_unstemmed Recurrent Spiking Networks Solve Planning Tasks
title_short Recurrent Spiking Networks Solve Planning Tasks
title_sort recurrent spiking networks solve planning tasks
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4758071/
https://www.ncbi.nlm.nih.gov/pubmed/26888174
http://dx.doi.org/10.1038/srep21142
work_keys_str_mv AT rueckertelmar recurrentspikingnetworkssolveplanningtasks
AT kappeldavid recurrentspikingnetworkssolveplanningtasks
AT tannebergdaniel recurrentspikingnetworkssolveplanningtasks
AT pecevskidejan recurrentspikingnetworkssolveplanningtasks
AT petersjan recurrentspikingnetworkssolveplanningtasks