Cargando…

EXODUS: Stable and efficient training of spiking neural networks

INTRODUCTION: Spiking Neural Networks (SNNs) are gaining significant traction in machine learning tasks where energy-efficiency is of utmost importance. Training such networks using the state-of-the-art back-propagation through time (BPTT) is, however, very time-consuming. Previous work employs an e...

Descripción completa

Detalles Bibliográficos
Autores principales: Bauer, Felix C., Lenz, Gregor, Haghighatshoar, Saeid, Sheik, Sadique
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9945199/
https://www.ncbi.nlm.nih.gov/pubmed/36845419
http://dx.doi.org/10.3389/fnins.2023.1110444
_version_ 1784892086416310272
author Bauer, Felix C.
Lenz, Gregor
Haghighatshoar, Saeid
Sheik, Sadique
author_facet Bauer, Felix C.
Lenz, Gregor
Haghighatshoar, Saeid
Sheik, Sadique
author_sort Bauer, Felix C.
collection PubMed
description INTRODUCTION: Spiking Neural Networks (SNNs) are gaining significant traction in machine learning tasks where energy-efficiency is of utmost importance. Training such networks using the state-of-the-art back-propagation through time (BPTT) is, however, very time-consuming. Previous work employs an efficient GPU-accelerated backpropagation algorithm called SLAYER, which speeds up training considerably. SLAYER, however, does not take into account the neuron reset mechanism while computing the gradients, which we argue to be the source of numerical instability. To counteract this, SLAYER introduces a gradient scale hyper parameter across layers, which needs manual tuning. METHODS: In this paper, we modify SLAYER and design an algorithm called EXODUS, that accounts for the neuron reset mechanism and applies the Implicit Function Theorem (IFT) to calculate the correct gradients (equivalent to those computed by BPTT). We furthermore eliminate the need for ad-hoc scaling of gradients, thus, reducing the training complexity tremendously. RESULTS: We demonstrate, via computer simulations, that EXODUS is numerically stable and achieves comparable or better performance than SLAYER especially in various tasks with SNNs that rely on temporal features.
format Online
Article
Text
id pubmed-9945199
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-99451992023-02-23 EXODUS: Stable and efficient training of spiking neural networks Bauer, Felix C. Lenz, Gregor Haghighatshoar, Saeid Sheik, Sadique Front Neurosci Neuroscience INTRODUCTION: Spiking Neural Networks (SNNs) are gaining significant traction in machine learning tasks where energy-efficiency is of utmost importance. Training such networks using the state-of-the-art back-propagation through time (BPTT) is, however, very time-consuming. Previous work employs an efficient GPU-accelerated backpropagation algorithm called SLAYER, which speeds up training considerably. SLAYER, however, does not take into account the neuron reset mechanism while computing the gradients, which we argue to be the source of numerical instability. To counteract this, SLAYER introduces a gradient scale hyper parameter across layers, which needs manual tuning. METHODS: In this paper, we modify SLAYER and design an algorithm called EXODUS, that accounts for the neuron reset mechanism and applies the Implicit Function Theorem (IFT) to calculate the correct gradients (equivalent to those computed by BPTT). We furthermore eliminate the need for ad-hoc scaling of gradients, thus, reducing the training complexity tremendously. RESULTS: We demonstrate, via computer simulations, that EXODUS is numerically stable and achieves comparable or better performance than SLAYER especially in various tasks with SNNs that rely on temporal features. Frontiers Media S.A. 2023-02-08 /pmc/articles/PMC9945199/ /pubmed/36845419 http://dx.doi.org/10.3389/fnins.2023.1110444 Text en Copyright © 2023 Bauer, Lenz, Haghighatshoar and Sheik. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Neuroscience
Bauer, Felix C.
Lenz, Gregor
Haghighatshoar, Saeid
Sheik, Sadique
EXODUS: Stable and efficient training of spiking neural networks
title EXODUS: Stable and efficient training of spiking neural networks
title_full EXODUS: Stable and efficient training of spiking neural networks
title_fullStr EXODUS: Stable and efficient training of spiking neural networks
title_full_unstemmed EXODUS: Stable and efficient training of spiking neural networks
title_short EXODUS: Stable and efficient training of spiking neural networks
title_sort exodus: stable and efficient training of spiking neural networks
topic Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9945199/
https://www.ncbi.nlm.nih.gov/pubmed/36845419
http://dx.doi.org/10.3389/fnins.2023.1110444
work_keys_str_mv AT bauerfelixc exodusstableandefficienttrainingofspikingneuralnetworks
AT lenzgregor exodusstableandefficienttrainingofspikingneuralnetworks
AT haghighatshoarsaeid exodusstableandefficienttrainingofspikingneuralnetworks
AT sheiksadique exodusstableandefficienttrainingofspikingneuralnetworks