Cargando…

Surrogate gradients for analog neuromorphic computing

To rapidly process temporal information at a low metabolic cost, biological neurons integrate inputs as an analog sum, but communicate with spikes, binary events in time. Analog neuromorphic hardware uses the same principles to emulate spiking neural networks with exceptional energy efficiency. Howe...

Descripción completa

Detalles Bibliográficos
Autores principales: Cramer, Benjamin, Billaudelle, Sebastian, Kanya, Simeon, Leibfried, Aron, Grübl, Andreas, Karasenko, Vitali, Pehle, Christian, Schreiber, Korbinian, Stradmann, Yannik, Weis, Johannes, Schemmel, Johannes, Zenke, Friedemann
Formato: Online Artículo Texto
Lenguaje:English
Publicado: National Academy of Sciences 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8794842/
https://www.ncbi.nlm.nih.gov/pubmed/35042792
http://dx.doi.org/10.1073/pnas.2109194119
_version_ 1784640913236033536
author Cramer, Benjamin
Billaudelle, Sebastian
Kanya, Simeon
Leibfried, Aron
Grübl, Andreas
Karasenko, Vitali
Pehle, Christian
Schreiber, Korbinian
Stradmann, Yannik
Weis, Johannes
Schemmel, Johannes
Zenke, Friedemann
author_facet Cramer, Benjamin
Billaudelle, Sebastian
Kanya, Simeon
Leibfried, Aron
Grübl, Andreas
Karasenko, Vitali
Pehle, Christian
Schreiber, Korbinian
Stradmann, Yannik
Weis, Johannes
Schemmel, Johannes
Zenke, Friedemann
author_sort Cramer, Benjamin
collection PubMed
description To rapidly process temporal information at a low metabolic cost, biological neurons integrate inputs as an analog sum, but communicate with spikes, binary events in time. Analog neuromorphic hardware uses the same principles to emulate spiking neural networks with exceptional energy efficiency. However, instantiating high-performing spiking networks on such hardware remains a significant challenge due to device mismatch and the lack of efficient training algorithms. Surrogate gradient learning has emerged as a promising training strategy for spiking networks, but its applicability for analog neuromorphic systems has not been demonstrated. Here, we demonstrate surrogate gradient learning on the BrainScaleS-2 analog neuromorphic system using an in-the-loop approach. We show that learning self-corrects for device mismatch, resulting in competitive spiking network performance on both vision and speech benchmarks. Our networks display sparse spiking activity with, on average, less than one spike per hidden neuron and input, perform inference at rates of up to 85,000 frames per second, and consume less than 200 mW. In summary, our work sets several benchmarks for low-energy spiking network processing on analog neuromorphic hardware and paves the way for future on-chip learning algorithms.
format Online
Article
Text
id pubmed-8794842
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher National Academy of Sciences
record_format MEDLINE/PubMed
spelling pubmed-87948422022-02-03 Surrogate gradients for analog neuromorphic computing Cramer, Benjamin Billaudelle, Sebastian Kanya, Simeon Leibfried, Aron Grübl, Andreas Karasenko, Vitali Pehle, Christian Schreiber, Korbinian Stradmann, Yannik Weis, Johannes Schemmel, Johannes Zenke, Friedemann Proc Natl Acad Sci U S A Physical Sciences To rapidly process temporal information at a low metabolic cost, biological neurons integrate inputs as an analog sum, but communicate with spikes, binary events in time. Analog neuromorphic hardware uses the same principles to emulate spiking neural networks with exceptional energy efficiency. However, instantiating high-performing spiking networks on such hardware remains a significant challenge due to device mismatch and the lack of efficient training algorithms. Surrogate gradient learning has emerged as a promising training strategy for spiking networks, but its applicability for analog neuromorphic systems has not been demonstrated. Here, we demonstrate surrogate gradient learning on the BrainScaleS-2 analog neuromorphic system using an in-the-loop approach. We show that learning self-corrects for device mismatch, resulting in competitive spiking network performance on both vision and speech benchmarks. Our networks display sparse spiking activity with, on average, less than one spike per hidden neuron and input, perform inference at rates of up to 85,000 frames per second, and consume less than 200 mW. In summary, our work sets several benchmarks for low-energy spiking network processing on analog neuromorphic hardware and paves the way for future on-chip learning algorithms. National Academy of Sciences 2022-01-14 2022-01-25 /pmc/articles/PMC8794842/ /pubmed/35042792 http://dx.doi.org/10.1073/pnas.2109194119 Text en Copyright © 2022 the Author(s). Published by PNAS. https://creativecommons.org/licenses/by/4.0/This open access article is distributed under Creative Commons Attribution License 4.0 (CC BY) (https://creativecommons.org/licenses/by/4.0/) .
spellingShingle Physical Sciences
Cramer, Benjamin
Billaudelle, Sebastian
Kanya, Simeon
Leibfried, Aron
Grübl, Andreas
Karasenko, Vitali
Pehle, Christian
Schreiber, Korbinian
Stradmann, Yannik
Weis, Johannes
Schemmel, Johannes
Zenke, Friedemann
Surrogate gradients for analog neuromorphic computing
title Surrogate gradients for analog neuromorphic computing
title_full Surrogate gradients for analog neuromorphic computing
title_fullStr Surrogate gradients for analog neuromorphic computing
title_full_unstemmed Surrogate gradients for analog neuromorphic computing
title_short Surrogate gradients for analog neuromorphic computing
title_sort surrogate gradients for analog neuromorphic computing
topic Physical Sciences
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8794842/
https://www.ncbi.nlm.nih.gov/pubmed/35042792
http://dx.doi.org/10.1073/pnas.2109194119
work_keys_str_mv AT cramerbenjamin surrogategradientsforanalogneuromorphiccomputing
AT billaudellesebastian surrogategradientsforanalogneuromorphiccomputing
AT kanyasimeon surrogategradientsforanalogneuromorphiccomputing
AT leibfriedaron surrogategradientsforanalogneuromorphiccomputing
AT grublandreas surrogategradientsforanalogneuromorphiccomputing
AT karasenkovitali surrogategradientsforanalogneuromorphiccomputing
AT pehlechristian surrogategradientsforanalogneuromorphiccomputing
AT schreiberkorbinian surrogategradientsforanalogneuromorphiccomputing
AT stradmannyannik surrogategradientsforanalogneuromorphiccomputing
AT weisjohannes surrogategradientsforanalogneuromorphiccomputing
AT schemmeljohannes surrogategradientsforanalogneuromorphiccomputing
AT zenkefriedemann surrogategradientsforanalogneuromorphiccomputing