Cargando…
Exploring the Connection Between Binary and Spiking Neural Networks
On-chip edge intelligence has necessitated the exploration of algorithmic techniques to reduce the compute requirements of current machine learning frameworks. This work aims to bridge the recent algorithmic progress in training Binary Neural Networks and Spiking Neural Networks—both of which are dr...
Autores principales: | , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Frontiers Media S.A.
2020
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7327094/ https://www.ncbi.nlm.nih.gov/pubmed/32670002 http://dx.doi.org/10.3389/fnins.2020.00535 |
_version_ | 1783552471508975616 |
---|---|
author | Lu, Sen Sengupta, Abhronil |
author_facet | Lu, Sen Sengupta, Abhronil |
author_sort | Lu, Sen |
collection | PubMed |
description | On-chip edge intelligence has necessitated the exploration of algorithmic techniques to reduce the compute requirements of current machine learning frameworks. This work aims to bridge the recent algorithmic progress in training Binary Neural Networks and Spiking Neural Networks—both of which are driven by the same motivation and yet synergies between the two have not been fully explored. We show that training Spiking Neural Networks in the extreme quantization regime results in near full precision accuracies on large-scale datasets like CIFAR-100 and ImageNet. An important implication of this work is that Binary Spiking Neural Networks can be enabled by “In-Memory” hardware accelerators catered for Binary Neural Networks without suffering any accuracy degradation due to binarization. We utilize standard training techniques for non-spiking networks to generate our spiking networks by conversion process and also perform an extensive empirical analysis and explore simple design-time and run-time optimization techniques for reducing inference latency of spiking networks (both for binary and full-precision models) by an order of magnitude over prior work. Our implementation source code and trained models are available at https://github.com/NeuroCompLab-psu/SNN-Conversion. |
format | Online Article Text |
id | pubmed-7327094 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2020 |
publisher | Frontiers Media S.A. |
record_format | MEDLINE/PubMed |
spelling | pubmed-73270942020-07-14 Exploring the Connection Between Binary and Spiking Neural Networks Lu, Sen Sengupta, Abhronil Front Neurosci Neuroscience On-chip edge intelligence has necessitated the exploration of algorithmic techniques to reduce the compute requirements of current machine learning frameworks. This work aims to bridge the recent algorithmic progress in training Binary Neural Networks and Spiking Neural Networks—both of which are driven by the same motivation and yet synergies between the two have not been fully explored. We show that training Spiking Neural Networks in the extreme quantization regime results in near full precision accuracies on large-scale datasets like CIFAR-100 and ImageNet. An important implication of this work is that Binary Spiking Neural Networks can be enabled by “In-Memory” hardware accelerators catered for Binary Neural Networks without suffering any accuracy degradation due to binarization. We utilize standard training techniques for non-spiking networks to generate our spiking networks by conversion process and also perform an extensive empirical analysis and explore simple design-time and run-time optimization techniques for reducing inference latency of spiking networks (both for binary and full-precision models) by an order of magnitude over prior work. Our implementation source code and trained models are available at https://github.com/NeuroCompLab-psu/SNN-Conversion. Frontiers Media S.A. 2020-06-24 /pmc/articles/PMC7327094/ /pubmed/32670002 http://dx.doi.org/10.3389/fnins.2020.00535 Text en Copyright © 2020 Lu and Sengupta. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. |
spellingShingle | Neuroscience Lu, Sen Sengupta, Abhronil Exploring the Connection Between Binary and Spiking Neural Networks |
title | Exploring the Connection Between Binary and Spiking Neural Networks |
title_full | Exploring the Connection Between Binary and Spiking Neural Networks |
title_fullStr | Exploring the Connection Between Binary and Spiking Neural Networks |
title_full_unstemmed | Exploring the Connection Between Binary and Spiking Neural Networks |
title_short | Exploring the Connection Between Binary and Spiking Neural Networks |
title_sort | exploring the connection between binary and spiking neural networks |
topic | Neuroscience |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7327094/ https://www.ncbi.nlm.nih.gov/pubmed/32670002 http://dx.doi.org/10.3389/fnins.2020.00535 |
work_keys_str_mv | AT lusen exploringtheconnectionbetweenbinaryandspikingneuralnetworks AT senguptaabhronil exploringtheconnectionbetweenbinaryandspikingneuralnetworks |