Cargando…
Effective Plug-Ins for Reducing Inference-Latency of Spiking Convolutional Neural Networks During Inference Phase
Convolutional Neural Networks (CNNs) are effective and mature in the field of classification, while Spiking Neural Networks (SNNs) are energy-saving for their sparsity of data flow and event-driven working mechanism. Previous work demonstrated that CNNs can be converted into equivalent Spiking Convo...
Autores principales: | , , , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Frontiers Media S.A.
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8558256/ https://www.ncbi.nlm.nih.gov/pubmed/34733147 http://dx.doi.org/10.3389/fncom.2021.697469 |
_version_ | 1784592516828364800 |
---|---|
author | Chen, Xuan Yuan, Xiaopeng Fu, Gaoming Luo, Yuanyong Yue, Tao Yan, Feng Wang, Yuxuan Pan, Hongbing |
author_facet | Chen, Xuan Yuan, Xiaopeng Fu, Gaoming Luo, Yuanyong Yue, Tao Yan, Feng Wang, Yuxuan Pan, Hongbing |
author_sort | Chen, Xuan |
collection | PubMed |
description | Convolutional Neural Networks (CNNs) are effective and mature in the field of classification, while Spiking Neural Networks (SNNs) are energy-saving for their sparsity of data flow and event-driven working mechanism. Previous work demonstrated that CNNs can be converted into equivalent Spiking Convolutional Neural Networks (SCNNs) without obvious accuracy loss, including different functional layers such as Convolutional (Conv), Fully Connected (FC), Avg-pooling, Max-pooling, and Batch-Normalization (BN) layers. To reduce inference-latency, existing researches mainly concentrated on the normalization of weights to increase the firing rate of neurons. There are also some approaches during training phase or altering the network architecture. However, little attention has been paid on the end of inference phase. From this new perspective, this paper presents 4 stopping criterions as low-cost plug-ins to reduce the inference-latency of SCNNs. The proposed methods are validated using MATLAB and PyTorch platforms with Spiking-AlexNet for CIFAR-10 dataset and Spiking-LeNet-5 for MNIST dataset. Simulation results reveal that, compared to the state-of-the-art methods, the proposed method can shorten the average inference-latency of Spiking-AlexNet from 892 to 267 time steps (almost 3.34 times faster) with the accuracy decline from 87.95 to 87.72%. With our methods, 4 types of Spiking-LeNet-5 only need 24–70 time steps per image with the accuracy decline not more than 0.1%, while models without our methods require 52–138 time steps, almost 1.92 to 3.21 times slower than us. |
format | Online Article Text |
id | pubmed-8558256 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | Frontiers Media S.A. |
record_format | MEDLINE/PubMed |
spelling | pubmed-85582562021-11-02 Effective Plug-Ins for Reducing Inference-Latency of Spiking Convolutional Neural Networks During Inference Phase Chen, Xuan Yuan, Xiaopeng Fu, Gaoming Luo, Yuanyong Yue, Tao Yan, Feng Wang, Yuxuan Pan, Hongbing Front Comput Neurosci Neuroscience Convolutional Neural Networks (CNNs) are effective and mature in the field of classification, while Spiking Neural Networks (SNNs) are energy-saving for their sparsity of data flow and event-driven working mechanism. Previous work demonstrated that CNNs can be converted into equivalent Spiking Convolutional Neural Networks (SCNNs) without obvious accuracy loss, including different functional layers such as Convolutional (Conv), Fully Connected (FC), Avg-pooling, Max-pooling, and Batch-Normalization (BN) layers. To reduce inference-latency, existing researches mainly concentrated on the normalization of weights to increase the firing rate of neurons. There are also some approaches during training phase or altering the network architecture. However, little attention has been paid on the end of inference phase. From this new perspective, this paper presents 4 stopping criterions as low-cost plug-ins to reduce the inference-latency of SCNNs. The proposed methods are validated using MATLAB and PyTorch platforms with Spiking-AlexNet for CIFAR-10 dataset and Spiking-LeNet-5 for MNIST dataset. Simulation results reveal that, compared to the state-of-the-art methods, the proposed method can shorten the average inference-latency of Spiking-AlexNet from 892 to 267 time steps (almost 3.34 times faster) with the accuracy decline from 87.95 to 87.72%. With our methods, 4 types of Spiking-LeNet-5 only need 24–70 time steps per image with the accuracy decline not more than 0.1%, while models without our methods require 52–138 time steps, almost 1.92 to 3.21 times slower than us. Frontiers Media S.A. 2021-10-18 /pmc/articles/PMC8558256/ /pubmed/34733147 http://dx.doi.org/10.3389/fncom.2021.697469 Text en Copyright © 2021 Chen, Yuan, Fu, Luo, Yue, Yan, Wang and Pan. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. |
spellingShingle | Neuroscience Chen, Xuan Yuan, Xiaopeng Fu, Gaoming Luo, Yuanyong Yue, Tao Yan, Feng Wang, Yuxuan Pan, Hongbing Effective Plug-Ins for Reducing Inference-Latency of Spiking Convolutional Neural Networks During Inference Phase |
title | Effective Plug-Ins for Reducing Inference-Latency of Spiking Convolutional Neural Networks During Inference Phase |
title_full | Effective Plug-Ins for Reducing Inference-Latency of Spiking Convolutional Neural Networks During Inference Phase |
title_fullStr | Effective Plug-Ins for Reducing Inference-Latency of Spiking Convolutional Neural Networks During Inference Phase |
title_full_unstemmed | Effective Plug-Ins for Reducing Inference-Latency of Spiking Convolutional Neural Networks During Inference Phase |
title_short | Effective Plug-Ins for Reducing Inference-Latency of Spiking Convolutional Neural Networks During Inference Phase |
title_sort | effective plug-ins for reducing inference-latency of spiking convolutional neural networks during inference phase |
topic | Neuroscience |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8558256/ https://www.ncbi.nlm.nih.gov/pubmed/34733147 http://dx.doi.org/10.3389/fncom.2021.697469 |
work_keys_str_mv | AT chenxuan effectivepluginsforreducinginferencelatencyofspikingconvolutionalneuralnetworksduringinferencephase AT yuanxiaopeng effectivepluginsforreducinginferencelatencyofspikingconvolutionalneuralnetworksduringinferencephase AT fugaoming effectivepluginsforreducinginferencelatencyofspikingconvolutionalneuralnetworksduringinferencephase AT luoyuanyong effectivepluginsforreducinginferencelatencyofspikingconvolutionalneuralnetworksduringinferencephase AT yuetao effectivepluginsforreducinginferencelatencyofspikingconvolutionalneuralnetworksduringinferencephase AT yanfeng effectivepluginsforreducinginferencelatencyofspikingconvolutionalneuralnetworksduringinferencephase AT wangyuxuan effectivepluginsforreducinginferencelatencyofspikingconvolutionalneuralnetworksduringinferencephase AT panhongbing effectivepluginsforreducinginferencelatencyofspikingconvolutionalneuralnetworksduringinferencephase |