Cargando…

Effective Plug-Ins for Reducing Inference-Latency of Spiking Convolutional Neural Networks During Inference Phase

Convolutional Neural Networks (CNNs) are effective and mature in the field of classification, while Spiking Neural Networks (SNNs) are energy-saving for their sparsity of data flow and event-driven working mechanism. Previous work demonstrated that CNNs can be converted into equivalent Spiking Convo...

Descripción completa

Detalles Bibliográficos
Autores principales: Chen, Xuan, Yuan, Xiaopeng, Fu, Gaoming, Luo, Yuanyong, Yue, Tao, Yan, Feng, Wang, Yuxuan, Pan, Hongbing
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8558256/
https://www.ncbi.nlm.nih.gov/pubmed/34733147
http://dx.doi.org/10.3389/fncom.2021.697469
_version_ 1784592516828364800
author Chen, Xuan
Yuan, Xiaopeng
Fu, Gaoming
Luo, Yuanyong
Yue, Tao
Yan, Feng
Wang, Yuxuan
Pan, Hongbing
author_facet Chen, Xuan
Yuan, Xiaopeng
Fu, Gaoming
Luo, Yuanyong
Yue, Tao
Yan, Feng
Wang, Yuxuan
Pan, Hongbing
author_sort Chen, Xuan
collection PubMed
description Convolutional Neural Networks (CNNs) are effective and mature in the field of classification, while Spiking Neural Networks (SNNs) are energy-saving for their sparsity of data flow and event-driven working mechanism. Previous work demonstrated that CNNs can be converted into equivalent Spiking Convolutional Neural Networks (SCNNs) without obvious accuracy loss, including different functional layers such as Convolutional (Conv), Fully Connected (FC), Avg-pooling, Max-pooling, and Batch-Normalization (BN) layers. To reduce inference-latency, existing researches mainly concentrated on the normalization of weights to increase the firing rate of neurons. There are also some approaches during training phase or altering the network architecture. However, little attention has been paid on the end of inference phase. From this new perspective, this paper presents 4 stopping criterions as low-cost plug-ins to reduce the inference-latency of SCNNs. The proposed methods are validated using MATLAB and PyTorch platforms with Spiking-AlexNet for CIFAR-10 dataset and Spiking-LeNet-5 for MNIST dataset. Simulation results reveal that, compared to the state-of-the-art methods, the proposed method can shorten the average inference-latency of Spiking-AlexNet from 892 to 267 time steps (almost 3.34 times faster) with the accuracy decline from 87.95 to 87.72%. With our methods, 4 types of Spiking-LeNet-5 only need 24–70 time steps per image with the accuracy decline not more than 0.1%, while models without our methods require 52–138 time steps, almost 1.92 to 3.21 times slower than us.
format Online
Article
Text
id pubmed-8558256
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-85582562021-11-02 Effective Plug-Ins for Reducing Inference-Latency of Spiking Convolutional Neural Networks During Inference Phase Chen, Xuan Yuan, Xiaopeng Fu, Gaoming Luo, Yuanyong Yue, Tao Yan, Feng Wang, Yuxuan Pan, Hongbing Front Comput Neurosci Neuroscience Convolutional Neural Networks (CNNs) are effective and mature in the field of classification, while Spiking Neural Networks (SNNs) are energy-saving for their sparsity of data flow and event-driven working mechanism. Previous work demonstrated that CNNs can be converted into equivalent Spiking Convolutional Neural Networks (SCNNs) without obvious accuracy loss, including different functional layers such as Convolutional (Conv), Fully Connected (FC), Avg-pooling, Max-pooling, and Batch-Normalization (BN) layers. To reduce inference-latency, existing researches mainly concentrated on the normalization of weights to increase the firing rate of neurons. There are also some approaches during training phase or altering the network architecture. However, little attention has been paid on the end of inference phase. From this new perspective, this paper presents 4 stopping criterions as low-cost plug-ins to reduce the inference-latency of SCNNs. The proposed methods are validated using MATLAB and PyTorch platforms with Spiking-AlexNet for CIFAR-10 dataset and Spiking-LeNet-5 for MNIST dataset. Simulation results reveal that, compared to the state-of-the-art methods, the proposed method can shorten the average inference-latency of Spiking-AlexNet from 892 to 267 time steps (almost 3.34 times faster) with the accuracy decline from 87.95 to 87.72%. With our methods, 4 types of Spiking-LeNet-5 only need 24–70 time steps per image with the accuracy decline not more than 0.1%, while models without our methods require 52–138 time steps, almost 1.92 to 3.21 times slower than us. Frontiers Media S.A. 2021-10-18 /pmc/articles/PMC8558256/ /pubmed/34733147 http://dx.doi.org/10.3389/fncom.2021.697469 Text en Copyright © 2021 Chen, Yuan, Fu, Luo, Yue, Yan, Wang and Pan. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Neuroscience
Chen, Xuan
Yuan, Xiaopeng
Fu, Gaoming
Luo, Yuanyong
Yue, Tao
Yan, Feng
Wang, Yuxuan
Pan, Hongbing
Effective Plug-Ins for Reducing Inference-Latency of Spiking Convolutional Neural Networks During Inference Phase
title Effective Plug-Ins for Reducing Inference-Latency of Spiking Convolutional Neural Networks During Inference Phase
title_full Effective Plug-Ins for Reducing Inference-Latency of Spiking Convolutional Neural Networks During Inference Phase
title_fullStr Effective Plug-Ins for Reducing Inference-Latency of Spiking Convolutional Neural Networks During Inference Phase
title_full_unstemmed Effective Plug-Ins for Reducing Inference-Latency of Spiking Convolutional Neural Networks During Inference Phase
title_short Effective Plug-Ins for Reducing Inference-Latency of Spiking Convolutional Neural Networks During Inference Phase
title_sort effective plug-ins for reducing inference-latency of spiking convolutional neural networks during inference phase
topic Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8558256/
https://www.ncbi.nlm.nih.gov/pubmed/34733147
http://dx.doi.org/10.3389/fncom.2021.697469
work_keys_str_mv AT chenxuan effectivepluginsforreducinginferencelatencyofspikingconvolutionalneuralnetworksduringinferencephase
AT yuanxiaopeng effectivepluginsforreducinginferencelatencyofspikingconvolutionalneuralnetworksduringinferencephase
AT fugaoming effectivepluginsforreducinginferencelatencyofspikingconvolutionalneuralnetworksduringinferencephase
AT luoyuanyong effectivepluginsforreducinginferencelatencyofspikingconvolutionalneuralnetworksduringinferencephase
AT yuetao effectivepluginsforreducinginferencelatencyofspikingconvolutionalneuralnetworksduringinferencephase
AT yanfeng effectivepluginsforreducinginferencelatencyofspikingconvolutionalneuralnetworksduringinferencephase
AT wangyuxuan effectivepluginsforreducinginferencelatencyofspikingconvolutionalneuralnetworksduringinferencephase
AT panhongbing effectivepluginsforreducinginferencelatencyofspikingconvolutionalneuralnetworksduringinferencephase