Cargando…

Exploring Optimized Spiking Neural Network Architectures for Classification Tasks on Embedded Platforms

In recent times, the usage of modern neuromorphic hardware for brain-inspired SNNs has grown exponentially. In the context of sparse input data, they are undertaking low power consumption for event-based neuromorphic hardware, specifically in the deeper layers. However, using deep ANNs for training...

Descripción completa

Detalles Bibliográficos
Autores principales: Syed, Tehreem, Kakani, Vijay, Cui, Xuenan, Kim, Hakil
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8125750/
https://www.ncbi.nlm.nih.gov/pubmed/34067080
http://dx.doi.org/10.3390/s21093240
_version_ 1783693592788729856
author Syed, Tehreem
Kakani, Vijay
Cui, Xuenan
Kim, Hakil
author_facet Syed, Tehreem
Kakani, Vijay
Cui, Xuenan
Kim, Hakil
author_sort Syed, Tehreem
collection PubMed
description In recent times, the usage of modern neuromorphic hardware for brain-inspired SNNs has grown exponentially. In the context of sparse input data, they are undertaking low power consumption for event-based neuromorphic hardware, specifically in the deeper layers. However, using deep ANNs for training spiking models is still considered as a tedious task. Until recently, various ANN to SNN conversion methods in the literature have been proposed to train deep SNN models. Nevertheless, these methods require hundreds to thousands of time-steps for training and still cannot attain good SNN performance. This work proposes a customized model (VGG, ResNet) architecture to train deep convolutional spiking neural networks. In this current study, the training is carried out using deep convolutional spiking neural networks with surrogate gradient descent backpropagation in a customized layer architecture similar to deep artificial neural networks. Moreover, this work also proposes fewer time-steps for training SNNs with surrogate gradient descent. During the training with surrogate gradient descent backpropagation, overfitting problems have been encountered. To overcome these problems, this work refines the SNN based dropout technique with surrogate gradient descent. The proposed customized SNN models achieve good classification results on both private and public datasets. In this work, several experiments have been carried out on an embedded platform (NVIDIA JETSON TX2 board), where the deployment of customized SNN models has been extensively conducted. Performance validations have been carried out in terms of processing time and inference accuracy between PC and embedded platforms, showing that the proposed customized models and training techniques are feasible for achieving a better performance on various datasets such as CIFAR-10, MNIST, SVHN, and private KITTI and Korean License plate dataset.
format Online
Article
Text
id pubmed-8125750
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-81257502021-05-17 Exploring Optimized Spiking Neural Network Architectures for Classification Tasks on Embedded Platforms Syed, Tehreem Kakani, Vijay Cui, Xuenan Kim, Hakil Sensors (Basel) Article In recent times, the usage of modern neuromorphic hardware for brain-inspired SNNs has grown exponentially. In the context of sparse input data, they are undertaking low power consumption for event-based neuromorphic hardware, specifically in the deeper layers. However, using deep ANNs for training spiking models is still considered as a tedious task. Until recently, various ANN to SNN conversion methods in the literature have been proposed to train deep SNN models. Nevertheless, these methods require hundreds to thousands of time-steps for training and still cannot attain good SNN performance. This work proposes a customized model (VGG, ResNet) architecture to train deep convolutional spiking neural networks. In this current study, the training is carried out using deep convolutional spiking neural networks with surrogate gradient descent backpropagation in a customized layer architecture similar to deep artificial neural networks. Moreover, this work also proposes fewer time-steps for training SNNs with surrogate gradient descent. During the training with surrogate gradient descent backpropagation, overfitting problems have been encountered. To overcome these problems, this work refines the SNN based dropout technique with surrogate gradient descent. The proposed customized SNN models achieve good classification results on both private and public datasets. In this work, several experiments have been carried out on an embedded platform (NVIDIA JETSON TX2 board), where the deployment of customized SNN models has been extensively conducted. Performance validations have been carried out in terms of processing time and inference accuracy between PC and embedded platforms, showing that the proposed customized models and training techniques are feasible for achieving a better performance on various datasets such as CIFAR-10, MNIST, SVHN, and private KITTI and Korean License plate dataset. MDPI 2021-05-07 /pmc/articles/PMC8125750/ /pubmed/34067080 http://dx.doi.org/10.3390/s21093240 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Syed, Tehreem
Kakani, Vijay
Cui, Xuenan
Kim, Hakil
Exploring Optimized Spiking Neural Network Architectures for Classification Tasks on Embedded Platforms
title Exploring Optimized Spiking Neural Network Architectures for Classification Tasks on Embedded Platforms
title_full Exploring Optimized Spiking Neural Network Architectures for Classification Tasks on Embedded Platforms
title_fullStr Exploring Optimized Spiking Neural Network Architectures for Classification Tasks on Embedded Platforms
title_full_unstemmed Exploring Optimized Spiking Neural Network Architectures for Classification Tasks on Embedded Platforms
title_short Exploring Optimized Spiking Neural Network Architectures for Classification Tasks on Embedded Platforms
title_sort exploring optimized spiking neural network architectures for classification tasks on embedded platforms
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8125750/
https://www.ncbi.nlm.nih.gov/pubmed/34067080
http://dx.doi.org/10.3390/s21093240
work_keys_str_mv AT syedtehreem exploringoptimizedspikingneuralnetworkarchitecturesforclassificationtasksonembeddedplatforms
AT kakanivijay exploringoptimizedspikingneuralnetworkarchitecturesforclassificationtasksonembeddedplatforms
AT cuixuenan exploringoptimizedspikingneuralnetworkarchitecturesforclassificationtasksonembeddedplatforms
AT kimhakil exploringoptimizedspikingneuralnetworkarchitecturesforclassificationtasksonembeddedplatforms