Cargando…

Unsupervised Adaptive Weight Pruning for Energy-Efficient Neuromorphic Systems

To tackle real-world challenges, deep and complex neural networks are generally used with a massive number of parameters, which require large memory size, extensive computational operations, and high energy consumption in neuromorphic hardware systems. In this work, we propose an unsupervised online...

Descripción completa

Detalles Bibliográficos
Autores principales: Guo, Wenzhe, Fouda, Mohammed E., Yantir, Hasan Erdem, Eltawil, Ahmed M., Salama, Khaled Nabil
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7689062/
https://www.ncbi.nlm.nih.gov/pubmed/33281549
http://dx.doi.org/10.3389/fnins.2020.598876
_version_ 1783613782948315136
author Guo, Wenzhe
Fouda, Mohammed E.
Yantir, Hasan Erdem
Eltawil, Ahmed M.
Salama, Khaled Nabil
author_facet Guo, Wenzhe
Fouda, Mohammed E.
Yantir, Hasan Erdem
Eltawil, Ahmed M.
Salama, Khaled Nabil
author_sort Guo, Wenzhe
collection PubMed
description To tackle real-world challenges, deep and complex neural networks are generally used with a massive number of parameters, which require large memory size, extensive computational operations, and high energy consumption in neuromorphic hardware systems. In this work, we propose an unsupervised online adaptive weight pruning method that dynamically removes non-critical weights from a spiking neural network (SNN) to reduce network complexity and improve energy efficiency. The adaptive pruning method explores neural dynamics and firing activity of SNNs and adapts the pruning threshold over time and neurons during training. The proposed adaptation scheme allows the network to effectively identify critical weights associated with each neuron by changing the pruning threshold dynamically over time and neurons. It balances the connection strength of neurons with the previous layer with adaptive thresholds and prevents weak neurons from failure after pruning. We also evaluated improvement in the energy efficiency of SNNs with our method by computing synaptic operations (SOPs). Simulation results and detailed analyses have revealed that applying adaptation in the pruning threshold can significantly improve network performance and reduce the number of SOPs. The pruned SNN with 800 excitatory neurons can achieve a 30% reduction in SOPs during training and a 55% reduction during inference, with only 0.44% accuracy loss on MNIST dataset. Compared with a previously reported online soft pruning method, the proposed adaptive pruning method shows 3.33% higher classification accuracy and 67% more reduction in SOPs. The effectiveness of our method was confirmed on different datasets and for different network sizes. Our evaluation showed that the implementation overhead of the adaptive method regarding speed, area, and energy is negligible in the network. Therefore, this work offers a promising solution for effective network compression and building highly energy-efficient neuromorphic systems in real-time applications.
format Online
Article
Text
id pubmed-7689062
institution National Center for Biotechnology Information
language English
publishDate 2020
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-76890622020-12-03 Unsupervised Adaptive Weight Pruning for Energy-Efficient Neuromorphic Systems Guo, Wenzhe Fouda, Mohammed E. Yantir, Hasan Erdem Eltawil, Ahmed M. Salama, Khaled Nabil Front Neurosci Neuroscience To tackle real-world challenges, deep and complex neural networks are generally used with a massive number of parameters, which require large memory size, extensive computational operations, and high energy consumption in neuromorphic hardware systems. In this work, we propose an unsupervised online adaptive weight pruning method that dynamically removes non-critical weights from a spiking neural network (SNN) to reduce network complexity and improve energy efficiency. The adaptive pruning method explores neural dynamics and firing activity of SNNs and adapts the pruning threshold over time and neurons during training. The proposed adaptation scheme allows the network to effectively identify critical weights associated with each neuron by changing the pruning threshold dynamically over time and neurons. It balances the connection strength of neurons with the previous layer with adaptive thresholds and prevents weak neurons from failure after pruning. We also evaluated improvement in the energy efficiency of SNNs with our method by computing synaptic operations (SOPs). Simulation results and detailed analyses have revealed that applying adaptation in the pruning threshold can significantly improve network performance and reduce the number of SOPs. The pruned SNN with 800 excitatory neurons can achieve a 30% reduction in SOPs during training and a 55% reduction during inference, with only 0.44% accuracy loss on MNIST dataset. Compared with a previously reported online soft pruning method, the proposed adaptive pruning method shows 3.33% higher classification accuracy and 67% more reduction in SOPs. The effectiveness of our method was confirmed on different datasets and for different network sizes. Our evaluation showed that the implementation overhead of the adaptive method regarding speed, area, and energy is negligible in the network. Therefore, this work offers a promising solution for effective network compression and building highly energy-efficient neuromorphic systems in real-time applications. Frontiers Media S.A. 2020-11-12 /pmc/articles/PMC7689062/ /pubmed/33281549 http://dx.doi.org/10.3389/fnins.2020.598876 Text en Copyright © 2020 Guo, Fouda, Yantir, Eltawil and Salama. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Neuroscience
Guo, Wenzhe
Fouda, Mohammed E.
Yantir, Hasan Erdem
Eltawil, Ahmed M.
Salama, Khaled Nabil
Unsupervised Adaptive Weight Pruning for Energy-Efficient Neuromorphic Systems
title Unsupervised Adaptive Weight Pruning for Energy-Efficient Neuromorphic Systems
title_full Unsupervised Adaptive Weight Pruning for Energy-Efficient Neuromorphic Systems
title_fullStr Unsupervised Adaptive Weight Pruning for Energy-Efficient Neuromorphic Systems
title_full_unstemmed Unsupervised Adaptive Weight Pruning for Energy-Efficient Neuromorphic Systems
title_short Unsupervised Adaptive Weight Pruning for Energy-Efficient Neuromorphic Systems
title_sort unsupervised adaptive weight pruning for energy-efficient neuromorphic systems
topic Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7689062/
https://www.ncbi.nlm.nih.gov/pubmed/33281549
http://dx.doi.org/10.3389/fnins.2020.598876
work_keys_str_mv AT guowenzhe unsupervisedadaptiveweightpruningforenergyefficientneuromorphicsystems
AT foudamohammede unsupervisedadaptiveweightpruningforenergyefficientneuromorphicsystems
AT yantirhasanerdem unsupervisedadaptiveweightpruningforenergyefficientneuromorphicsystems
AT eltawilahmedm unsupervisedadaptiveweightpruningforenergyefficientneuromorphicsystems
AT salamakhalednabil unsupervisedadaptiveweightpruningforenergyefficientneuromorphicsystems