Cargando…
Unsupervised Adaptive Weight Pruning for Energy-Efficient Neuromorphic Systems
To tackle real-world challenges, deep and complex neural networks are generally used with a massive number of parameters, which require large memory size, extensive computational operations, and high energy consumption in neuromorphic hardware systems. In this work, we propose an unsupervised online...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Frontiers Media S.A.
2020
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7689062/ https://www.ncbi.nlm.nih.gov/pubmed/33281549 http://dx.doi.org/10.3389/fnins.2020.598876 |
_version_ | 1783613782948315136 |
---|---|
author | Guo, Wenzhe Fouda, Mohammed E. Yantir, Hasan Erdem Eltawil, Ahmed M. Salama, Khaled Nabil |
author_facet | Guo, Wenzhe Fouda, Mohammed E. Yantir, Hasan Erdem Eltawil, Ahmed M. Salama, Khaled Nabil |
author_sort | Guo, Wenzhe |
collection | PubMed |
description | To tackle real-world challenges, deep and complex neural networks are generally used with a massive number of parameters, which require large memory size, extensive computational operations, and high energy consumption in neuromorphic hardware systems. In this work, we propose an unsupervised online adaptive weight pruning method that dynamically removes non-critical weights from a spiking neural network (SNN) to reduce network complexity and improve energy efficiency. The adaptive pruning method explores neural dynamics and firing activity of SNNs and adapts the pruning threshold over time and neurons during training. The proposed adaptation scheme allows the network to effectively identify critical weights associated with each neuron by changing the pruning threshold dynamically over time and neurons. It balances the connection strength of neurons with the previous layer with adaptive thresholds and prevents weak neurons from failure after pruning. We also evaluated improvement in the energy efficiency of SNNs with our method by computing synaptic operations (SOPs). Simulation results and detailed analyses have revealed that applying adaptation in the pruning threshold can significantly improve network performance and reduce the number of SOPs. The pruned SNN with 800 excitatory neurons can achieve a 30% reduction in SOPs during training and a 55% reduction during inference, with only 0.44% accuracy loss on MNIST dataset. Compared with a previously reported online soft pruning method, the proposed adaptive pruning method shows 3.33% higher classification accuracy and 67% more reduction in SOPs. The effectiveness of our method was confirmed on different datasets and for different network sizes. Our evaluation showed that the implementation overhead of the adaptive method regarding speed, area, and energy is negligible in the network. Therefore, this work offers a promising solution for effective network compression and building highly energy-efficient neuromorphic systems in real-time applications. |
format | Online Article Text |
id | pubmed-7689062 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2020 |
publisher | Frontiers Media S.A. |
record_format | MEDLINE/PubMed |
spelling | pubmed-76890622020-12-03 Unsupervised Adaptive Weight Pruning for Energy-Efficient Neuromorphic Systems Guo, Wenzhe Fouda, Mohammed E. Yantir, Hasan Erdem Eltawil, Ahmed M. Salama, Khaled Nabil Front Neurosci Neuroscience To tackle real-world challenges, deep and complex neural networks are generally used with a massive number of parameters, which require large memory size, extensive computational operations, and high energy consumption in neuromorphic hardware systems. In this work, we propose an unsupervised online adaptive weight pruning method that dynamically removes non-critical weights from a spiking neural network (SNN) to reduce network complexity and improve energy efficiency. The adaptive pruning method explores neural dynamics and firing activity of SNNs and adapts the pruning threshold over time and neurons during training. The proposed adaptation scheme allows the network to effectively identify critical weights associated with each neuron by changing the pruning threshold dynamically over time and neurons. It balances the connection strength of neurons with the previous layer with adaptive thresholds and prevents weak neurons from failure after pruning. We also evaluated improvement in the energy efficiency of SNNs with our method by computing synaptic operations (SOPs). Simulation results and detailed analyses have revealed that applying adaptation in the pruning threshold can significantly improve network performance and reduce the number of SOPs. The pruned SNN with 800 excitatory neurons can achieve a 30% reduction in SOPs during training and a 55% reduction during inference, with only 0.44% accuracy loss on MNIST dataset. Compared with a previously reported online soft pruning method, the proposed adaptive pruning method shows 3.33% higher classification accuracy and 67% more reduction in SOPs. The effectiveness of our method was confirmed on different datasets and for different network sizes. Our evaluation showed that the implementation overhead of the adaptive method regarding speed, area, and energy is negligible in the network. Therefore, this work offers a promising solution for effective network compression and building highly energy-efficient neuromorphic systems in real-time applications. Frontiers Media S.A. 2020-11-12 /pmc/articles/PMC7689062/ /pubmed/33281549 http://dx.doi.org/10.3389/fnins.2020.598876 Text en Copyright © 2020 Guo, Fouda, Yantir, Eltawil and Salama. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. |
spellingShingle | Neuroscience Guo, Wenzhe Fouda, Mohammed E. Yantir, Hasan Erdem Eltawil, Ahmed M. Salama, Khaled Nabil Unsupervised Adaptive Weight Pruning for Energy-Efficient Neuromorphic Systems |
title | Unsupervised Adaptive Weight Pruning for Energy-Efficient Neuromorphic Systems |
title_full | Unsupervised Adaptive Weight Pruning for Energy-Efficient Neuromorphic Systems |
title_fullStr | Unsupervised Adaptive Weight Pruning for Energy-Efficient Neuromorphic Systems |
title_full_unstemmed | Unsupervised Adaptive Weight Pruning for Energy-Efficient Neuromorphic Systems |
title_short | Unsupervised Adaptive Weight Pruning for Energy-Efficient Neuromorphic Systems |
title_sort | unsupervised adaptive weight pruning for energy-efficient neuromorphic systems |
topic | Neuroscience |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7689062/ https://www.ncbi.nlm.nih.gov/pubmed/33281549 http://dx.doi.org/10.3389/fnins.2020.598876 |
work_keys_str_mv | AT guowenzhe unsupervisedadaptiveweightpruningforenergyefficientneuromorphicsystems AT foudamohammede unsupervisedadaptiveweightpruningforenergyefficientneuromorphicsystems AT yantirhasanerdem unsupervisedadaptiveweightpruningforenergyefficientneuromorphicsystems AT eltawilahmedm unsupervisedadaptiveweightpruningforenergyefficientneuromorphicsystems AT salamakhalednabil unsupervisedadaptiveweightpruningforenergyefficientneuromorphicsystems |