Cargando…
Dynamically Optimizing Network Structure Based on Synaptic Pruning in the Brain
Most neural networks need to predefine the network architecture empirically, which may cause over-fitting or under-fitting. Besides, a large number of parameters in a fully connected network leads to the prohibitively expensive computational cost and storage overhead, which makes the model hard to b...
Autores principales: | , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Frontiers Media S.A.
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8220807/ https://www.ncbi.nlm.nih.gov/pubmed/34177473 http://dx.doi.org/10.3389/fnsys.2021.620558 |
_version_ | 1783711218701172736 |
---|---|
author | Zhao, Feifei Zeng, Yi |
author_facet | Zhao, Feifei Zeng, Yi |
author_sort | Zhao, Feifei |
collection | PubMed |
description | Most neural networks need to predefine the network architecture empirically, which may cause over-fitting or under-fitting. Besides, a large number of parameters in a fully connected network leads to the prohibitively expensive computational cost and storage overhead, which makes the model hard to be deployed on mobile devices. Dynamically optimizing the network architecture by pruning unused synapses is a promising technique for solving this problem. Most existing pruning methods focus on reducing the redundancy of deep convolutional neural networks by pruning unimportant filters or weights, at the cost of accuracy drop. In this paper, we propose an effective brain-inspired synaptic pruning method to dynamically modulate the network architecture and simultaneously improve network performance. The proposed model is biologically inspired as it dynamically eliminates redundant connections based on the synaptic pruning rules used during the brain's development. Connections are pruned if they are not activated or less activated multiple times consecutively. Extensive experiments demonstrate the effectiveness of our method on classification tasks of different complexity with the MNIST, Fashion MNIST, and CIFAR-10 datasets. Experimental results reveal that even for a compact network, the proposed method can also remove up to 59–90% of the connections, with relative improvement in learning speed and accuracy. |
format | Online Article Text |
id | pubmed-8220807 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | Frontiers Media S.A. |
record_format | MEDLINE/PubMed |
spelling | pubmed-82208072021-06-24 Dynamically Optimizing Network Structure Based on Synaptic Pruning in the Brain Zhao, Feifei Zeng, Yi Front Syst Neurosci Neuroscience Most neural networks need to predefine the network architecture empirically, which may cause over-fitting or under-fitting. Besides, a large number of parameters in a fully connected network leads to the prohibitively expensive computational cost and storage overhead, which makes the model hard to be deployed on mobile devices. Dynamically optimizing the network architecture by pruning unused synapses is a promising technique for solving this problem. Most existing pruning methods focus on reducing the redundancy of deep convolutional neural networks by pruning unimportant filters or weights, at the cost of accuracy drop. In this paper, we propose an effective brain-inspired synaptic pruning method to dynamically modulate the network architecture and simultaneously improve network performance. The proposed model is biologically inspired as it dynamically eliminates redundant connections based on the synaptic pruning rules used during the brain's development. Connections are pruned if they are not activated or less activated multiple times consecutively. Extensive experiments demonstrate the effectiveness of our method on classification tasks of different complexity with the MNIST, Fashion MNIST, and CIFAR-10 datasets. Experimental results reveal that even for a compact network, the proposed method can also remove up to 59–90% of the connections, with relative improvement in learning speed and accuracy. Frontiers Media S.A. 2021-06-04 /pmc/articles/PMC8220807/ /pubmed/34177473 http://dx.doi.org/10.3389/fnsys.2021.620558 Text en Copyright © 2021 Zhao and Zeng. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. |
spellingShingle | Neuroscience Zhao, Feifei Zeng, Yi Dynamically Optimizing Network Structure Based on Synaptic Pruning in the Brain |
title | Dynamically Optimizing Network Structure Based on Synaptic Pruning in the Brain |
title_full | Dynamically Optimizing Network Structure Based on Synaptic Pruning in the Brain |
title_fullStr | Dynamically Optimizing Network Structure Based on Synaptic Pruning in the Brain |
title_full_unstemmed | Dynamically Optimizing Network Structure Based on Synaptic Pruning in the Brain |
title_short | Dynamically Optimizing Network Structure Based on Synaptic Pruning in the Brain |
title_sort | dynamically optimizing network structure based on synaptic pruning in the brain |
topic | Neuroscience |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8220807/ https://www.ncbi.nlm.nih.gov/pubmed/34177473 http://dx.doi.org/10.3389/fnsys.2021.620558 |
work_keys_str_mv | AT zhaofeifei dynamicallyoptimizingnetworkstructurebasedonsynapticpruninginthebrain AT zengyi dynamicallyoptimizingnetworkstructurebasedonsynapticpruninginthebrain |