Cargando…

Dynamically Optimizing Network Structure Based on Synaptic Pruning in the Brain

Most neural networks need to predefine the network architecture empirically, which may cause over-fitting or under-fitting. Besides, a large number of parameters in a fully connected network leads to the prohibitively expensive computational cost and storage overhead, which makes the model hard to b...

Descripción completa

Detalles Bibliográficos
Autores principales: Zhao, Feifei, Zeng, Yi
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8220807/
https://www.ncbi.nlm.nih.gov/pubmed/34177473
http://dx.doi.org/10.3389/fnsys.2021.620558
_version_ 1783711218701172736
author Zhao, Feifei
Zeng, Yi
author_facet Zhao, Feifei
Zeng, Yi
author_sort Zhao, Feifei
collection PubMed
description Most neural networks need to predefine the network architecture empirically, which may cause over-fitting or under-fitting. Besides, a large number of parameters in a fully connected network leads to the prohibitively expensive computational cost and storage overhead, which makes the model hard to be deployed on mobile devices. Dynamically optimizing the network architecture by pruning unused synapses is a promising technique for solving this problem. Most existing pruning methods focus on reducing the redundancy of deep convolutional neural networks by pruning unimportant filters or weights, at the cost of accuracy drop. In this paper, we propose an effective brain-inspired synaptic pruning method to dynamically modulate the network architecture and simultaneously improve network performance. The proposed model is biologically inspired as it dynamically eliminates redundant connections based on the synaptic pruning rules used during the brain's development. Connections are pruned if they are not activated or less activated multiple times consecutively. Extensive experiments demonstrate the effectiveness of our method on classification tasks of different complexity with the MNIST, Fashion MNIST, and CIFAR-10 datasets. Experimental results reveal that even for a compact network, the proposed method can also remove up to 59–90% of the connections, with relative improvement in learning speed and accuracy.
format Online
Article
Text
id pubmed-8220807
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-82208072021-06-24 Dynamically Optimizing Network Structure Based on Synaptic Pruning in the Brain Zhao, Feifei Zeng, Yi Front Syst Neurosci Neuroscience Most neural networks need to predefine the network architecture empirically, which may cause over-fitting or under-fitting. Besides, a large number of parameters in a fully connected network leads to the prohibitively expensive computational cost and storage overhead, which makes the model hard to be deployed on mobile devices. Dynamically optimizing the network architecture by pruning unused synapses is a promising technique for solving this problem. Most existing pruning methods focus on reducing the redundancy of deep convolutional neural networks by pruning unimportant filters or weights, at the cost of accuracy drop. In this paper, we propose an effective brain-inspired synaptic pruning method to dynamically modulate the network architecture and simultaneously improve network performance. The proposed model is biologically inspired as it dynamically eliminates redundant connections based on the synaptic pruning rules used during the brain's development. Connections are pruned if they are not activated or less activated multiple times consecutively. Extensive experiments demonstrate the effectiveness of our method on classification tasks of different complexity with the MNIST, Fashion MNIST, and CIFAR-10 datasets. Experimental results reveal that even for a compact network, the proposed method can also remove up to 59–90% of the connections, with relative improvement in learning speed and accuracy. Frontiers Media S.A. 2021-06-04 /pmc/articles/PMC8220807/ /pubmed/34177473 http://dx.doi.org/10.3389/fnsys.2021.620558 Text en Copyright © 2021 Zhao and Zeng. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Neuroscience
Zhao, Feifei
Zeng, Yi
Dynamically Optimizing Network Structure Based on Synaptic Pruning in the Brain
title Dynamically Optimizing Network Structure Based on Synaptic Pruning in the Brain
title_full Dynamically Optimizing Network Structure Based on Synaptic Pruning in the Brain
title_fullStr Dynamically Optimizing Network Structure Based on Synaptic Pruning in the Brain
title_full_unstemmed Dynamically Optimizing Network Structure Based on Synaptic Pruning in the Brain
title_short Dynamically Optimizing Network Structure Based on Synaptic Pruning in the Brain
title_sort dynamically optimizing network structure based on synaptic pruning in the brain
topic Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8220807/
https://www.ncbi.nlm.nih.gov/pubmed/34177473
http://dx.doi.org/10.3389/fnsys.2021.620558
work_keys_str_mv AT zhaofeifei dynamicallyoptimizingnetworkstructurebasedonsynapticpruninginthebrain
AT zengyi dynamicallyoptimizingnetworkstructurebasedonsynapticpruninginthebrain