Cargando…

Differential Evolution Based Layer-Wise Weight Pruning for Compressing Deep Neural Networks

Deep neural networks have evolved significantly in the past decades and are now able to achieve better progression of sensor data. Nonetheless, most of the deep models verify the ruling maxim in deep learning—bigger is better—so they have very complex structures. As the models become more complex, t...

Descripción completa

Detalles Bibliográficos
Autores principales: Wu, Tao, Li, Xiaoyang, Zhou, Deyun, Li, Na, Shi, Jiao
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7865320/
https://www.ncbi.nlm.nih.gov/pubmed/33525527
http://dx.doi.org/10.3390/s21030880
_version_ 1783647818632658944
author Wu, Tao
Li, Xiaoyang
Zhou, Deyun
Li, Na
Shi, Jiao
author_facet Wu, Tao
Li, Xiaoyang
Zhou, Deyun
Li, Na
Shi, Jiao
author_sort Wu, Tao
collection PubMed
description Deep neural networks have evolved significantly in the past decades and are now able to achieve better progression of sensor data. Nonetheless, most of the deep models verify the ruling maxim in deep learning—bigger is better—so they have very complex structures. As the models become more complex, the computational complexity and resource consumption of these deep models are increasing significantly, making them difficult to perform on resource-limited platforms, such as sensor platforms. In this paper, we observe that different layers often have different pruning requirements, and propose a differential evolutionary layer-wise weight pruning method. Firstly, the pruning sensitivity of each layer is analyzed, and then the network is compressed by iterating the weight pruning process. Unlike some other methods that deal with pruning ratio by greedy ways or statistical analysis, we establish an optimization model to find the optimal pruning sensitivity set for each layer. Differential evolution is an effective method based on population optimization which can be used to address this task. Furthermore, we adopt a strategy to recovery some of the removed connections to increase the capacity of the pruned model during the fine-tuning phase. The effectiveness of our method has been demonstrated in experimental studies. Our method compresses the number of weight parameters in LeNet-300-100, LeNet-5, AlexNet and VGG16 by [Formula: see text] , [Formula: see text] , [Formula: see text] and [Formula: see text] , respectively.
format Online
Article
Text
id pubmed-7865320
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-78653202021-02-07 Differential Evolution Based Layer-Wise Weight Pruning for Compressing Deep Neural Networks Wu, Tao Li, Xiaoyang Zhou, Deyun Li, Na Shi, Jiao Sensors (Basel) Article Deep neural networks have evolved significantly in the past decades and are now able to achieve better progression of sensor data. Nonetheless, most of the deep models verify the ruling maxim in deep learning—bigger is better—so they have very complex structures. As the models become more complex, the computational complexity and resource consumption of these deep models are increasing significantly, making them difficult to perform on resource-limited platforms, such as sensor platforms. In this paper, we observe that different layers often have different pruning requirements, and propose a differential evolutionary layer-wise weight pruning method. Firstly, the pruning sensitivity of each layer is analyzed, and then the network is compressed by iterating the weight pruning process. Unlike some other methods that deal with pruning ratio by greedy ways or statistical analysis, we establish an optimization model to find the optimal pruning sensitivity set for each layer. Differential evolution is an effective method based on population optimization which can be used to address this task. Furthermore, we adopt a strategy to recovery some of the removed connections to increase the capacity of the pruned model during the fine-tuning phase. The effectiveness of our method has been demonstrated in experimental studies. Our method compresses the number of weight parameters in LeNet-300-100, LeNet-5, AlexNet and VGG16 by [Formula: see text] , [Formula: see text] , [Formula: see text] and [Formula: see text] , respectively. MDPI 2021-01-28 /pmc/articles/PMC7865320/ /pubmed/33525527 http://dx.doi.org/10.3390/s21030880 Text en © 2021 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Wu, Tao
Li, Xiaoyang
Zhou, Deyun
Li, Na
Shi, Jiao
Differential Evolution Based Layer-Wise Weight Pruning for Compressing Deep Neural Networks
title Differential Evolution Based Layer-Wise Weight Pruning for Compressing Deep Neural Networks
title_full Differential Evolution Based Layer-Wise Weight Pruning for Compressing Deep Neural Networks
title_fullStr Differential Evolution Based Layer-Wise Weight Pruning for Compressing Deep Neural Networks
title_full_unstemmed Differential Evolution Based Layer-Wise Weight Pruning for Compressing Deep Neural Networks
title_short Differential Evolution Based Layer-Wise Weight Pruning for Compressing Deep Neural Networks
title_sort differential evolution based layer-wise weight pruning for compressing deep neural networks
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7865320/
https://www.ncbi.nlm.nih.gov/pubmed/33525527
http://dx.doi.org/10.3390/s21030880
work_keys_str_mv AT wutao differentialevolutionbasedlayerwiseweightpruningforcompressingdeepneuralnetworks
AT lixiaoyang differentialevolutionbasedlayerwiseweightpruningforcompressingdeepneuralnetworks
AT zhoudeyun differentialevolutionbasedlayerwiseweightpruningforcompressingdeepneuralnetworks
AT lina differentialevolutionbasedlayerwiseweightpruningforcompressingdeepneuralnetworks
AT shijiao differentialevolutionbasedlayerwiseweightpruningforcompressingdeepneuralnetworks