Cargando…
Differential Evolution Based Layer-Wise Weight Pruning for Compressing Deep Neural Networks
Deep neural networks have evolved significantly in the past decades and are now able to achieve better progression of sensor data. Nonetheless, most of the deep models verify the ruling maxim in deep learning—bigger is better—so they have very complex structures. As the models become more complex, t...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7865320/ https://www.ncbi.nlm.nih.gov/pubmed/33525527 http://dx.doi.org/10.3390/s21030880 |
_version_ | 1783647818632658944 |
---|---|
author | Wu, Tao Li, Xiaoyang Zhou, Deyun Li, Na Shi, Jiao |
author_facet | Wu, Tao Li, Xiaoyang Zhou, Deyun Li, Na Shi, Jiao |
author_sort | Wu, Tao |
collection | PubMed |
description | Deep neural networks have evolved significantly in the past decades and are now able to achieve better progression of sensor data. Nonetheless, most of the deep models verify the ruling maxim in deep learning—bigger is better—so they have very complex structures. As the models become more complex, the computational complexity and resource consumption of these deep models are increasing significantly, making them difficult to perform on resource-limited platforms, such as sensor platforms. In this paper, we observe that different layers often have different pruning requirements, and propose a differential evolutionary layer-wise weight pruning method. Firstly, the pruning sensitivity of each layer is analyzed, and then the network is compressed by iterating the weight pruning process. Unlike some other methods that deal with pruning ratio by greedy ways or statistical analysis, we establish an optimization model to find the optimal pruning sensitivity set for each layer. Differential evolution is an effective method based on population optimization which can be used to address this task. Furthermore, we adopt a strategy to recovery some of the removed connections to increase the capacity of the pruned model during the fine-tuning phase. The effectiveness of our method has been demonstrated in experimental studies. Our method compresses the number of weight parameters in LeNet-300-100, LeNet-5, AlexNet and VGG16 by [Formula: see text] , [Formula: see text] , [Formula: see text] and [Formula: see text] , respectively. |
format | Online Article Text |
id | pubmed-7865320 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-78653202021-02-07 Differential Evolution Based Layer-Wise Weight Pruning for Compressing Deep Neural Networks Wu, Tao Li, Xiaoyang Zhou, Deyun Li, Na Shi, Jiao Sensors (Basel) Article Deep neural networks have evolved significantly in the past decades and are now able to achieve better progression of sensor data. Nonetheless, most of the deep models verify the ruling maxim in deep learning—bigger is better—so they have very complex structures. As the models become more complex, the computational complexity and resource consumption of these deep models are increasing significantly, making them difficult to perform on resource-limited platforms, such as sensor platforms. In this paper, we observe that different layers often have different pruning requirements, and propose a differential evolutionary layer-wise weight pruning method. Firstly, the pruning sensitivity of each layer is analyzed, and then the network is compressed by iterating the weight pruning process. Unlike some other methods that deal with pruning ratio by greedy ways or statistical analysis, we establish an optimization model to find the optimal pruning sensitivity set for each layer. Differential evolution is an effective method based on population optimization which can be used to address this task. Furthermore, we adopt a strategy to recovery some of the removed connections to increase the capacity of the pruned model during the fine-tuning phase. The effectiveness of our method has been demonstrated in experimental studies. Our method compresses the number of weight parameters in LeNet-300-100, LeNet-5, AlexNet and VGG16 by [Formula: see text] , [Formula: see text] , [Formula: see text] and [Formula: see text] , respectively. MDPI 2021-01-28 /pmc/articles/PMC7865320/ /pubmed/33525527 http://dx.doi.org/10.3390/s21030880 Text en © 2021 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Wu, Tao Li, Xiaoyang Zhou, Deyun Li, Na Shi, Jiao Differential Evolution Based Layer-Wise Weight Pruning for Compressing Deep Neural Networks |
title | Differential Evolution Based Layer-Wise Weight Pruning for Compressing Deep Neural Networks |
title_full | Differential Evolution Based Layer-Wise Weight Pruning for Compressing Deep Neural Networks |
title_fullStr | Differential Evolution Based Layer-Wise Weight Pruning for Compressing Deep Neural Networks |
title_full_unstemmed | Differential Evolution Based Layer-Wise Weight Pruning for Compressing Deep Neural Networks |
title_short | Differential Evolution Based Layer-Wise Weight Pruning for Compressing Deep Neural Networks |
title_sort | differential evolution based layer-wise weight pruning for compressing deep neural networks |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7865320/ https://www.ncbi.nlm.nih.gov/pubmed/33525527 http://dx.doi.org/10.3390/s21030880 |
work_keys_str_mv | AT wutao differentialevolutionbasedlayerwiseweightpruningforcompressingdeepneuralnetworks AT lixiaoyang differentialevolutionbasedlayerwiseweightpruningforcompressingdeepneuralnetworks AT zhoudeyun differentialevolutionbasedlayerwiseweightpruningforcompressingdeepneuralnetworks AT lina differentialevolutionbasedlayerwiseweightpruningforcompressingdeepneuralnetworks AT shijiao differentialevolutionbasedlayerwiseweightpruningforcompressingdeepneuralnetworks |