Cargando…
Rethinking Weight Decay for Efficient Neural Network Pruning
Introduced in the late 1980s for generalization purposes, pruning has now become a staple for compressing deep neural networks. Despite many innovations in recent decades, pruning approaches still face core issues that hinder their performance or scalability. Drawing inspiration from early work in t...
Autores principales: | , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8950981/ https://www.ncbi.nlm.nih.gov/pubmed/35324619 http://dx.doi.org/10.3390/jimaging8030064 |
_version_ | 1784675273695821824 |
---|---|
author | Tessier, Hugo Gripon, Vincent Léonardon, Mathieu Arzel, Matthieu Hannagan, Thomas Bertrand, David |
author_facet | Tessier, Hugo Gripon, Vincent Léonardon, Mathieu Arzel, Matthieu Hannagan, Thomas Bertrand, David |
author_sort | Tessier, Hugo |
collection | PubMed |
description | Introduced in the late 1980s for generalization purposes, pruning has now become a staple for compressing deep neural networks. Despite many innovations in recent decades, pruning approaches still face core issues that hinder their performance or scalability. Drawing inspiration from early work in the field, and especially the use of weight decay to achieve sparsity, we introduce Selective Weight Decay (SWD), which carries out efficient, continuous pruning throughout training. Our approach, theoretically grounded on Lagrangian smoothing, is versatile and can be applied to multiple tasks, networks, and pruning structures. We show that SWD compares favorably to state-of-the-art approaches, in terms of performance-to-parameters ratio, on the CIFAR-10, Cora, and ImageNet ILSVRC2012 datasets. |
format | Online Article Text |
id | pubmed-8950981 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-89509812022-03-26 Rethinking Weight Decay for Efficient Neural Network Pruning Tessier, Hugo Gripon, Vincent Léonardon, Mathieu Arzel, Matthieu Hannagan, Thomas Bertrand, David J Imaging Article Introduced in the late 1980s for generalization purposes, pruning has now become a staple for compressing deep neural networks. Despite many innovations in recent decades, pruning approaches still face core issues that hinder their performance or scalability. Drawing inspiration from early work in the field, and especially the use of weight decay to achieve sparsity, we introduce Selective Weight Decay (SWD), which carries out efficient, continuous pruning throughout training. Our approach, theoretically grounded on Lagrangian smoothing, is versatile and can be applied to multiple tasks, networks, and pruning structures. We show that SWD compares favorably to state-of-the-art approaches, in terms of performance-to-parameters ratio, on the CIFAR-10, Cora, and ImageNet ILSVRC2012 datasets. MDPI 2022-03-04 /pmc/articles/PMC8950981/ /pubmed/35324619 http://dx.doi.org/10.3390/jimaging8030064 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Tessier, Hugo Gripon, Vincent Léonardon, Mathieu Arzel, Matthieu Hannagan, Thomas Bertrand, David Rethinking Weight Decay for Efficient Neural Network Pruning |
title | Rethinking Weight Decay for Efficient Neural Network Pruning |
title_full | Rethinking Weight Decay for Efficient Neural Network Pruning |
title_fullStr | Rethinking Weight Decay for Efficient Neural Network Pruning |
title_full_unstemmed | Rethinking Weight Decay for Efficient Neural Network Pruning |
title_short | Rethinking Weight Decay for Efficient Neural Network Pruning |
title_sort | rethinking weight decay for efficient neural network pruning |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8950981/ https://www.ncbi.nlm.nih.gov/pubmed/35324619 http://dx.doi.org/10.3390/jimaging8030064 |
work_keys_str_mv | AT tessierhugo rethinkingweightdecayforefficientneuralnetworkpruning AT griponvincent rethinkingweightdecayforefficientneuralnetworkpruning AT leonardonmathieu rethinkingweightdecayforefficientneuralnetworkpruning AT arzelmatthieu rethinkingweightdecayforefficientneuralnetworkpruning AT hannaganthomas rethinkingweightdecayforefficientneuralnetworkpruning AT bertranddavid rethinkingweightdecayforefficientneuralnetworkpruning |