Cargando…

Rethinking Weight Decay for Efficient Neural Network Pruning

Introduced in the late 1980s for generalization purposes, pruning has now become a staple for compressing deep neural networks. Despite many innovations in recent decades, pruning approaches still face core issues that hinder their performance or scalability. Drawing inspiration from early work in t...

Descripción completa

Detalles Bibliográficos
Autores principales: Tessier, Hugo, Gripon, Vincent, Léonardon, Mathieu, Arzel, Matthieu, Hannagan, Thomas, Bertrand, David
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8950981/
https://www.ncbi.nlm.nih.gov/pubmed/35324619
http://dx.doi.org/10.3390/jimaging8030064
_version_ 1784675273695821824
author Tessier, Hugo
Gripon, Vincent
Léonardon, Mathieu
Arzel, Matthieu
Hannagan, Thomas
Bertrand, David
author_facet Tessier, Hugo
Gripon, Vincent
Léonardon, Mathieu
Arzel, Matthieu
Hannagan, Thomas
Bertrand, David
author_sort Tessier, Hugo
collection PubMed
description Introduced in the late 1980s for generalization purposes, pruning has now become a staple for compressing deep neural networks. Despite many innovations in recent decades, pruning approaches still face core issues that hinder their performance or scalability. Drawing inspiration from early work in the field, and especially the use of weight decay to achieve sparsity, we introduce Selective Weight Decay (SWD), which carries out efficient, continuous pruning throughout training. Our approach, theoretically grounded on Lagrangian smoothing, is versatile and can be applied to multiple tasks, networks, and pruning structures. We show that SWD compares favorably to state-of-the-art approaches, in terms of performance-to-parameters ratio, on the CIFAR-10, Cora, and ImageNet ILSVRC2012 datasets.
format Online
Article
Text
id pubmed-8950981
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-89509812022-03-26 Rethinking Weight Decay for Efficient Neural Network Pruning Tessier, Hugo Gripon, Vincent Léonardon, Mathieu Arzel, Matthieu Hannagan, Thomas Bertrand, David J Imaging Article Introduced in the late 1980s for generalization purposes, pruning has now become a staple for compressing deep neural networks. Despite many innovations in recent decades, pruning approaches still face core issues that hinder their performance or scalability. Drawing inspiration from early work in the field, and especially the use of weight decay to achieve sparsity, we introduce Selective Weight Decay (SWD), which carries out efficient, continuous pruning throughout training. Our approach, theoretically grounded on Lagrangian smoothing, is versatile and can be applied to multiple tasks, networks, and pruning structures. We show that SWD compares favorably to state-of-the-art approaches, in terms of performance-to-parameters ratio, on the CIFAR-10, Cora, and ImageNet ILSVRC2012 datasets. MDPI 2022-03-04 /pmc/articles/PMC8950981/ /pubmed/35324619 http://dx.doi.org/10.3390/jimaging8030064 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Tessier, Hugo
Gripon, Vincent
Léonardon, Mathieu
Arzel, Matthieu
Hannagan, Thomas
Bertrand, David
Rethinking Weight Decay for Efficient Neural Network Pruning
title Rethinking Weight Decay for Efficient Neural Network Pruning
title_full Rethinking Weight Decay for Efficient Neural Network Pruning
title_fullStr Rethinking Weight Decay for Efficient Neural Network Pruning
title_full_unstemmed Rethinking Weight Decay for Efficient Neural Network Pruning
title_short Rethinking Weight Decay for Efficient Neural Network Pruning
title_sort rethinking weight decay for efficient neural network pruning
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8950981/
https://www.ncbi.nlm.nih.gov/pubmed/35324619
http://dx.doi.org/10.3390/jimaging8030064
work_keys_str_mv AT tessierhugo rethinkingweightdecayforefficientneuralnetworkpruning
AT griponvincent rethinkingweightdecayforefficientneuralnetworkpruning
AT leonardonmathieu rethinkingweightdecayforefficientneuralnetworkpruning
AT arzelmatthieu rethinkingweightdecayforefficientneuralnetworkpruning
AT hannaganthomas rethinkingweightdecayforefficientneuralnetworkpruning
AT bertranddavid rethinkingweightdecayforefficientneuralnetworkpruning