Cargando…

Model pruning based on filter similarity for edge device deployment

Filter pruning is widely used for inference acceleration and compatibility with off-the-shelf hardware devices. Some filter pruning methods have proposed various criteria to approximate the importance of filters, and then sort the filters globally or locally to prune the redundant parameters. Howeve...

Descripción completa

Detalles Bibliográficos
Autores principales: Wu, Tingting, Song, Chunhe, Zeng, Peng
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10017522/
https://www.ncbi.nlm.nih.gov/pubmed/36937554
http://dx.doi.org/10.3389/fnbot.2023.1132679
_version_ 1784907606501883904
author Wu, Tingting
Song, Chunhe
Zeng, Peng
author_facet Wu, Tingting
Song, Chunhe
Zeng, Peng
author_sort Wu, Tingting
collection PubMed
description Filter pruning is widely used for inference acceleration and compatibility with off-the-shelf hardware devices. Some filter pruning methods have proposed various criteria to approximate the importance of filters, and then sort the filters globally or locally to prune the redundant parameters. However, the current criterion-based methods have problems: (1) parameters with smaller criterion values for extracting edge features are easily ignored, and (2) there is a strong correlation between different criteria, resulting in similar pruning structures. In this article, we propose a novel simple but effective pruning method based on filter similarity, which is used to evaluate the similarity between filters instead of the importance of a single filter. The proposed method first calculates the similarity of the filters pairwise in one convolutional layer and then obtains the similarity distribution. Finally, the filters with high similarity to others are deleted from the distribution or set to zero. In addition, the proposed algorithm does not need to specify the pruning rate for each layer, and only needs to set the desired FLOPs or parameter reduction to obtain the final compression model. We also provide iterative pruning strategies for hard pruning and soft pruning to satisfy the tradeoff requirements of accuracy and memory in different scenarios. Extensive experiments on various representative benchmark datasets across different network architectures demonstrate the effectiveness of our proposed method. For example, on CIFAR10, the proposed algorithm achieves 61.1% FLOPs reduction by removing 58.3% of the parameters, with no loss in Top-1 accuracy on ResNet-56; and reduces 53.05% FLOPs on ResNet-50 with only 0.29% Top-1 accuracy degradation on ILSVRC-2012.
format Online
Article
Text
id pubmed-10017522
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-100175222023-03-17 Model pruning based on filter similarity for edge device deployment Wu, Tingting Song, Chunhe Zeng, Peng Front Neurorobot Neuroscience Filter pruning is widely used for inference acceleration and compatibility with off-the-shelf hardware devices. Some filter pruning methods have proposed various criteria to approximate the importance of filters, and then sort the filters globally or locally to prune the redundant parameters. However, the current criterion-based methods have problems: (1) parameters with smaller criterion values for extracting edge features are easily ignored, and (2) there is a strong correlation between different criteria, resulting in similar pruning structures. In this article, we propose a novel simple but effective pruning method based on filter similarity, which is used to evaluate the similarity between filters instead of the importance of a single filter. The proposed method first calculates the similarity of the filters pairwise in one convolutional layer and then obtains the similarity distribution. Finally, the filters with high similarity to others are deleted from the distribution or set to zero. In addition, the proposed algorithm does not need to specify the pruning rate for each layer, and only needs to set the desired FLOPs or parameter reduction to obtain the final compression model. We also provide iterative pruning strategies for hard pruning and soft pruning to satisfy the tradeoff requirements of accuracy and memory in different scenarios. Extensive experiments on various representative benchmark datasets across different network architectures demonstrate the effectiveness of our proposed method. For example, on CIFAR10, the proposed algorithm achieves 61.1% FLOPs reduction by removing 58.3% of the parameters, with no loss in Top-1 accuracy on ResNet-56; and reduces 53.05% FLOPs on ResNet-50 with only 0.29% Top-1 accuracy degradation on ILSVRC-2012. Frontiers Media S.A. 2023-03-02 /pmc/articles/PMC10017522/ /pubmed/36937554 http://dx.doi.org/10.3389/fnbot.2023.1132679 Text en Copyright © 2023 Wu, Song and Zeng. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Neuroscience
Wu, Tingting
Song, Chunhe
Zeng, Peng
Model pruning based on filter similarity for edge device deployment
title Model pruning based on filter similarity for edge device deployment
title_full Model pruning based on filter similarity for edge device deployment
title_fullStr Model pruning based on filter similarity for edge device deployment
title_full_unstemmed Model pruning based on filter similarity for edge device deployment
title_short Model pruning based on filter similarity for edge device deployment
title_sort model pruning based on filter similarity for edge device deployment
topic Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10017522/
https://www.ncbi.nlm.nih.gov/pubmed/36937554
http://dx.doi.org/10.3389/fnbot.2023.1132679
work_keys_str_mv AT wutingting modelpruningbasedonfiltersimilarityforedgedevicedeployment
AT songchunhe modelpruningbasedonfiltersimilarityforedgedevicedeployment
AT zengpeng modelpruningbasedonfiltersimilarityforedgedevicedeployment