Cargando…

Retrain or Not Retrain? - Efficient Pruning Methods of Deep CNN Networks

Nowadays, convolutional neural networks (CNN) play a major role in image processing tasks like image classification, object detection, semantic segmentation. Very often CNN networks have from several to hundred stacked layers with several megabytes of weights. One of the possible techniques to reduc...

Descripción completa

Detalles Bibliográficos
Autores principales: Pietron, Marcin, Wielgosz, Maciej
Formato: Online Artículo Texto
Lenguaje:English
Publicado: 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7304033/
http://dx.doi.org/10.1007/978-3-030-50420-5_34
_version_ 1783548184599986176
author Pietron, Marcin
Wielgosz, Maciej
author_facet Pietron, Marcin
Wielgosz, Maciej
author_sort Pietron, Marcin
collection PubMed
description Nowadays, convolutional neural networks (CNN) play a major role in image processing tasks like image classification, object detection, semantic segmentation. Very often CNN networks have from several to hundred stacked layers with several megabytes of weights. One of the possible techniques to reduce complexity and memory footprint is pruning. Pruning is a process of removing weights which connect neurons from two adjacent layers in the network. The process of finding near optimal solution with specified and acceptable drop in accuracy can be more sophisticated when DL model has higher number of convolutional layers. In the paper few approaches based on retraining and no retraining are described and compared together.
format Online
Article
Text
id pubmed-7304033
institution National Center for Biotechnology Information
language English
publishDate 2020
record_format MEDLINE/PubMed
spelling pubmed-73040332020-06-19 Retrain or Not Retrain? - Efficient Pruning Methods of Deep CNN Networks Pietron, Marcin Wielgosz, Maciej Computational Science – ICCS 2020 Article Nowadays, convolutional neural networks (CNN) play a major role in image processing tasks like image classification, object detection, semantic segmentation. Very often CNN networks have from several to hundred stacked layers with several megabytes of weights. One of the possible techniques to reduce complexity and memory footprint is pruning. Pruning is a process of removing weights which connect neurons from two adjacent layers in the network. The process of finding near optimal solution with specified and acceptable drop in accuracy can be more sophisticated when DL model has higher number of convolutional layers. In the paper few approaches based on retraining and no retraining are described and compared together. 2020-05-22 /pmc/articles/PMC7304033/ http://dx.doi.org/10.1007/978-3-030-50420-5_34 Text en © Springer Nature Switzerland AG 2020 This article is made available via the PMC Open Access Subset for unrestricted research re-use and secondary analysis in any form or by any means with acknowledgement of the original source. These permissions are granted for the duration of the World Health Organization (WHO) declaration of COVID-19 as a global pandemic.
spellingShingle Article
Pietron, Marcin
Wielgosz, Maciej
Retrain or Not Retrain? - Efficient Pruning Methods of Deep CNN Networks
title Retrain or Not Retrain? - Efficient Pruning Methods of Deep CNN Networks
title_full Retrain or Not Retrain? - Efficient Pruning Methods of Deep CNN Networks
title_fullStr Retrain or Not Retrain? - Efficient Pruning Methods of Deep CNN Networks
title_full_unstemmed Retrain or Not Retrain? - Efficient Pruning Methods of Deep CNN Networks
title_short Retrain or Not Retrain? - Efficient Pruning Methods of Deep CNN Networks
title_sort retrain or not retrain? - efficient pruning methods of deep cnn networks
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7304033/
http://dx.doi.org/10.1007/978-3-030-50420-5_34
work_keys_str_mv AT pietronmarcin retrainornotretrainefficientpruningmethodsofdeepcnnnetworks
AT wielgoszmaciej retrainornotretrainefficientpruningmethodsofdeepcnnnetworks