Cargando…
GradFreeBits: Gradient-Free Bit Allocation for Mixed-Precision Neural Networks
Quantized neural networks (QNNs) are among the main approaches for deploying deep neural networks on low-resource edge devices. Training QNNs using different levels of precision throughout the network (mixed-precision quantization) typically achieves superior trade-offs between performance and compu...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9787339/ https://www.ncbi.nlm.nih.gov/pubmed/36560141 http://dx.doi.org/10.3390/s22249772 |
_version_ | 1784858487767957504 |
---|---|
author | Bodner, Benjamin Jacob Ben-Shalom, Gil Treister, Eran |
author_facet | Bodner, Benjamin Jacob Ben-Shalom, Gil Treister, Eran |
author_sort | Bodner, Benjamin Jacob |
collection | PubMed |
description | Quantized neural networks (QNNs) are among the main approaches for deploying deep neural networks on low-resource edge devices. Training QNNs using different levels of precision throughout the network (mixed-precision quantization) typically achieves superior trade-offs between performance and computational load. However, optimizing the different precision levels of QNNs can be complicated, as the values of the bit allocations are discrete and difficult to differentiate for. Moreover, adequately accounting for the dependencies between the bit allocation of different layers is not straightforward. To meet these challenges, in this work, we propose GradFreeBits: a novel joint optimization scheme for training mixed-precision QNNs, which alternates between gradient-based optimization for the weights and gradient-free optimization for the bit allocation. Our method achieves a better or on par performance with the current state-of-the-art low-precision classification networks on CIFAR10/100 and ImageNet, semantic segmentation networks on Cityscapes, and several graph neural networks benchmarks. Furthermore, our approach can be extended to a variety of other applications involving neural networks used in conjunction with parameters that are difficult to optimize for. |
format | Online Article Text |
id | pubmed-9787339 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-97873392022-12-24 GradFreeBits: Gradient-Free Bit Allocation for Mixed-Precision Neural Networks Bodner, Benjamin Jacob Ben-Shalom, Gil Treister, Eran Sensors (Basel) Article Quantized neural networks (QNNs) are among the main approaches for deploying deep neural networks on low-resource edge devices. Training QNNs using different levels of precision throughout the network (mixed-precision quantization) typically achieves superior trade-offs between performance and computational load. However, optimizing the different precision levels of QNNs can be complicated, as the values of the bit allocations are discrete and difficult to differentiate for. Moreover, adequately accounting for the dependencies between the bit allocation of different layers is not straightforward. To meet these challenges, in this work, we propose GradFreeBits: a novel joint optimization scheme for training mixed-precision QNNs, which alternates between gradient-based optimization for the weights and gradient-free optimization for the bit allocation. Our method achieves a better or on par performance with the current state-of-the-art low-precision classification networks on CIFAR10/100 and ImageNet, semantic segmentation networks on Cityscapes, and several graph neural networks benchmarks. Furthermore, our approach can be extended to a variety of other applications involving neural networks used in conjunction with parameters that are difficult to optimize for. MDPI 2022-12-13 /pmc/articles/PMC9787339/ /pubmed/36560141 http://dx.doi.org/10.3390/s22249772 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Bodner, Benjamin Jacob Ben-Shalom, Gil Treister, Eran GradFreeBits: Gradient-Free Bit Allocation for Mixed-Precision Neural Networks |
title | GradFreeBits: Gradient-Free Bit Allocation for Mixed-Precision Neural Networks |
title_full | GradFreeBits: Gradient-Free Bit Allocation for Mixed-Precision Neural Networks |
title_fullStr | GradFreeBits: Gradient-Free Bit Allocation for Mixed-Precision Neural Networks |
title_full_unstemmed | GradFreeBits: Gradient-Free Bit Allocation for Mixed-Precision Neural Networks |
title_short | GradFreeBits: Gradient-Free Bit Allocation for Mixed-Precision Neural Networks |
title_sort | gradfreebits: gradient-free bit allocation for mixed-precision neural networks |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9787339/ https://www.ncbi.nlm.nih.gov/pubmed/36560141 http://dx.doi.org/10.3390/s22249772 |
work_keys_str_mv | AT bodnerbenjaminjacob gradfreebitsgradientfreebitallocationformixedprecisionneuralnetworks AT benshalomgil gradfreebitsgradientfreebitallocationformixedprecisionneuralnetworks AT treistereran gradfreebitsgradientfreebitallocationformixedprecisionneuralnetworks |