Cargando…

How Many Bits Does it Take to Quantize Your Neural Network?

Quantization converts neural networks into low-bit fixed-point computations which can be carried out by efficient integer-only hardware, and is standard practice for the deployment of neural networks on real-time embedded devices. However, like their real-numbered counterpart, quantized networks are...

Descripción completa

Detalles Bibliográficos
Autores principales: Giacobbe, Mirco, Henzinger, Thomas A., Lechner, Mathias
Formato: Online Artículo Texto
Lenguaje:English
Publicado: 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7480702/
http://dx.doi.org/10.1007/978-3-030-45237-7_5
_version_ 1783580462803845120
author Giacobbe, Mirco
Henzinger, Thomas A.
Lechner, Mathias
author_facet Giacobbe, Mirco
Henzinger, Thomas A.
Lechner, Mathias
author_sort Giacobbe, Mirco
collection PubMed
description Quantization converts neural networks into low-bit fixed-point computations which can be carried out by efficient integer-only hardware, and is standard practice for the deployment of neural networks on real-time embedded devices. However, like their real-numbered counterpart, quantized networks are not immune to malicious misclassification caused by adversarial attacks. We investigate how quantization affects a network’s robustness to adversarial attacks, which is a formal verification question. We show that neither robustness nor non-robustness are monotonic with changing the number of bits for the representation and, also, neither are preserved by quantization from a real-numbered network. For this reason, we introduce a verification method for quantized neural networks which, using SMT solving over bit-vectors, accounts for their exact, bit-precise semantics. We built a tool and analyzed the effect of quantization on a classifier for the MNIST dataset. We demonstrate that, compared to our method, existing methods for the analysis of real-numbered networks often derive false conclusions about their quantizations, both when determining robustness and when detecting attacks, and that existing methods for quantized networks often miss attacks. Furthermore, we applied our method beyond robustness, showing how the number of bits in quantization enlarges the gender bias of a predictor for students’ grades.
format Online
Article
Text
id pubmed-7480702
institution National Center for Biotechnology Information
language English
publishDate 2020
record_format MEDLINE/PubMed
spelling pubmed-74807022020-09-10 How Many Bits Does it Take to Quantize Your Neural Network? Giacobbe, Mirco Henzinger, Thomas A. Lechner, Mathias Tools and Algorithms for the Construction and Analysis of Systems Article Quantization converts neural networks into low-bit fixed-point computations which can be carried out by efficient integer-only hardware, and is standard practice for the deployment of neural networks on real-time embedded devices. However, like their real-numbered counterpart, quantized networks are not immune to malicious misclassification caused by adversarial attacks. We investigate how quantization affects a network’s robustness to adversarial attacks, which is a formal verification question. We show that neither robustness nor non-robustness are monotonic with changing the number of bits for the representation and, also, neither are preserved by quantization from a real-numbered network. For this reason, we introduce a verification method for quantized neural networks which, using SMT solving over bit-vectors, accounts for their exact, bit-precise semantics. We built a tool and analyzed the effect of quantization on a classifier for the MNIST dataset. We demonstrate that, compared to our method, existing methods for the analysis of real-numbered networks often derive false conclusions about their quantizations, both when determining robustness and when detecting attacks, and that existing methods for quantized networks often miss attacks. Furthermore, we applied our method beyond robustness, showing how the number of bits in quantization enlarges the gender bias of a predictor for students’ grades. 2020-03-13 /pmc/articles/PMC7480702/ http://dx.doi.org/10.1007/978-3-030-45237-7_5 Text en © The Author(s) 2020 Open Access This chapter is licensed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license and indicate if changes were made. The images or other third party material in this chapter are included in the chapter's Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the chapter's Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder.
spellingShingle Article
Giacobbe, Mirco
Henzinger, Thomas A.
Lechner, Mathias
How Many Bits Does it Take to Quantize Your Neural Network?
title How Many Bits Does it Take to Quantize Your Neural Network?
title_full How Many Bits Does it Take to Quantize Your Neural Network?
title_fullStr How Many Bits Does it Take to Quantize Your Neural Network?
title_full_unstemmed How Many Bits Does it Take to Quantize Your Neural Network?
title_short How Many Bits Does it Take to Quantize Your Neural Network?
title_sort how many bits does it take to quantize your neural network?
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7480702/
http://dx.doi.org/10.1007/978-3-030-45237-7_5
work_keys_str_mv AT giacobbemirco howmanybitsdoesittaketoquantizeyourneuralnetwork
AT henzingerthomasa howmanybitsdoesittaketoquantizeyourneuralnetwork
AT lechnermathias howmanybitsdoesittaketoquantizeyourneuralnetwork