Cargando…
A Hardware-Friendly Low-Bit Power-of-Two Quantization Method for CNNs and Its FPGA Implementation
To address the problems of convolutional neural networks (CNNs) consuming more hardware resources (such as DSPs and RAMs on FPGAs) and their accuracy, efficiency, and resources being difficult to balance, meaning they cannot meet the requirements of industrial applications, we proposed an innovative...
Autores principales: | , , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9460272/ https://www.ncbi.nlm.nih.gov/pubmed/36081072 http://dx.doi.org/10.3390/s22176618 |
_version_ | 1784786706569887744 |
---|---|
author | Sui, Xuefu Lv, Qunbo Bai, Yang Zhu, Baoyu Zhi, Liangjie Yang, Yuanbo Tan, Zheng |
author_facet | Sui, Xuefu Lv, Qunbo Bai, Yang Zhu, Baoyu Zhi, Liangjie Yang, Yuanbo Tan, Zheng |
author_sort | Sui, Xuefu |
collection | PubMed |
description | To address the problems of convolutional neural networks (CNNs) consuming more hardware resources (such as DSPs and RAMs on FPGAs) and their accuracy, efficiency, and resources being difficult to balance, meaning they cannot meet the requirements of industrial applications, we proposed an innovative low-bit power-of-two quantization method: the global sign-based network quantization (GSNQ). This method involves designing different quantization ranges according to the sign of the weights, which can provide a larger quantization-value range. Combined with the fine-grained and multi-scale global retraining method proposed in this paper, the accuracy loss of low-bit quantization can be effectively reduced. We also proposed a novel convolutional algorithm using shift operations to replace multiplication to help to deploy the GSNQ quantized models on FPGAs. Quantization comparison experiments performed on LeNet-5, AlexNet, VGG-Net, ResNet, and GoogLeNet showed that GSNQ has higher accuracy than most existing methods and achieves “lossless” quantization (i.e., the accuracy of the quantized CNN model is higher than the baseline) at low-bit quantization in most cases. FPGA comparison experiments showed that our convolutional algorithm does not occupy on-chip DSPs, and it also has a low comprehensive occupancy in terms of on-chip LUTs and FFs, which can effectively improve the computational parallelism, and this proves that GSNQ has good hardware-adaptation capability. This study provides theoretical and experimental support for the industrial application of CNNs. |
format | Online Article Text |
id | pubmed-9460272 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-94602722022-09-10 A Hardware-Friendly Low-Bit Power-of-Two Quantization Method for CNNs and Its FPGA Implementation Sui, Xuefu Lv, Qunbo Bai, Yang Zhu, Baoyu Zhi, Liangjie Yang, Yuanbo Tan, Zheng Sensors (Basel) Article To address the problems of convolutional neural networks (CNNs) consuming more hardware resources (such as DSPs and RAMs on FPGAs) and their accuracy, efficiency, and resources being difficult to balance, meaning they cannot meet the requirements of industrial applications, we proposed an innovative low-bit power-of-two quantization method: the global sign-based network quantization (GSNQ). This method involves designing different quantization ranges according to the sign of the weights, which can provide a larger quantization-value range. Combined with the fine-grained and multi-scale global retraining method proposed in this paper, the accuracy loss of low-bit quantization can be effectively reduced. We also proposed a novel convolutional algorithm using shift operations to replace multiplication to help to deploy the GSNQ quantized models on FPGAs. Quantization comparison experiments performed on LeNet-5, AlexNet, VGG-Net, ResNet, and GoogLeNet showed that GSNQ has higher accuracy than most existing methods and achieves “lossless” quantization (i.e., the accuracy of the quantized CNN model is higher than the baseline) at low-bit quantization in most cases. FPGA comparison experiments showed that our convolutional algorithm does not occupy on-chip DSPs, and it also has a low comprehensive occupancy in terms of on-chip LUTs and FFs, which can effectively improve the computational parallelism, and this proves that GSNQ has good hardware-adaptation capability. This study provides theoretical and experimental support for the industrial application of CNNs. MDPI 2022-09-01 /pmc/articles/PMC9460272/ /pubmed/36081072 http://dx.doi.org/10.3390/s22176618 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Sui, Xuefu Lv, Qunbo Bai, Yang Zhu, Baoyu Zhi, Liangjie Yang, Yuanbo Tan, Zheng A Hardware-Friendly Low-Bit Power-of-Two Quantization Method for CNNs and Its FPGA Implementation |
title | A Hardware-Friendly Low-Bit Power-of-Two Quantization Method for CNNs and Its FPGA Implementation |
title_full | A Hardware-Friendly Low-Bit Power-of-Two Quantization Method for CNNs and Its FPGA Implementation |
title_fullStr | A Hardware-Friendly Low-Bit Power-of-Two Quantization Method for CNNs and Its FPGA Implementation |
title_full_unstemmed | A Hardware-Friendly Low-Bit Power-of-Two Quantization Method for CNNs and Its FPGA Implementation |
title_short | A Hardware-Friendly Low-Bit Power-of-Two Quantization Method for CNNs and Its FPGA Implementation |
title_sort | hardware-friendly low-bit power-of-two quantization method for cnns and its fpga implementation |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9460272/ https://www.ncbi.nlm.nih.gov/pubmed/36081072 http://dx.doi.org/10.3390/s22176618 |
work_keys_str_mv | AT suixuefu ahardwarefriendlylowbitpoweroftwoquantizationmethodforcnnsanditsfpgaimplementation AT lvqunbo ahardwarefriendlylowbitpoweroftwoquantizationmethodforcnnsanditsfpgaimplementation AT baiyang ahardwarefriendlylowbitpoweroftwoquantizationmethodforcnnsanditsfpgaimplementation AT zhubaoyu ahardwarefriendlylowbitpoweroftwoquantizationmethodforcnnsanditsfpgaimplementation AT zhiliangjie ahardwarefriendlylowbitpoweroftwoquantizationmethodforcnnsanditsfpgaimplementation AT yangyuanbo ahardwarefriendlylowbitpoweroftwoquantizationmethodforcnnsanditsfpgaimplementation AT tanzheng ahardwarefriendlylowbitpoweroftwoquantizationmethodforcnnsanditsfpgaimplementation AT suixuefu hardwarefriendlylowbitpoweroftwoquantizationmethodforcnnsanditsfpgaimplementation AT lvqunbo hardwarefriendlylowbitpoweroftwoquantizationmethodforcnnsanditsfpgaimplementation AT baiyang hardwarefriendlylowbitpoweroftwoquantizationmethodforcnnsanditsfpgaimplementation AT zhubaoyu hardwarefriendlylowbitpoweroftwoquantizationmethodforcnnsanditsfpgaimplementation AT zhiliangjie hardwarefriendlylowbitpoweroftwoquantizationmethodforcnnsanditsfpgaimplementation AT yangyuanbo hardwarefriendlylowbitpoweroftwoquantizationmethodforcnnsanditsfpgaimplementation AT tanzheng hardwarefriendlylowbitpoweroftwoquantizationmethodforcnnsanditsfpgaimplementation |