Cargando…

A Post-training Quantization Method for the Design of Fixed-Point-Based FPGA/ASIC Hardware Accelerators for LSTM/GRU Algorithms

Recurrent Neural Networks (RNNs) have become important tools for tasks such as speech recognition, text generation, or natural language processing. However, their inference may involve up to billions of operations and their large number of parameters leads to large storage size and runtime memory us...

Descripción completa

Detalles Bibliográficos
Autores principales: Rapuano, Emilio, Pacini, Tommaso, Fanucci, Luca
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Hindawi 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9117057/
https://www.ncbi.nlm.nih.gov/pubmed/35602644
http://dx.doi.org/10.1155/2022/9485933
_version_ 1784710246719029248
author Rapuano, Emilio
Pacini, Tommaso
Fanucci, Luca
author_facet Rapuano, Emilio
Pacini, Tommaso
Fanucci, Luca
author_sort Rapuano, Emilio
collection PubMed
description Recurrent Neural Networks (RNNs) have become important tools for tasks such as speech recognition, text generation, or natural language processing. However, their inference may involve up to billions of operations and their large number of parameters leads to large storage size and runtime memory usage. These reasons impede the adoption of these models in real-time, on-the-edge applications. Field-Programmable Gate Arrays (FPGAs) and Application-Specific Integrated Circuits (ASICs) have emerged as promising solutions for the hardware acceleration of these algorithms, thanks to their degree of customization of compute data paths and memory subsystems, which makes them take the maximum advantage from compression techniques for what concerns area, timing, and power consumption. In contrast to the extensive study in compression and quantization for plain feed forward neural networks in the literature, little attention has been paid to reducing the computational resource requirements of RNNs. This work proposes a new effective methodology for the post-training quantization of RNNs. In particular, we focus on the quantization of Long Short-Term Memory (LSTM) RNNs and Gated Recurrent Unit (GRU) RNNs. The proposed quantization strategy is meant to be a detailed guideline toward the design of custom hardware accelerators for LSTM/GRU-based algorithms to be implemented on FPGA or ASIC devices using fixed-point arithmetic only. We applied our methods to LSTM/GRU models pretrained on the IMDb sentiment classification dataset and Penn TreeBank language modelling dataset, thus comparing each quantized model to its floating-point counterpart. The results show the possibility to achieve up to 90% memory footprint reduction in both cases, obtaining less than 1% loss in accuracy and even a slight improvement in the Perplexity per word metric, respectively. The results are presented showing the various trade-offs between memory footprint reduction and accuracy changes, demonstrating the benefits of the proposed methodology even in comparison with other works from the literature.
format Online
Article
Text
id pubmed-9117057
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Hindawi
record_format MEDLINE/PubMed
spelling pubmed-91170572022-05-19 A Post-training Quantization Method for the Design of Fixed-Point-Based FPGA/ASIC Hardware Accelerators for LSTM/GRU Algorithms Rapuano, Emilio Pacini, Tommaso Fanucci, Luca Comput Intell Neurosci Research Article Recurrent Neural Networks (RNNs) have become important tools for tasks such as speech recognition, text generation, or natural language processing. However, their inference may involve up to billions of operations and their large number of parameters leads to large storage size and runtime memory usage. These reasons impede the adoption of these models in real-time, on-the-edge applications. Field-Programmable Gate Arrays (FPGAs) and Application-Specific Integrated Circuits (ASICs) have emerged as promising solutions for the hardware acceleration of these algorithms, thanks to their degree of customization of compute data paths and memory subsystems, which makes them take the maximum advantage from compression techniques for what concerns area, timing, and power consumption. In contrast to the extensive study in compression and quantization for plain feed forward neural networks in the literature, little attention has been paid to reducing the computational resource requirements of RNNs. This work proposes a new effective methodology for the post-training quantization of RNNs. In particular, we focus on the quantization of Long Short-Term Memory (LSTM) RNNs and Gated Recurrent Unit (GRU) RNNs. The proposed quantization strategy is meant to be a detailed guideline toward the design of custom hardware accelerators for LSTM/GRU-based algorithms to be implemented on FPGA or ASIC devices using fixed-point arithmetic only. We applied our methods to LSTM/GRU models pretrained on the IMDb sentiment classification dataset and Penn TreeBank language modelling dataset, thus comparing each quantized model to its floating-point counterpart. The results show the possibility to achieve up to 90% memory footprint reduction in both cases, obtaining less than 1% loss in accuracy and even a slight improvement in the Perplexity per word metric, respectively. The results are presented showing the various trade-offs between memory footprint reduction and accuracy changes, demonstrating the benefits of the proposed methodology even in comparison with other works from the literature. Hindawi 2022-05-11 /pmc/articles/PMC9117057/ /pubmed/35602644 http://dx.doi.org/10.1155/2022/9485933 Text en Copyright © 2022 Emilio Rapuano et al. https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
spellingShingle Research Article
Rapuano, Emilio
Pacini, Tommaso
Fanucci, Luca
A Post-training Quantization Method for the Design of Fixed-Point-Based FPGA/ASIC Hardware Accelerators for LSTM/GRU Algorithms
title A Post-training Quantization Method for the Design of Fixed-Point-Based FPGA/ASIC Hardware Accelerators for LSTM/GRU Algorithms
title_full A Post-training Quantization Method for the Design of Fixed-Point-Based FPGA/ASIC Hardware Accelerators for LSTM/GRU Algorithms
title_fullStr A Post-training Quantization Method for the Design of Fixed-Point-Based FPGA/ASIC Hardware Accelerators for LSTM/GRU Algorithms
title_full_unstemmed A Post-training Quantization Method for the Design of Fixed-Point-Based FPGA/ASIC Hardware Accelerators for LSTM/GRU Algorithms
title_short A Post-training Quantization Method for the Design of Fixed-Point-Based FPGA/ASIC Hardware Accelerators for LSTM/GRU Algorithms
title_sort post-training quantization method for the design of fixed-point-based fpga/asic hardware accelerators for lstm/gru algorithms
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9117057/
https://www.ncbi.nlm.nih.gov/pubmed/35602644
http://dx.doi.org/10.1155/2022/9485933
work_keys_str_mv AT rapuanoemilio aposttrainingquantizationmethodforthedesignoffixedpointbasedfpgaasichardwareacceleratorsforlstmgrualgorithms
AT pacinitommaso aposttrainingquantizationmethodforthedesignoffixedpointbasedfpgaasichardwareacceleratorsforlstmgrualgorithms
AT fanucciluca aposttrainingquantizationmethodforthedesignoffixedpointbasedfpgaasichardwareacceleratorsforlstmgrualgorithms
AT rapuanoemilio posttrainingquantizationmethodforthedesignoffixedpointbasedfpgaasichardwareacceleratorsforlstmgrualgorithms
AT pacinitommaso posttrainingquantizationmethodforthedesignoffixedpointbasedfpgaasichardwareacceleratorsforlstmgrualgorithms
AT fanucciluca posttrainingquantizationmethodforthedesignoffixedpointbasedfpgaasichardwareacceleratorsforlstmgrualgorithms