Cargando…

Resource-constrained FPGA/DNN co-design

Deep neural networks (DNNs) have demonstrated super performance in most learning tasks. However, a DNN typically contains a large number of parameters and operations, requiring a high-end processing platform for high-speed execution. To address this challenge, hardware-and-software co-design strateg...

Descripción completa

Detalles Bibliográficos
Autores principales: Zhang, Zhichao, Kouzani, Abbas Z.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Springer London 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8122185/
https://www.ncbi.nlm.nih.gov/pubmed/34025038
http://dx.doi.org/10.1007/s00521-021-06113-4
_version_ 1783692534352969728
author Zhang, Zhichao
Kouzani, Abbas Z.
author_facet Zhang, Zhichao
Kouzani, Abbas Z.
author_sort Zhang, Zhichao
collection PubMed
description Deep neural networks (DNNs) have demonstrated super performance in most learning tasks. However, a DNN typically contains a large number of parameters and operations, requiring a high-end processing platform for high-speed execution. To address this challenge, hardware-and-software co-design strategies, which involve joint DNN optimization and hardware implementation, can be applied. These strategies reduce the parameters and operations of the DNN, and fit it into a low-resource processing platform. In this paper, a DNN model is used for the analysis of the data captured using an electrochemical method to determine the concentration of a neurotransmitter and the recoding electrode. Next, a DNN miniaturization algorithm is introduced, involving combined pruning and compression, to reduce the DNN resource utilization. Here, the DNN is transformed to have sparse parameters by pruning a percentage of its weights. The Lempel–Ziv–Welch algorithm is then applied to compress the sparse DNN. Next, a DNN overlay is developed, combining the decompression of the DNN parameters and DNN inference, to allow the execution of the DNN on a FPGA on the PYNQ-Z2 board. This approach helps avoid the need for inclusion of a complex quantization algorithm. It compresses the DNN by a factor of 6.18, leading to about 50% reduction in the resource utilization on the FPGA.
format Online
Article
Text
id pubmed-8122185
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher Springer London
record_format MEDLINE/PubMed
spelling pubmed-81221852021-05-17 Resource-constrained FPGA/DNN co-design Zhang, Zhichao Kouzani, Abbas Z. Neural Comput Appl Original Article Deep neural networks (DNNs) have demonstrated super performance in most learning tasks. However, a DNN typically contains a large number of parameters and operations, requiring a high-end processing platform for high-speed execution. To address this challenge, hardware-and-software co-design strategies, which involve joint DNN optimization and hardware implementation, can be applied. These strategies reduce the parameters and operations of the DNN, and fit it into a low-resource processing platform. In this paper, a DNN model is used for the analysis of the data captured using an electrochemical method to determine the concentration of a neurotransmitter and the recoding electrode. Next, a DNN miniaturization algorithm is introduced, involving combined pruning and compression, to reduce the DNN resource utilization. Here, the DNN is transformed to have sparse parameters by pruning a percentage of its weights. The Lempel–Ziv–Welch algorithm is then applied to compress the sparse DNN. Next, a DNN overlay is developed, combining the decompression of the DNN parameters and DNN inference, to allow the execution of the DNN on a FPGA on the PYNQ-Z2 board. This approach helps avoid the need for inclusion of a complex quantization algorithm. It compresses the DNN by a factor of 6.18, leading to about 50% reduction in the resource utilization on the FPGA. Springer London 2021-05-15 2021 /pmc/articles/PMC8122185/ /pubmed/34025038 http://dx.doi.org/10.1007/s00521-021-06113-4 Text en © The Author(s), under exclusive licence to Springer-Verlag London Ltd., part of Springer Nature 2021 This article is made available via the PMC Open Access Subset for unrestricted research re-use and secondary analysis in any form or by any means with acknowledgement of the original source. These permissions are granted for the duration of the World Health Organization (WHO) declaration of COVID-19 as a global pandemic.
spellingShingle Original Article
Zhang, Zhichao
Kouzani, Abbas Z.
Resource-constrained FPGA/DNN co-design
title Resource-constrained FPGA/DNN co-design
title_full Resource-constrained FPGA/DNN co-design
title_fullStr Resource-constrained FPGA/DNN co-design
title_full_unstemmed Resource-constrained FPGA/DNN co-design
title_short Resource-constrained FPGA/DNN co-design
title_sort resource-constrained fpga/dnn co-design
topic Original Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8122185/
https://www.ncbi.nlm.nih.gov/pubmed/34025038
http://dx.doi.org/10.1007/s00521-021-06113-4
work_keys_str_mv AT zhangzhichao resourceconstrainedfpgadnncodesign
AT kouzaniabbasz resourceconstrainedfpgadnncodesign