Cargando…

Designing Deep Learning Hardware Accelerator and Efficiency Evaluation

With the swift development of deep learning applications, the convolutional neural network (CNN) has brought a tremendous challenge to traditional processors to fulfil computing requirements. It is urgent to embrace new strategies to improve efficiency and diminish energy consumption. Currently, div...

Descripción completa

Detalles Bibliográficos
Autores principales: Qi, Zhi, Chen, Weijian, Naqvi, Rizwan Ali, Siddique, Kamran
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Hindawi 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9300348/
https://www.ncbi.nlm.nih.gov/pubmed/35875766
http://dx.doi.org/10.1155/2022/1291103
_version_ 1784751192736268288
author Qi, Zhi
Chen, Weijian
Naqvi, Rizwan Ali
Siddique, Kamran
author_facet Qi, Zhi
Chen, Weijian
Naqvi, Rizwan Ali
Siddique, Kamran
author_sort Qi, Zhi
collection PubMed
description With the swift development of deep learning applications, the convolutional neural network (CNN) has brought a tremendous challenge to traditional processors to fulfil computing requirements. It is urgent to embrace new strategies to improve efficiency and diminish energy consumption. Currently, diverse accelerator strategies for CNN computation based on the field-programmable gate array (FPGA) platform have been gradually explored because they have edges of high parallelism, low power consumption, and better programmability. This paper first illustrates state-of-the-art FPGA-based accelerator design by emphasizing the contributions and limitations of existing research works. Subsequently, we demonstrated significant concepts of parallel computing (PC) in the convolution algorithm and discussed how to accomplish parallelism based on the FPGA hardware structure. Eventually, with the proposed CPU+ FPGA framework, we performed experiments and compared the performance against traditional computation strategies in terms of the operation efficiency and energy consumption ratio. The results revealed that the efficiency of the FPGA platform is much higher than that of the central processing unit and graphics processing unit.
format Online
Article
Text
id pubmed-9300348
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Hindawi
record_format MEDLINE/PubMed
spelling pubmed-93003482022-07-21 Designing Deep Learning Hardware Accelerator and Efficiency Evaluation Qi, Zhi Chen, Weijian Naqvi, Rizwan Ali Siddique, Kamran Comput Intell Neurosci Research Article With the swift development of deep learning applications, the convolutional neural network (CNN) has brought a tremendous challenge to traditional processors to fulfil computing requirements. It is urgent to embrace new strategies to improve efficiency and diminish energy consumption. Currently, diverse accelerator strategies for CNN computation based on the field-programmable gate array (FPGA) platform have been gradually explored because they have edges of high parallelism, low power consumption, and better programmability. This paper first illustrates state-of-the-art FPGA-based accelerator design by emphasizing the contributions and limitations of existing research works. Subsequently, we demonstrated significant concepts of parallel computing (PC) in the convolution algorithm and discussed how to accomplish parallelism based on the FPGA hardware structure. Eventually, with the proposed CPU+ FPGA framework, we performed experiments and compared the performance against traditional computation strategies in terms of the operation efficiency and energy consumption ratio. The results revealed that the efficiency of the FPGA platform is much higher than that of the central processing unit and graphics processing unit. Hindawi 2022-07-13 /pmc/articles/PMC9300348/ /pubmed/35875766 http://dx.doi.org/10.1155/2022/1291103 Text en Copyright © 2022 Zhi Qi et al. https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
spellingShingle Research Article
Qi, Zhi
Chen, Weijian
Naqvi, Rizwan Ali
Siddique, Kamran
Designing Deep Learning Hardware Accelerator and Efficiency Evaluation
title Designing Deep Learning Hardware Accelerator and Efficiency Evaluation
title_full Designing Deep Learning Hardware Accelerator and Efficiency Evaluation
title_fullStr Designing Deep Learning Hardware Accelerator and Efficiency Evaluation
title_full_unstemmed Designing Deep Learning Hardware Accelerator and Efficiency Evaluation
title_short Designing Deep Learning Hardware Accelerator and Efficiency Evaluation
title_sort designing deep learning hardware accelerator and efficiency evaluation
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9300348/
https://www.ncbi.nlm.nih.gov/pubmed/35875766
http://dx.doi.org/10.1155/2022/1291103
work_keys_str_mv AT qizhi designingdeeplearninghardwareacceleratorandefficiencyevaluation
AT chenweijian designingdeeplearninghardwareacceleratorandefficiencyevaluation
AT naqvirizwanali designingdeeplearninghardwareacceleratorandefficiencyevaluation
AT siddiquekamran designingdeeplearninghardwareacceleratorandefficiencyevaluation