Cargando…
Efficient Layer-Wise N:M Sparse CNN Accelerator with Flexible SPEC: Sparse Processing Element Clusters
Recently, the layer-wise N:M fine-grained sparse neural network algorithm (i.e., every M-weights contains N non-zero values) has attracted tremendous attention, as it can effectively reduce the computational complexity with negligible accuracy loss. However, the speed-up potential of this algorithm...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10057003/ https://www.ncbi.nlm.nih.gov/pubmed/36984936 http://dx.doi.org/10.3390/mi14030528 |
_version_ | 1785016258848096256 |
---|---|
author | Xie, Xiaoru Zhu, Mingyu Lu, Siyuan Wang, Zhongfeng |
author_facet | Xie, Xiaoru Zhu, Mingyu Lu, Siyuan Wang, Zhongfeng |
author_sort | Xie, Xiaoru |
collection | PubMed |
description | Recently, the layer-wise N:M fine-grained sparse neural network algorithm (i.e., every M-weights contains N non-zero values) has attracted tremendous attention, as it can effectively reduce the computational complexity with negligible accuracy loss. However, the speed-up potential of this algorithm will not be fully exploited if the right hardware support is lacking. In this work, we design an efficient accelerator for the N:M sparse convolutional neural networks (CNNs) with layer-wise sparse patterns. First, we analyze the performances of different processing element (PE) structures and extensions to construct the flexible PE architecture. Second, the variable sparse convolutional dimensions and sparse ratios are involved in the hardware design. With a sparse PE cluster (SPEC) design, the hardware can efficiently accelerate CNNs with the layer-wise N:M pattern. Finally, we employ the proposed SPEC into the CNN accelerator with flexible network-on-chip and specially designed dataflow. We implement hardware accelerators on Xilinx ZCU102 FPGA and Xilinx VCU118 FPGA and evaluate them with classical CNNs such as Alexnet, VGG-16, and ResNet-50. Compared with existing accelerators designed for structured and unstructured pruned networks, our design achieves the best performance in terms of power efficiency. |
format | Online Article Text |
id | pubmed-10057003 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-100570032023-03-30 Efficient Layer-Wise N:M Sparse CNN Accelerator with Flexible SPEC: Sparse Processing Element Clusters Xie, Xiaoru Zhu, Mingyu Lu, Siyuan Wang, Zhongfeng Micromachines (Basel) Article Recently, the layer-wise N:M fine-grained sparse neural network algorithm (i.e., every M-weights contains N non-zero values) has attracted tremendous attention, as it can effectively reduce the computational complexity with negligible accuracy loss. However, the speed-up potential of this algorithm will not be fully exploited if the right hardware support is lacking. In this work, we design an efficient accelerator for the N:M sparse convolutional neural networks (CNNs) with layer-wise sparse patterns. First, we analyze the performances of different processing element (PE) structures and extensions to construct the flexible PE architecture. Second, the variable sparse convolutional dimensions and sparse ratios are involved in the hardware design. With a sparse PE cluster (SPEC) design, the hardware can efficiently accelerate CNNs with the layer-wise N:M pattern. Finally, we employ the proposed SPEC into the CNN accelerator with flexible network-on-chip and specially designed dataflow. We implement hardware accelerators on Xilinx ZCU102 FPGA and Xilinx VCU118 FPGA and evaluate them with classical CNNs such as Alexnet, VGG-16, and ResNet-50. Compared with existing accelerators designed for structured and unstructured pruned networks, our design achieves the best performance in terms of power efficiency. MDPI 2023-02-24 /pmc/articles/PMC10057003/ /pubmed/36984936 http://dx.doi.org/10.3390/mi14030528 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Xie, Xiaoru Zhu, Mingyu Lu, Siyuan Wang, Zhongfeng Efficient Layer-Wise N:M Sparse CNN Accelerator with Flexible SPEC: Sparse Processing Element Clusters |
title | Efficient Layer-Wise N:M Sparse CNN Accelerator with Flexible SPEC: Sparse Processing Element Clusters |
title_full | Efficient Layer-Wise N:M Sparse CNN Accelerator with Flexible SPEC: Sparse Processing Element Clusters |
title_fullStr | Efficient Layer-Wise N:M Sparse CNN Accelerator with Flexible SPEC: Sparse Processing Element Clusters |
title_full_unstemmed | Efficient Layer-Wise N:M Sparse CNN Accelerator with Flexible SPEC: Sparse Processing Element Clusters |
title_short | Efficient Layer-Wise N:M Sparse CNN Accelerator with Flexible SPEC: Sparse Processing Element Clusters |
title_sort | efficient layer-wise n:m sparse cnn accelerator with flexible spec: sparse processing element clusters |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10057003/ https://www.ncbi.nlm.nih.gov/pubmed/36984936 http://dx.doi.org/10.3390/mi14030528 |
work_keys_str_mv | AT xiexiaoru efficientlayerwisenmsparsecnnacceleratorwithflexiblespecsparseprocessingelementclusters AT zhumingyu efficientlayerwisenmsparsecnnacceleratorwithflexiblespecsparseprocessingelementclusters AT lusiyuan efficientlayerwisenmsparsecnnacceleratorwithflexiblespecsparseprocessingelementclusters AT wangzhongfeng efficientlayerwisenmsparsecnnacceleratorwithflexiblespecsparseprocessingelementclusters |