Cargando…

Acceleration of Scientific Deep Learning Models on Heterogeneous Computing Platform with Intel FPGA

<!--HTML-->AI and deep learning have been widely used and shown great promise in recent scientific research activities. Deep neural network (DNN) models are proven to be highly efficient in big data analytic application for scientific experiments. However, traditional CPU-based sequential comp...

Descripción completa

Detalles Bibliográficos
Autor principal: Vallecorsa, Sofia
Lenguaje:eng
Publicado: 2019
Materias:
Acceso en línea:http://cds.cern.ch/record/2691452
_version_ 1780963856757030912
author Vallecorsa, Sofia
author_facet Vallecorsa, Sofia
author_sort Vallecorsa, Sofia
collection CERN
description <!--HTML-->AI and deep learning have been widely used and shown great promise in recent scientific research activities. Deep neural network (DNN) models are proven to be highly efficient in big data analytic application for scientific experiments. However, traditional CPU-based sequential computing can no longer meet the requirements of applications which are compute intensive and requiring low latency and high throughput. Heterogeneous computing (HGC), with CPUs integrated with accelerators such as GPUs and FPGAs, offers unique capabilities to accelerate DNNs. Collaborating researchers at SHREC at the University of Florida, NERSC at Lawrence Berkeley National Lab, CERN openlab, Dell EMC, and Intel are studying the application of HGC to scientific problems using DNN models. Our current work focuses on the use of FPGAs to accelerate the inferencing stage of the HGC workflow, using case studies of three state-of-the-art DNN models: HEP-CNN and CosmoGAN developed by NERSC, and 3DGAN developed by CERN openlab. Based on the Intel Deep Learning Acceleration (DLA) suite from Intel, we developed custom FPGA primitives and optimized the existing architecture for maximizing inferencing performance. Using Intel distribution of OpenVINO, we are able to accelerate the case study models running on an Intel Programmable Acceleration Card (PAC) equipped with an Arria 10 GX FPGA. In the ISC19 IXPUG Workshop, we presented our HGC framework and initial results for both HEP-CNN and CosmoGAN, using the native implementation of OpenVINO. With the help of the custom FPGA primitives in the DLA, we were able to improve the inferencing result for HEP-CNN and make a prediction of the optimal inference performance for CosmoGAN. We achieved a speedup from 3x to 6x for a single Arria 10 GX FPGA against a single core (single thread) of a server-class Intel Skylake CPU. For the IXPUG Annual Conference 2019, we will present our recent customization of the DLA architecture to implement the 3D convolution FPGA primitives for the 3DGAN model. We will also demonstrate additional improvements in the inference performance for HEP-CNN and CosmoGAN with the new DLA implementation. The details of our DLA customization, along with results in terms of comparison against the Skylake CPU, will be presented in this work.
id cern-2691452
institution Organización Europea para la Investigación Nuclear
language eng
publishDate 2019
record_format invenio
spelling cern-26914522022-11-02T22:24:40Zhttp://cds.cern.ch/record/2691452engVallecorsa, SofiaAcceleration of Scientific Deep Learning Models on Heterogeneous Computing Platform with Intel FPGAIXPUG 2019 Annual Conference at CERNother events or meetings<!--HTML-->AI and deep learning have been widely used and shown great promise in recent scientific research activities. Deep neural network (DNN) models are proven to be highly efficient in big data analytic application for scientific experiments. However, traditional CPU-based sequential computing can no longer meet the requirements of applications which are compute intensive and requiring low latency and high throughput. Heterogeneous computing (HGC), with CPUs integrated with accelerators such as GPUs and FPGAs, offers unique capabilities to accelerate DNNs. Collaborating researchers at SHREC at the University of Florida, NERSC at Lawrence Berkeley National Lab, CERN openlab, Dell EMC, and Intel are studying the application of HGC to scientific problems using DNN models. Our current work focuses on the use of FPGAs to accelerate the inferencing stage of the HGC workflow, using case studies of three state-of-the-art DNN models: HEP-CNN and CosmoGAN developed by NERSC, and 3DGAN developed by CERN openlab. Based on the Intel Deep Learning Acceleration (DLA) suite from Intel, we developed custom FPGA primitives and optimized the existing architecture for maximizing inferencing performance. Using Intel distribution of OpenVINO, we are able to accelerate the case study models running on an Intel Programmable Acceleration Card (PAC) equipped with an Arria 10 GX FPGA. In the ISC19 IXPUG Workshop, we presented our HGC framework and initial results for both HEP-CNN and CosmoGAN, using the native implementation of OpenVINO. With the help of the custom FPGA primitives in the DLA, we were able to improve the inferencing result for HEP-CNN and make a prediction of the optimal inference performance for CosmoGAN. We achieved a speedup from 3x to 6x for a single Arria 10 GX FPGA against a single core (single thread) of a server-class Intel Skylake CPU. For the IXPUG Annual Conference 2019, we will present our recent customization of the DLA architecture to implement the 3D convolution FPGA primitives for the 3DGAN model. We will also demonstrate additional improvements in the inference performance for HEP-CNN and CosmoGAN with the new DLA implementation. The details of our DLA customization, along with results in terms of comparison against the Skylake CPU, will be presented in this work.oai:cds.cern.ch:26914522019
spellingShingle other events or meetings
Vallecorsa, Sofia
Acceleration of Scientific Deep Learning Models on Heterogeneous Computing Platform with Intel FPGA
title Acceleration of Scientific Deep Learning Models on Heterogeneous Computing Platform with Intel FPGA
title_full Acceleration of Scientific Deep Learning Models on Heterogeneous Computing Platform with Intel FPGA
title_fullStr Acceleration of Scientific Deep Learning Models on Heterogeneous Computing Platform with Intel FPGA
title_full_unstemmed Acceleration of Scientific Deep Learning Models on Heterogeneous Computing Platform with Intel FPGA
title_short Acceleration of Scientific Deep Learning Models on Heterogeneous Computing Platform with Intel FPGA
title_sort acceleration of scientific deep learning models on heterogeneous computing platform with intel fpga
topic other events or meetings
url http://cds.cern.ch/record/2691452
work_keys_str_mv AT vallecorsasofia accelerationofscientificdeeplearningmodelsonheterogeneouscomputingplatformwithintelfpga
AT vallecorsasofia ixpug2019annualconferenceatcern