Cargando…

Accelerate Scientific Deep Learning Models on Heterogeneous Computing Platform with FPGA

AI and deep learning are experiencing explosive growth in almost every domain involving analysis of big data. Deep learning using Deep Neural Networks (DNNs) has shown great promise for such scientific data analysis applications. However, traditional CPU-based sequential computing without special in...

Descripción completa

Detalles Bibliográficos
Autores principales: Jiang, Chao, Ojika, David, Vallecorsa, Sofia, Kurth, Thorsten, Prabhat, Patel, Bhavesh, Lam, Herman
Lenguaje:eng
Publicado: 2020
Materias:
Acceso en línea:https://dx.doi.org/10.1051/epjconf/202024509014
http://cds.cern.ch/record/2753441
_version_ 1780969444759044096
author Jiang, Chao
Ojika, David
Vallecorsa, Sofia
Kurth, Thorsten
Prabhat
Patel, Bhavesh
Lam, Herman
author_facet Jiang, Chao
Ojika, David
Vallecorsa, Sofia
Kurth, Thorsten
Prabhat
Patel, Bhavesh
Lam, Herman
author_sort Jiang, Chao
collection CERN
description AI and deep learning are experiencing explosive growth in almost every domain involving analysis of big data. Deep learning using Deep Neural Networks (DNNs) has shown great promise for such scientific data analysis applications. However, traditional CPU-based sequential computing without special instructions can no longer meet the requirements of mission-critical applications, which are compute-intensive and require low latency and high throughput. Heterogeneous computing (HGC), with CPUs integrated with GPUs, FPGAs, and other science-targeted accelerators, offers unique capabilities to accelerate DNNs. Collaborating researchers at SHREC1at the University of Florida, CERN Openlab, NERSC2at Lawrence Berkeley National Lab, Dell EMC, and Intel are studying the application of heterogeneous computing (HGC) to scientific problems using DNN models. This paper focuses on the use of FPGAs to accelerate the inferencing stage of the HGC workflow. We present case studies and results in inferencing state-of-the-art DNN models for scientific data analysis, using Intel distribution of OpenVINO, running on an Intel Programmable Acceleration Card (PAC) equipped with an Arria 10 GX FPGA. Using the Intel Deep Learning Acceleration (DLA) development suite to optimize existing FPGA primitives and develop new ones, we were able accelerate the scientific DNN models under study with a speedup from 2.46x to 9.59x for a single Arria 10 FPGA against a single core (single thread) of a server-class Skylake CPU.
id oai-inspirehep.net-1832149
institution Organización Europea para la Investigación Nuclear
language eng
publishDate 2020
record_format invenio
spelling oai-inspirehep.net-18321492021-03-04T08:10:02Zdoi:10.1051/epjconf/202024509014http://cds.cern.ch/record/2753441engJiang, ChaoOjika, DavidVallecorsa, SofiaKurth, ThorstenPrabhatPatel, BhaveshLam, HermanAccelerate Scientific Deep Learning Models on Heterogeneous Computing Platform with FPGAComputing and ComputersAI and deep learning are experiencing explosive growth in almost every domain involving analysis of big data. Deep learning using Deep Neural Networks (DNNs) has shown great promise for such scientific data analysis applications. However, traditional CPU-based sequential computing without special instructions can no longer meet the requirements of mission-critical applications, which are compute-intensive and require low latency and high throughput. Heterogeneous computing (HGC), with CPUs integrated with GPUs, FPGAs, and other science-targeted accelerators, offers unique capabilities to accelerate DNNs. Collaborating researchers at SHREC1at the University of Florida, CERN Openlab, NERSC2at Lawrence Berkeley National Lab, Dell EMC, and Intel are studying the application of heterogeneous computing (HGC) to scientific problems using DNN models. This paper focuses on the use of FPGAs to accelerate the inferencing stage of the HGC workflow. We present case studies and results in inferencing state-of-the-art DNN models for scientific data analysis, using Intel distribution of OpenVINO, running on an Intel Programmable Acceleration Card (PAC) equipped with an Arria 10 GX FPGA. Using the Intel Deep Learning Acceleration (DLA) development suite to optimize existing FPGA primitives and develop new ones, we were able accelerate the scientific DNN models under study with a speedup from 2.46x to 9.59x for a single Arria 10 FPGA against a single core (single thread) of a server-class Skylake CPU.oai:inspirehep.net:18321492020
spellingShingle Computing and Computers
Jiang, Chao
Ojika, David
Vallecorsa, Sofia
Kurth, Thorsten
Prabhat
Patel, Bhavesh
Lam, Herman
Accelerate Scientific Deep Learning Models on Heterogeneous Computing Platform with FPGA
title Accelerate Scientific Deep Learning Models on Heterogeneous Computing Platform with FPGA
title_full Accelerate Scientific Deep Learning Models on Heterogeneous Computing Platform with FPGA
title_fullStr Accelerate Scientific Deep Learning Models on Heterogeneous Computing Platform with FPGA
title_full_unstemmed Accelerate Scientific Deep Learning Models on Heterogeneous Computing Platform with FPGA
title_short Accelerate Scientific Deep Learning Models on Heterogeneous Computing Platform with FPGA
title_sort accelerate scientific deep learning models on heterogeneous computing platform with fpga
topic Computing and Computers
url https://dx.doi.org/10.1051/epjconf/202024509014
http://cds.cern.ch/record/2753441
work_keys_str_mv AT jiangchao acceleratescientificdeeplearningmodelsonheterogeneouscomputingplatformwithfpga
AT ojikadavid acceleratescientificdeeplearningmodelsonheterogeneouscomputingplatformwithfpga
AT vallecorsasofia acceleratescientificdeeplearningmodelsonheterogeneouscomputingplatformwithfpga
AT kurththorsten acceleratescientificdeeplearningmodelsonheterogeneouscomputingplatformwithfpga
AT prabhat acceleratescientificdeeplearningmodelsonheterogeneouscomputingplatformwithfpga
AT patelbhavesh acceleratescientificdeeplearningmodelsonheterogeneouscomputingplatformwithfpga
AT lamherman acceleratescientificdeeplearningmodelsonheterogeneouscomputingplatformwithfpga