Issue |
EPJ Web Conf.
Volume 245, 2020
24th International Conference on Computing in High Energy and Nuclear Physics (CHEP 2019)
|
|
---|---|---|
Article Number | 09014 | |
Number of page(s) | 10 | |
Section | 9 - Exascale Science | |
DOI | https://doi.org/10.1051/epjconf/202024509014 | |
Published online | 16 November 2020 |
https://doi.org/10.1051/epjconf/202024509014
Accelerate Scientific Deep Learning Models on Heterogeneous Computing Platform with FPGA
1
SHREC: NSF Center for Space, High-Performance, and Resilient Computing, University of Florida
2
CERN openlab
3
NVIDIA
4
National Energy Research Scientific Computing Center
5
Dell EMC
* e-mail: jc19chaoj@ufl.edu
** e-mail: davido@ufl.edu
*** e-mail: sofia.vallecorsa@cern.ch
**** e-mail: prabhat@lbl.gov
† e-mail: bhavesh.a.patel@dell.com
‡ e-mail: hlam@ufl.edu
Published online: 16 November 2020
AI and deep learning are experiencing explosive growth in almost every domain involving analysis of big data. Deep learning using Deep Neural Networks (DNNs) has shown great promise for such scientific data analysis applications. However, traditional CPU-based sequential computing without special instructions can no longer meet the requirements of mission-critical applications, which are compute-intensive and require low latency and high throughput. Heterogeneous computing (HGC), with CPUs integrated with GPUs, FPGAs, and other science-targeted accelerators, offers unique capabilities to accelerate DNNs. Collaborating researchers at SHREC1at the University of Florida, CERN Openlab, NERSC2at Lawrence Berkeley National Lab, Dell EMC, and Intel are studying the application of heterogeneous computing (HGC) to scientific problems using DNN models. This paper focuses on the use of FPGAs to accelerate the inferencing stage of the HGC workflow. We present case studies and results in inferencing state-of-the-art DNN models for scientific data analysis, using Intel distribution of OpenVINO, running on an Intel Programmable Acceleration Card (PAC) equipped with an Arria 10 GX FPGA. Using the Intel Deep Learning Acceleration (DLA) development suite to optimize existing FPGA primitives and develop new ones, we were able accelerate the scientific DNN models under study with a speedup from 2.46x to 9.59x for a single Arria 10 FPGA against a single core (single thread) of a server-class Skylake CPU.
© The Authors, published by EDP Sciences, 2020
This is an Open Access article distributed under the terms of the Creative Commons Attribution License 4.0, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
Current usage metrics show cumulative count of Article Views (full-text article views including HTML views, PDF and ePub downloads, according to the available data) and Abstracts Views on Vision4Press platform.
Data correspond to usage on the plateform after 2015. The current usage metrics is available 48-96 hours after online publication and is updated daily on week days.
Initial download of the metrics may take a while.