[institut] [phys4phys] SCL online seminar: Vladimir Loncar, Thursday, 11 February, 14:00

Antun Balaz antun at ipb.ac.rs
Fri Feb 5 09:23:54 CET 2021


Dear colleagues,

You are cordially invited to the SCL online seminar of the Center for the Study of Complex Systems, which will be held on Thursday, 11 February 2021 at 14:00 on Zoom (link is given below). The talk entitled

hls4ml: Fast inference of deep neural networks in FPGAs

will be given by Dr. Vladimir Lončar (Scientific Computing Laboratory, Center for the Study of Complex Systems, Institute of Physics Belgrade and CERN). Abstract of the talk:

With edge computing, real-time inference of deep neural networks (DNNs) on custom hardware has become increasingly relevant. Smartphone companies are incorporating Artificial Intelligence (AI) chips in their design for on-device inference to improve user experience and tighten data security, and the autonomous vehicle industry is turning to application-specific integrated circuits (ASICs) to keep the latency low. While the typical acceptable latency for real-time inference in applications like those above is O(1) ms, other applications require sub-microsecond inference. For instance, high-frequency trading machine learning (ML) algorithms are running on field-programmable gate arrays (FPGAs), highly accurate devices, to make decisions within nanoseconds. At the extreme inference spectrum end of both the low-latency (as in high-frequency trading) and limited-area (as in smartphone applications) is the processing of data from proton-proton collisions at the Large Hadron Collider (LHC) at CERN. Here, latencies of O(1) microsecond are required and resources are strictly limited. To address these challenges, we have developed hls4ml, an open-source library that converts pre-trained ML models into FPGA firmware, targeting extreme low-latency inference in order to stay within the strict constraints imposed by the CERN particle detectors.

In this talk, we will describe the essential features of the hls4ml workflow and network optimization techniques, including how to reduce the footprint of a machine learning model using state-of-the art techniques such as model pruning and quantization through quantization aware training.

*********************************************
Time: Feb 11, 2021 14:00 Belgrade
https://us02web.zoom.us/j/85795076891?pwd=a1dtQU9yK1QzVEVkV1JrVDlMQ1lYQT09

Meeting ID: 857 9507 6891
Passcode: 738237
*********************************************

Best regards,
Antun Balaž

-----
Dr. Antun Balaž
Research Professor

http://www.scl.rs/antun
Phone: +381 11 3713152
Fax: +381 11 3162190

Scientific Computing Laboratory
Center for the Study of Complex Systems

Institute of Physics Belgrade
National Institute of the Republic of Serbia
Pregrevica 118, 11080 Belgrade, Serbia
-----



More information about the institut mailing list