ϟ
 
DOI: 10.1088/2632-2153/acc0d7
¤ OpenAccess: Gold
This work has “Gold” OA status. This means it is published in an Open Access journal that is indexed by the DOAJ.

Ultra-low latency recurrent neural network inference on FPGAs for physics applications with hls4ml

Elham E Khoda,Dylan Rankin,Rafael Teixeira de Lima,Philip Harris,Scott Hauck,Shih-Chieh Hsu,Michael Kagan,Vladimir Lončar,Chaitanya Paikara,R. Bhima Rao,Sioni Summers,Caterina Vernieri,Aaron Wang

Field-programmable gate array
Latency (audio)
Inference
2023
Abstract Recurrent neural networks have been shown to be effective architectures for many tasks in high energy physics, and thus have been widely adopted. Their use in low-latency environments has, however, been limited as a result of the difficulties of implementing recurrent architectures on field-programmable gate arrays (FPGAs). In this paper we present an implementation of two types of recurrent neural network layers—long short-term memory and gated recurrent unit—within the hls4ml framework. We demonstrate that our implementation is capable of producing effective designs for both small and large models, and can be customized to meet specific design requirements for inference latencies and FPGA resources. We show the performance and synthesized designs for multiple neural networks, many of which are trained specifically for jet identification tasks at the CERN Large Hadron Collider.
Loading...
    Cite this:
Generate Citation
Powered by Citationsy*
    Ultra-low latency recurrent neural network inference on FPGAs for physics applications with hls4ml” is a paper by Elham E Khoda Dylan Rankin Rafael Teixeira de Lima Philip Harris Scott Hauck Shih-Chieh Hsu Michael Kagan Vladimir Lončar Chaitanya Paikara R. Bhima Rao Sioni Summers Caterina Vernieri Aaron Wang published in 2023. It has an Open Access status of “gold”. You can read and download a PDF Full Text of this paper here.