Ultra-low latency recurrent neural network inference on FPGAs for physics applications with hls4ml

Khoda, Elham E and Rankin, Dylan and Teixeira de Lima, Rafael and Harris, Philip and Hauck, Scott and Hsu, Shih-Chieh and Kagan, Michael and Loncar, Vladimir and Paikara, Chaitanya and Rao, Richa and Summers, Sioni and Vernieri, Caterina and Wang, Aaron (2023) Ultra-low latency recurrent neural network inference on FPGAs for physics applications with hls4ml. Machine Learning: Science and Technology, 4 (2). 025004. ISSN 2632-2153

[thumbnail of Khoda_2023_Mach._Learn.__Sci._Technol._4_025004.pdf] Text
Khoda_2023_Mach._Learn.__Sci._Technol._4_025004.pdf - Published Version

Download (975kB)

Abstract

Recurrent neural networks have been shown to be effective architectures for many tasks in high energy physics, and thus have been widely adopted. Their use in low-latency environments has, however, been limited as a result of the difficulties of implementing recurrent architectures on field-programmable gate arrays (FPGAs). In this paper we present an implementation of two types of recurrent neural network layers—long short-term memory and gated recurrent unit—within the hls4ml framework. We demonstrate that our implementation is capable of producing effective designs for both small and large models, and can be customized to meet specific design requirements for inference latencies and FPGA resources. We show the performance and synthesized designs for multiple neural networks, many of which are trained specifically for jet identification tasks at the CERN Large Hadron Collider.

Item Type: Article
Subjects: East Asian Archive > Multidisciplinary
Depositing User: Unnamed user with email support@eastasianarchive.com
Date Deposited: 12 Jul 2023 12:57
Last Modified: 09 Sep 2025 03:48
URI: http://authors.go2articles.com/id/eprint/1254

Actions (login required)

View Item
View Item