ϟ

J. Ngadiuba

Here are all the papers by J. Ngadiuba that you can download and read on OA.mg.
J. Ngadiuba’s last known institution is . Download J. Ngadiuba PDFs here.

Claim this Profile →
DOI: 10.1088/1748-0221/13/07/p07027
2018
Cited 269 times
Fast inference of deep neural networks in FPGAs for particle physics
Recent results at the Large Hadron Collider (LHC) have pointed to enhanced physics capabilities through the improvement of the real-time event processing techniques. Machine learning methods are ubiquitous and have proven to be very powerful in LHC physics, and particle physics as a whole. However, exploration of the use of such techniques in low-latency, low-power FPGA hardware has only just begun. FPGA-based trigger and data acquisition (DAQ) systems have extremely low, sub-microsecond latency requirements that are unique to particle physics. We present a case study for neural network inference in FPGAs focusing on a classifier for jet substructure which would enable, among many other physics scenarios, searches for new dark sector particles and novel measurements of the Higgs boson. While we focus on a specific example, the lessons are far-reaching. We develop a package based on High-Level Synthesis (HLS) called hls4ml to build machine learning models in FPGAs. The use of HLS increases accessibility across a broad user community and allows for a drastic decrease in firmware development time. We map out FPGA resource usage and latency versus neural network hyperparameters to identify the problems in particle physics that would benefit from performing neural network inference with FPGAs. For our example jet substructure model, we fit well within the available resources of modern FPGAs with a latency on the scale of 100 ns.
DOI: 10.1038/s42256-021-00356-5
2021
Cited 98 times
Automatic heterogeneous quantization of deep neural networks for low-latency inference on the edge for particle detectors
Although the quest for more accurate solutions is pushing deep learning research towards larger and more complex algorithms, edge devices demand efficient inference and therefore reduction in model size, latency and energy consumption. One technique to limit model size is quantization, which implies using fewer bits to represent weights and biases. Such an approach usually results in a decline in performance. Here, we introduce a method for designing optimally heterogeneously quantized versions of deep neural network models for minimum-energy, high-accuracy, nanosecond inference and fully automated deployment on chip. With a per-layer, per-parameter type automatic quantization procedure, sampling from a wide range of quantizers, model energy consumption and size are minimized while high accuracy is maintained. This is crucial for the event selection procedure in proton–proton collisions at the CERN Large Hadron Collider, where resources are strictly limited and a latency of $${\mathcal{O}}(1)\,\upmu{\rm{s}}$$ is required. Nanosecond inference and a resource consumption reduced by a factor of 50 when implemented on field-programmable gate array hardware are achieved. With edge computing on custom hardware, real-time inference with deep neural networks can reach the nanosecond timescale. An important application in this regime is event processing at particle collision detectors like those at the Large Hadron Collider (LHC). To ensure high performance as well as reduced resource consumption, a method is developed, and made available as an extension of the Keras library, to automatically design optimal quantization of the different layers in a deep neural network.
DOI: 10.21468/scipostphys.12.1.043
2022
Cited 50 times
The Dark Machines Anomaly Score Challenge: Benchmark Data and Model Independent Event Classification for the Large Hadron Collider
We describe the outcome of a data challenge conducted as part of the Dark Machines Initiative and the Les Houches 2019 workshop on Physics at TeV colliders. The challenged aims at detecting signals of new physics at the LHC using unsupervised machine learning algorithms. First, we propose how an anomaly score could be implemented to define model-independent signal regions in LHC searches. We define and describe a large benchmark dataset, consisting of >1 Billion simulated LHC events corresponding to $10~\rm{fb}^{-1}$ of proton-proton collisions at a center-of-mass energy of 13 TeV. We then review a wide range of anomaly detection and density estimation algorithms, developed in the context of the data challenge, and we measure their performance in a set of realistic analysis environments. We draw a number of useful conclusions that will aid the development of unsupervised new physics searches during the third run of the LHC, and provide our benchmark dataset for future studies at https://www.phenoMLdata.org. Code to reproduce the analysis is provided at https://github.com/bostdiek/DarkMachines-UnsupervisedChallenge.
DOI: 10.1088/2632-2153/aba042
2020
Cited 60 times
Compressing deep neural networks on FPGAs to binary and ternary precision with <tt>hls4ml</tt>
We present the implementation of binary and ternary neural networks in the hls4ml library, designed to automatically convert deep neural network models to digital circuits with field-programmable gate arrays (FPGA) firmware. Starting from benchmark models trained with floating point precision, we investigate different strategies to reduce the network's resource consumption by reducing the numerical precision of the network parameters to binary or ternary. We discuss the trade-off between model accuracy and resource consumption. In addition, we show how to balance between latency and accuracy by retaining full precision on a selected subset of network components. As an example, we consider two multiclass classification tasks: handwritten digit recognition with the MNIST data set and jet identification with simulated proton-proton collisions at the CERN Large Hadron Collider. The binary and ternary implementation has similar performance to the higher precision implementation while using drastically fewer FPGA resources.
DOI: 10.1088/2632-2153/ac0ea1
2021
Cited 53 times
Fast convolutional neural networks on FPGAs with hls4ml
Abstract We introduce an automated tool for deploying ultra low-latency, low-power deep neural networks with convolutional layers on field-programmable gate arrays (FPGAs). By extending the hls4ml library, we demonstrate an inference latency of 5 µ s using convolutional architectures, targeting microsecond latency applications like those at the CERN Large Hadron Collider. Considering benchmark models trained on the Street View House Numbers Dataset, we demonstrate various methods for model compression in order to fit the computational constraints of a typical FPGA device used in trigger and data acquisition systems of particle detectors. In particular, we discuss pruning and quantization-aware training, and demonstrate how resource utilization can be significantly reduced with little to no loss in model accuracy. We show that the FPGA critical resource consumption can be reduced by 97% with zero loss in model accuracy, and by 99% when tolerating a 6% accuracy degradation.
DOI: 10.3389/fdata.2020.598927
2021
Cited 41 times
Distance-Weighted Graph Neural Networks on FPGAs for Real-Time Particle Reconstruction in High Energy Physics
Graph neural networks have been shown to achieve excellent performance for several crucial tasks in particle physics, such as charged particle tracking, jet tagging, and clustering. An important domain for the application of these networks is the FGPA-based first layer of real-time data filtering at the CERN Large Hadron Collider, which has strict latency and resource constraints. We discuss how to design distance-weighted graph networks that can be executed with a latency of less than 1$\mu\mathrm{s}$ on an FPGA. To do so, we consider a representative task associated to particle reconstruction and identification in a next-generation calorimeter operating at a particle collider. We use a graph network architecture developed for such purposes, and apply additional simplifications to match the computing constraints of Level-1 trigger systems, including weight quantization. Using the $\mathtt{hls4ml}$ library, we convert the compressed models into firmware to be implemented on an FPGA. Performance of the synthesized models is presented both in terms of inference accuracy and resource usage.
DOI: 10.1007/s41781-019-0027-2
2019
Cited 43 times
FPGA-Accelerated Machine Learning Inference as a Service for Particle Physics Computing
Large-scale particle physics experiments face challenging demands for high-throughput computing resources both now and in the future. New heterogeneous computing paradigms on dedicated hardware with increased parallelization, such as Field Programmable Gate Arrays (FPGAs), offer exciting solutions with large potential gains. The growing applications of machine learning algorithms in particle physics for simulation, reconstruction, and analysis are naturally deployed on such platforms. We demonstrate that the acceleration of machine learning inference as a web service represents a heterogeneous computing solution for particle physics experiments that potentially requires minimal modification to the current computing model. As examples, we retrain the ResNet-50 convolutional neural network to demonstrate state-of-the-art performance for top quark jet tagging at the LHC and apply a ResNet-50 model with transfer learning for neutrino event classification. Using Project Brainwave by Microsoft to accelerate the ResNet-50 image classification model, we achieve average inference times of 60 (10) ms with our experimental physics software framework using Brainwave as a cloud (edge or on-premises) service, representing an improvement by a factor of approximately 30 (175) in model inference latency over traditional CPU inference in current experimental hardware. A single FPGA service accessed by many CPUs achieves a throughput of 600–700 inferences per second using an image batch of one, comparable to large batch-size GPU throughput and significantly better than small batch-size GPU throughput. Deployed as an edge or cloud service for the particle physics computing model, coprocessor accelerators can have a higher duty cycle and are potentially much more cost-effective.
DOI: 10.1088/1748-0221/15/05/p05026
2020
Cited 41 times
Fast inference of Boosted Decision Trees in FPGAs for particle physics
We describe the implementation of Boosted Decision Trees in the hls4ml library, which allows the translation of a trained model into FPGA firmware through an automated conversion process. Thanks to its fully on-chip implementation, hls4ml performs inference of Boosted Decision Tree models with extremely low latency. With a typical latency less than 100 ns, this solution is suitable for FPGA-based real-time processing, such as in the Level-1 Trigger system of a collider experiment. These developments open up prospects for physicists to deploy BDTs in FPGAs for identifying the origin of jets, better reconstructing the energies of muons, and enabling better selection of rare signal processes.
DOI: 10.3389/fdata.2022.787421
2022
Cited 23 times
Applications and Techniques for Fast Machine Learning in Science
In this community review report, we discuss applications and techniques for fast machine learning (ML) in science-the concept of integrating powerful ML methods into the real-time experimental data processing loop to accelerate scientific discovery. The material for the report builds on two workshops held by the Fast ML for Science community and covers three main areas: applications for fast ML across a number of scientific domains; techniques for training and implementing performant and resource-efficient ML algorithms; and computing architectures, platforms, and technologies for deploying these algorithms. We also present overlapping challenges across the multiple scientific domains where common solutions can be found. This community report is intended to give plenty of examples and inspiration for scientific discovery through integrated and accelerated ML solutions. This is followed by a high-level overview and organization of technical advances, including an abundance of pointers to source material, which can enable these breakthroughs.
DOI: 10.1038/s42256-022-00441-3
2022
Cited 22 times
Autoencoders on field-programmable gate arrays for real-time, unsupervised new physics detection at 40 MHz at the Large Hadron Collider
To study the physics of fundamental particles and their interactions, the Large Hadron Collider was constructed at CERN, where protons collide to create new particles measured by detectors. Collisions occur at a frequency of 40 MHz, and with an event size of roughly 1 MB it is impossible to read out and store the generated amount of data from the detector and therefore a multi-tiered, real-time filtering system is required. In this paper, we show how to adapt and deploy deep-learning-based autoencoders for the unsupervised detection of new physics signatures in the challenging environment of a real-time event selection system at the Large Hadron Collider. The first-stage filter, implemented on custom electronics, decides within a few microseconds whether an event should be kept or discarded. At this stage, the rate is reduced from 40 MHz to about 100 kHz. We demonstrate the deployment of an unsupervised selection algorithm on this custom electronics, running in as little as 80 ns and enhancing the signal-over-background ratio by three orders of magnitude. This work enables the practical deployment of these networks during the next data-taking campaign of the Large Hadron Collider. The Large Hadron Collider produces 40 million collision events per second, most of which need to be discarded by a real-time filtering system. Unsupervised deep learning algorithms are developed and deployed on custom electronics to search for rare events indicating new physics, rather than for specific events led by theory.
DOI: 10.1109/tns.2021.3087100
2021
Cited 27 times
A Reconfigurable Neural Network ASIC for Detector Front-End Data Compression at the HL-LHC
Despite advances in the programmable logic capabilities of modern trigger systems, a significant bottleneck remains in the amount of data to be transported from the detector to off-detector logic where trigger decisions are made. We demonstrate that a neural network autoencoder model can be implemented in a radiation tolerant ASIC to perform lossy data compression alleviating the data transmission problem while preserving critical information of the detector energy profile. For our application, we consider the high-granularity calorimeter from the CMS experiment at the CERN Large Hadron Collider. The advantage of the machine learning approach is in the flexibility and configurability of the algorithm. By changing the neural network weights, a unique data compression algorithm can be deployed for each sensor in different detector regions, and changing detector or collider conditions. To meet area, performance, and power constraints, we perform a quantization-aware training to create an optimized neural network hardware implementation. The design is achieved through the use of high-level synthesis tools and the hls4ml framework, and was processed through synthesis and physical layout flows based on a LP CMOS 65 nm technology node. The flow anticipates 200 Mrad of ionizing radiation to select gates, and reports a total area of 3.6 mm^2 and consumes 95 mW of power. The simulated energy consumption per inference is 2.4 nJ. This is the first radiation tolerant on-detector ASIC implementation of a neural network that has been designed for particle physics applications.
DOI: 10.3389/fdata.2022.803685
2022
Cited 15 times
Improving Variational Autoencoders for New Physics Detection at the LHC With Normalizing Flows
We investigate how to improve new physics detection strategies exploiting variational autoencoders and normalizing flows for anomaly detection at the Large Hadron Collider. As a working example, we consider the DarkMachines challenge dataset. We show how different design choices (e.g., event representations, anomaly score definitions, network architectures) affect the result on specific benchmark new physics models. Once a baseline is established, we discuss how to improve the anomaly detection accuracy by exploiting normalizing flow layers in the latent space of the variational autoencoder.
DOI: 10.1038/s41597-022-01187-8
2022
Cited 13 times
LHC physics dataset for unsupervised New Physics detection at 40 MHz
Abstract In the particle detectors at the Large Hadron Collider, hundreds of millions of proton-proton collisions are produced every second. If one could store the whole data stream produced in these collisions, tens of terabytes of data would be written to disk every second. The general-purpose experiments ATLAS and CMS reduce this overwhelming data volume to a sustainable level, by deciding in real-time whether each collision event should be kept for further analysis or be discarded. We introduce a dataset of proton collision events that emulates a typical data stream collected by such a real-time processing system, pre-filtered by requiring the presence of at least one electron or muon. This dataset could be used to develop novel event selection strategies and assess their sensitivity to new phenomena. In particular, we intend to stimulate a community-based effort towards the design of novel algorithms for performing unsupervised new physics detection, customized to fit the bandwidth, latency and computational resource constraints of the real-time event selection system of a typical particle detector.
2020
Cited 18 times
Automatic deep heterogeneous quantization of Deep Neural Networks for ultra low-area, low-latency inference on the edge at particle colliders
While the quest for more accurate solutions is pushing deep learning research towards larger and more complex algorithms, edge devices demand efficient inference i.e. reduction in model size, latency and energy consumption. A technique to limit model size is quantization, i.e. using fewer bits to represent weights and biases. Such an approach usually results in a decline in performance. Here, we introduce a novel method for designing optimally heterogeneously quantized versions of deep neural network models for minimum-energy, high-accuracy, nanosecond inference and fully automated deployment on chip. With a per-layer, per-parameter type automatic quantization procedure, sampling from a wide range of quantizers, model energy consumption and size are minimized while high accuracy is maintained. This is crucial for the event selection procedure in proton-proton collisions at the CERN Large Hadron Collider, where resources are strictly limited and a latency of ${\mathcal O}(1)~\mu$s is required. Nanosecond inference and a resource consumption reduced by a factor of $50$ when implemented on FPGA hardware is achieved.
2020
Cited 16 times
Ultra Low-latency, Low-area Inference Accelerators using Heterogeneous Deep Quantization with QKeras and hls4ml
In this paper, we introduce the QKeras library, an extension of the Keras library allowing for the creation of heterogeneously quantized versions of deep neural network models, through drop-in replacement of Keras layers. These models are trained quantization-aware, where the user can trade off model area or energy consumption by accuracy. We demonstrate how the reduction of numerical precision, through quantization-aware training, significantly reduces resource consumption while retaining high accuracy when implemented on FPGA hardware. Together with the hls4ml library, this allows for a fully automated deployment of quantized Keras models on chip, crucial for ultra low-latency inference. As a benchmark problem, we consider a classification task for the triggering of events in proton-proton collisions at the CERN Large Hadron Collider, where a latency of ${\mathcal O}(1)~\mu$s is required.
DOI: 10.48550/arxiv.2012.01563
2020
Cited 15 times
Accelerated Charged Particle Tracking with Graph Neural Networks on FPGAs
We develop and study FPGA implementations of algorithms for charged particle tracking based on graph neural networks. The two complementary FPGA designs are based on OpenCL, a framework for writing programs that execute across heterogeneous platforms, and hls4ml, a high-level-synthesis-based compiler for neural network to firmware conversion. We evaluate and compare the resource usage, latency, and tracking performance of our implementations based on a benchmark dataset. We find a considerable speedup over CPU-based execution is possible, potentially enabling such algorithms to be used effectively in future computing workflows and the FPGA-based Level-1 trigger at the CERN Large Hadron Collider.
DOI: 10.1109/asap52443.2021.00025
2021
Cited 13 times
Accelerating Recurrent Neural Networks for Gravitational Wave Experiments
This paper presents novel reconfigurable architectures for reducing the latency of recurrent neural networks (RNNs) that are used for detecting gravitational waves. Gravitational interferometers such as the LIGO detectors capture cosmic events such as black hole mergers which happen at unknown times and of varying durations, producing time-series data. We have developed a new architecture capable of accelerating RNN inference for analyzing time-series data from LIGO detectors. This architecture is based on optimizing the initiation intervals (II) in a multi-layer LSTM (Long Short-Term Memory) network, by identifying appropriate reuse factors for each layer. A customizable template for this architecture has been designed, which enables the generation of low-latency FPGA designs with efficient resource utilization using high-level synthesis tools. The proposed approach has been evaluated based on two LSTM models, targeting a ZYNQ 7045 FPGA and a U250 FPGA. Experimental results show that with balanced II, the number of DSPs can be reduced up to 42% while achieving the same IIs. When compared to other FPGA-based LSTM designs, our design can achieve about 4.92 to 12.4 times lower latency.
DOI: 10.1016/j.nima.2015.12.003
2016
Cited 16 times
The pixel tracking telescope at the Fermilab Test Beam Facility
An all silicon pixel telescope has been assembled and used at the Fermilab Test Beam Facility (FTBF) since 2009 to provide precise tracking information for different test beam experiments with a wide range of Detectors Under Test (DUTs) requiring high resolution measurement of the track impact point. The telescope is based on CMS pixel modules left over from the CMS forward pixel production. Eight planes are arranged to achieve a resolution of less than 8 μm on the 120 GeV proton beam transverse coordinate at the DUT position. In order to achieve such resolution with 100×150 μm2 pixel cells, the planes were tilted to 25 degrees to maximize charge sharing between pixels. Crucial for obtaining this performance is the alignment software, called Monicelli, specifically designed and optimized for this system. This paper will describe the telescope hardware, the data acquisition system and the alignment software constituting this particle tracking system for test beam users.
DOI: 10.48550/arxiv.2401.08777
2024
Robust Anomaly Detection for Particle Physics Using Multi-Background Representation Learning
Anomaly, or out-of-distribution, detection is a promising tool for aiding discoveries of new particles or processes in particle physics. In this work, we identify and address two overlooked opportunities to improve anomaly detection for high-energy physics. First, rather than train a generative model on the single most dominant background process, we build detection algorithms using representation learning from multiple background types, thus taking advantage of more information to improve estimation of what is relevant for detection. Second, we generalize decorrelation to the multi-background setting, thus directly enforcing a more complete definition of robustness for anomaly detection. We demonstrate the benefit of the proposed robust multi-background anomaly detection algorithms on a high-dimensional dataset of particle decays at the Large Hadron Collider.
DOI: 10.48550/arxiv.2402.01876
2024
Sets are all you need: Ultrafast jet classification on FPGAs for HL-LHC
We study various machine learning based algorithms for performing accurate jet flavor classification on field-programmable gate arrays and demonstrate how latency and resource consumption scale with the input size and choice of algorithm. These architectures provide an initial design for models that could be used for tagging at the CERN LHC during its high-luminosity phase. The high-luminosity upgrade will lead to a five-fold increase in its instantaneous luminosity for proton-proton collisions and, in turn, higher data volume and complexity, such as the availability of jet constituents. Through quantization-aware training and efficient hardware implementations, we show that O(100) ns inference of complex architectures such as deep sets and interaction networks is feasible at a low computational resource cost.
DOI: 10.48550/arxiv.2405.00645
2024
Gradient-based Automatic Per-Weight Mixed Precision Quantization for Neural Networks On-Chip
Model size and inference speed at deployment time, are major challenges in many deep learning applications. A promising strategy to overcome these challenges is quantization. However, a straightforward uniform quantization to very low precision can result in significant accuracy loss. Mixed-precision quantization, based on the idea that certain parts of the network can accommodate lower precision without compromising performance compared to other parts, offers a potential solution. In this work, we present High Granularity Quantization (HGQ), an innovative quantization-aware training method designed to fine-tune the per-weight and per-activation precision in an automatic way for ultra-low latency and low power neural networks which are to be deployed on FPGAs. We demonstrate that HGQ can outperform existing methods by a substantial margin, achieving resource reduction by up to a factor of 20 and latency improvement by a factor of 5 while preserving accuracy.
2021
Cited 8 times
arXiv : The Dark Machines Anomaly Score Challenge: Benchmark Data and Model Independent Event Classification for the Large Hadron Collider
We describe the outcome of a data challenge conducted as part of the Dark Machines Initiative and the Les Houches 2019 workshop on Physics at TeV colliders. The challenged aims at detecting signals of new physics at the LHC using unsupervised machine learning algorithms. First, we propose how an anomaly score could be implemented to define model-independent signal regions in LHC searches. We define and describe a large benchmark dataset, consisting of >1 Billion simulated LHC events corresponding to $10~\rm{fb}^{-1}$ of proton-proton collisions at a center-of-mass energy of 13 TeV. We then review a wide range of anomaly detection and density estimation algorithms, developed in the context of the data challenge, and we measure their performance in a set of realistic analysis environments. We draw a number of useful conclusions that will aid the development of unsupervised new physics searches during the third run of the LHC, and provide our benchmark dataset for future studies at this https URL. Code to reproduce the analysis is provided at this https URL.
DOI: 10.1016/j.nima.2012.10.011
2013
Cited 8 times
Test-beam studies of diamond sensors for SLHC
Abstract Diamond sensors are studied as an alternative to silicon sensors to withstand the high radiation doses that are expected in future upgrades of the pixel detectors for the SLHC. Diamond pixel sensors are intrinsically radiation hard and are considered as a possible solution for the innermost tracker layers close to the interaction point where current silicon sensors cannot cope with the harsh radiation environment.An effort to study possible candidates for the upgrades is undergoing using the Fermilab test-beam facility (FTBF), where diamonds and 3D silicon sensors have been studied. Using a CMS pixel-based telescope built and installed at the FTBF, we are studying charge collection efficiencies for un-irradiated and irradiated devices bump-bonded to the CMS PSI46 pixel readout chip. A description of the test-beam effort and preliminary results on diamond sensors will be presented.
DOI: 10.1088/2632-2153/ac9cb5
2022
Cited 3 times
Real-time semantic segmentation on FPGAs for autonomous vehicles with hls4ml
Abstract In this paper, we investigate how field programmable gate arrays can serve as hardware accelerators for real-time semantic segmentation tasks relevant for autonomous driving. Considering compressed versions of the ENet convolutional neural network architecture, we demonstrate a fully-on-chip deployment with a latency of 4.9 ms per image, using less than 30% of the available resources on a Xilinx ZCU102 evaluation board. The latency is reduced to 3 ms per image when increasing the batch size to ten, corresponding to the use case where the autonomous vehicle receives inputs from multiple cameras simultaneously. We show, through aggressive filter reduction and heterogeneous quantization-aware training, and an optimized implementation of convolutional layers, that the power consumption and resource utilization can be significantly reduced while maintaining accuracy on the Cityscapes dataset.
DOI: 10.1051/epjconf/202024506039
2020
Cited 5 times
New Physics Agnostic Selections For New Physics Searches
We discuss a model-independent strategy for boosting new physics searches with the help of an unsupervised anomaly detection algorithm. Prior to a search, each input event is preprocessed by the algorithm - a variational autoencoder (VAE). Based on the loss assigned to each event, input data can be split into a background control sample and a signal enriched sample. Following this strategy, one can enhance the sensitivity to new physics with no assumption on the underlying new physics signature. Our results show that a typical BSM search on the signal enriched group is more sensitive than an equivalent search on the original dataset.
DOI: 10.1016/j.nima.2014.06.029
2014
Cited 4 times
Pre- and post-irradiation performance of FBK 3D silicon pixel detectors for CMS
In preparation for the tenfold luminosity upgrade of the Large Hadron Collider (the HL-LHC) around 2020, three-dimensional (3D) silicon pixel sensors are being developed as a radiation-hard candidate to replace the planar ones currently being used in the CMS pixel detector. This study examines an early batch of FBK sensors (named ATLAS08) of three 3D pixel geometries: 1E, 2E, and 4E, which respectively contain one, two, and four readout electrodes for each pixel, passing completely through the bulk. We present electrical characteristics and beam test performance results for each detector before and after irradiation. The maximum fluence applied is 3.5×1015 n eq/cm2.
DOI: 10.1145/3289602.3293986
2019
Cited 4 times
Fast Inference of Deep Neural Networks for Real-time Particle Physics Applications
Machine learning methods are ubiquitous and have proven to be very powerful in LHC physics, and particle physics as a whole. However, exploration of such techniques in low-latency, low-power FPGA (Field Programmable Gate Array) hardware has only just begun. FPGA-based trigger and data acquisition systems have extremely low, sub-microsecond latency requirements that are unique to particle physics. We present a case study for neural network inference in FPGAs focusing on a classifier for jet substructure which would enable many new physics measurements. While we focus on a specific example, the lessons are far-reaching. A compiler package is developed based on High-Level Synthesis (HLS) called HLS4ML to build machine learning models in FPGAs. The use of HLS increases accessibility across a broad user community and allows for a drastic decrease in firmware development time. We map out FPGA resource usage and latency versus neural network hyperparameters to allow for directed resource tuning in the low latency environment and assess the impact on our benchmark Physics performance scenario For our example jet substructure model, we fit well within the available resources of modern FPGAs with latency on the scale of 100~ns.
DOI: 10.48550/arxiv.2008.13636
2020
Cited 4 times
HL-LHC Computing Review: Common Tools and Community Software
Common and community software packages, such as ROOT, Geant4 and event generators have been a key part of the LHC's success so far and continued development and optimisation will be critical in the future. The challenges are driven by an ambitious physics programme, notably the LHC accelerator upgrade to high-luminosity, HL-LHC, and the corresponding detector upgrades of ATLAS and CMS. In this document we address the issues for software that is used in multiple experiments (usually even more widely than ATLAS and CMS) and maintained by teams of developers who are either not linked to a particular experiment or who contribute to common software within the context of their experiment activity. We also give space to general considerations for future software and projects that tackle upcoming challenges, no matter who writes it, which is an area where community convergence on best practice is extremely useful.
DOI: 10.1016/j.nima.2013.04.048
2013
Cited 3 times
Performance of CMS 3D silicon pixel detectors before and after irradiation
Three-dimensional (3D) silicon detectors are emerging as one of the most promising technologies for the innermost layers of tracking devices for the foreseen upgrades of the LHC. 3D sensors compatible with the CMS readout, fabricated at FBK (Trento, Italy), were tested in the laboratory and with a 120 GeV/c proton beam at the FNAL test beam facility, before and after irradiation up to a fluence of 3.5×1015neq/cm2. Preliminary results of the data analysis are presented.
DOI: 10.1016/j.nima.2012.11.076
2013
Cited 3 times
3D-FBK pixel sensors with CMS readout: First test results
Abstract Silicon 3D detectors consist of an array of columnar electrodes of both doping types which penetrate entirely in the detector bulk, perpendicularly to the surface. They are emerging as one of the most promising technologies for innermost layers of tracking devices for the foreseen upgrades of the LHC. Until recently, properties of 3D sensors have been investigated mostly with ATLAS readout electronics. 3D pixel sensors compatible with the CMS readout were first fabricated at SINTEF (Oslo, Norway), and more recently at FBK (Trento, Italy) and CNM (Barcelona, Spain). Several sensors with different electrode configurations, bump-bonded with the CMS pixel PSI46 readout chip, were characterized in laboratory and tested at Fermilab with a proton beam of 120 GeV/ c . Preliminary results of the data analysis are presented.
2021
Cited 3 times
hls4ml: An Open-Source Codesign Workflow to Empower Scientific Low-Power Machine Learning Devices
Accessible machine learning algorithms, software, and diagnostic tools for energy-efficient devices and systems are extremely valuable across a broad range of application domains. In scientific domains, real-time near-sensor processing can drastically improve experimental design and accelerate scientific discoveries. To support domain scientists, we have developed hls4ml, an open-source software-hardware codesign workflow to interpret and translate machine learning algorithms for implementation with both FPGA and ASIC technologies. We expand on previous hls4ml work by extending capabilities and techniques towards low-power implementations and increased usability: new Python APIs, quantization-aware pruning, end-to-end FPGA workflows, long pipeline kernels for low power, and new device backends include an ASIC workflow. Taken together, these and continued efforts in hls4ml will arm a new generation of domain scientists with accessible, efficient, and powerful tools for machine-learning-accelerated discovery.
DOI: 10.1051/epjconf/202125104027
2021
Cited 3 times
Jet Single Shot Detection
We apply object detection techniques based on Convolutional Neural Networks to jet reconstruction and identification at the CERN Large Hadron Collider. In particular, we focus on CaloJet reconstruction, representing each event as an image composed of calorimeter cells and using a Single Shot Detection network, called Jet-SSD. The model performs simultaneous localization and classification and additional regression tasks to measure jet features. We investigate Ternary Weight Networks with weights constrained to {-1, 0, 1} times a layer- and channel-dependent scaling factors. We show that the quantized version of the network closely matches the performance of its full-precision equivalent.
DOI: 10.5281/zenodo.5046389
2021
Cited 3 times
Unsupervised New Physics detection at 40 MHz: Training Dataset
Unsupervised New Physics detection at 40 MHz data challenge Training dataset, consisting of a cocktail of Standard Model collision events (simulation of LHC 13 TeV proton-proton collisions) pre-filtered by a requirement of a muon or electron with 23 GeV transverse momentum. Data format description available on the data challenge web page: https://mpp-hep.github.io/ADC2021/
DOI: 10.1016/j.nima.2013.07.042
2013
Testbeam and laboratory test results of irradiated 3D CMS pixel detectors
The CMS silicon pixel detector is the tracking device closest to the LHC p–p collisions, which precisely reconstructs the charged particle trajectories. The planar technology used in the current innermost layer of the pixel detector will reach the design limit for radiation hardness at the end of Phase I upgrade and will need to be replaced before the Phase II upgrade in 2020. Due to its unprecedented performance in harsh radiation environments, 3D silicon technology is under consideration as a possible replacement of planar technology for the High Luminosity-LHC or HL-LHC. 3D silicon detectors are fabricated by the Deep Reactive-Ion-Etching (DRIE) technique which allows p- and n-type electrodes to be processed through the silicon substrate as opposed to being implanted through the silicon surface. The 3D CMS pixel devices presented in this paper were processed at FBK. They were bump bonded to the current CMS pixel readout chip, tested in the laboratory, and testbeams carried out at FNAL with the proton beam of 120 GeV/c. In this paper we present the laboratory and beam test results for the irradiated 3D CMS pixel devices.
DOI: 10.3389/fdata.2023.1301942
2023
Corrigendum: Applications and techniques for fast machine learning in science
[This corrects the article DOI: 10.3389/fdata.2022.787421.].
DOI: 10.48550/arxiv.2311.14160
2023
Efficient and Robust Jet Tagging at the LHC with Knowledge Distillation
The challenging environment of real-time data processing systems at the Large Hadron Collider (LHC) strictly limits the computational complexity of algorithms that can be deployed. For deep learning models, this implies that only models with low computational complexity that have weak inductive bias are feasible. To address this issue, we utilize knowledge distillation to leverage both the performance of large models and the reduced computational complexity of small ones. In this paper, we present an implementation of knowledge distillation, demonstrating an overall boost in the student models' performance for the task of classifying jets at the LHC. Furthermore, by using a teacher model with a strong inductive bias of Lorentz symmetry, we show that we can induce the same inductive bias in the student model which leads to better robustness against arbitrary Lorentz boost.
DOI: 10.48550/arxiv.2311.17162
2023
Fast Particle-based Anomaly Detection Algorithm with Variational Autoencoder
Model-agnostic anomaly detection is one of the promising approaches in the search for new beyond the standard model physics. In this paper, we present Set-VAE, a particle-based variational autoencoder (VAE) anomaly detection algorithm. We demonstrate a 2x signal efficiency gain compared with traditional subjettiness-based jet selection. Furthermore, with an eye to the future deployment to trigger systems, we propose the CLIP-VAE, which reduces the inference-time cost of anomaly detection by using the KL-divergence loss as the anomaly score, resulting in a 2x acceleration in latency and reducing the caching requirement.
DOI: 10.5281/zenodo.5055454
2021
Unsupervised New Physics detection at 40 MHz: LQ -&gt; b tau Signal Benchmark Dataset
Unsupervised New Physics detection at 40 MHz data challenge Signal Benchmark Dataset consisting of Leptoquarks -&gt; b tau decays produced in collision events (simulation of LHC 13 TeV proton-proton collisions) pre-filtered by a requirement of a muon or electron with 23 GeV transverse momentum. Data format description available on the data challenge web page: https://mpp-hep.github.io/ADC2021/
DOI: 10.1088/2632-2153/ac7a02
2022
Lightweight jet reconstruction and identification as an object detection task
Abstract We apply object detection techniques based on deep convolutional blocks to end-to-end jet identification and reconstruction tasks encountered at the CERN large hadron collider (LHC). Collision events produced at the LHC and represented as an image composed of calorimeter and tracker cells are given as an input to a Single Shot Detection network. The algorithm, named PFJet-SSD performs simultaneous localization, classification and regression tasks to cluster jets and reconstruct their features. This all-in-one single feed-forward pass gives advantages in terms of execution time and an improved accuracy w.r.t. traditional rule-based methods. A further gain is obtained from network slimming, homogeneous quantization, and optimized runtime for meeting memory and latency constraints of a typical real-time processing environment. We experiment with 8-bit and ternary quantization, benchmarking their accuracy and inference latency against a single-precision floating-point. We show that the ternary network closely matches the performance of its full-precision equivalent and outperforms the state-of-the-art rule-based algorithm. Finally, we report the inference latency on different hardware platforms and discuss future applications.
DOI: 10.3389/frai.2022.999173
2022
Editorial: Efficient AI in particle physics and astrophysics
EDITORIAL article Front. Artif. Intell., 30 September 2022Sec. Big Data and AI in High Energy Physics Volume 5 - 2022 | https://doi.org/10.3389/frai.2022.999173
DOI: 10.2172/1570210
2019
FPGAs as a Service to Accelerate Machine Learning Inference [PowerPoint]
Large-scale particle physics experiments face challenging demands for high-throughput computing resources both now and in the future. New heterogeneous computing paradigms on dedicated hardware with increased parallelization, such as Field Programmable Gate Arrays (FPGAs), offer exciting solutions with large potential gains. The growing applications of machine learning algorithms in particle physics for simulation, reconstruction, and analysis are naturally deployed on such platforms. We demonstrate that the acceleration of machine learning inference as a web service represents a heterogeneous computing solution for particle physics experiments that requires minimal modification to the current computing model. As examples, we retrain the ResNet-50 convolutional neural network to demonstrate state-of-the-art performance for top quark jet tagging at the LHC, and apply a ResNet-50 model with transfer learning for neutrino event classification. Using Project Brainwave b y Microsoft to accelerate the ResNet-50 image classification model, we achieve average inference times of 60 (10) milliseconds with our experimental physics software framework using Brainwave as a cloud (edge or on-premises) service, representing an improvement by a factor of approximately 30 (175) in model inference latency over traditional CPU inference in current experimental hardware. A single FPGA service accessed by many CPUs achieves a throughput of 600--700 inferences per second using an image batch of one, comparable to large batch-size GPU throughput and significantly better than small batch-size GPU throughput. Deployed as an edge or cloud service for the particle physics computing model, coprocessor accelerators can have a higher duty cycle and are potentially much more cost-effective.
DOI: 10.48550/arxiv.2002.02837
2020
Report on the ECFA Early-Career Researchers Debate on the 2020 European Strategy Update for Particle Physics
A group of Early-Career Researchers (ECRs) has been given a mandate from the European Committee for Future Accelerators (ECFA) to debate the topics of the current European Strategy Update (ESU) for Particle Physics and to summarise the outcome in a brief document [1]. A full-day debate with 180 delegates was held at CERN, followed by a survey collecting quantitative input. During the debate, the ECRs discussed future colliders in terms of the physics prospects, their implications for accelerator and detector technology as well as computing and software. The discussion was organised into several topic areas. From these areas two common themes were particularly highlighted by the ECRs: sociological and human aspects; and issues of the environmental impact and sustainability of our research.
DOI: 10.5281/zenodo.5046446
2021
Unsupervised New Physics detection at 40 MHz: A -&gt; 4 leptons Signal Benchmark Dataset
Unsupervised New Physics detection at 40 MHz data challenge Signal Benchmark Dataset consisting of A -&gt; 4 leptons decays produced in collision events (simulation of LHC 13 TeV proton-proton collisions) pre-filtered by a requirement of a muon or electron with 23 GeV transverse momentum. Data format description available on the data challenge web page: https://mpp-hep.github.io/ADC2021/
DOI: 10.5281/zenodo.5061633
2021
Unsupervised New Physics detection at 40 MHz: h^0 -&gt; tau tau Signal Benchmark Dataset
Unsupervised New Physics detection at 40 MHz data challenge Signal Benchmark Dataset consisting of h^0 -&gt; tau tau decays produced in collision events (simulation of LHC 13 TeV proton-proton collisions) pre-filtered by a requirement of a muon or electron with 23 GeV transverse momentum. Data format description available on the data challenge web page: https://mpp-hep.github.io/ADC2021/
DOI: 10.5281/zenodo.5061688
2021
Unsupervised New Physics detection at 40 MHz: h+ -&gt; tau nu Signal Benchmark Dataset
Unsupervised New Physics detection at 40 MHz data challenge Signal Benchmark Dataset consisting of h+ -&gt; tau nu decays produced in collision events (simulation of LHC 13 TeV proton-proton collisions) pre-filtered by a requirement of a muon or electron with 23 GeV transverse momentum. Data format description available on the data challenge web page: https://mpp-hep.github.io/ADC2021/
DOI: 10.1088/1748-0221/8/06/p06006
2013
Tracking performance of a single-crystal and a polycrystalline diamond pixel-detector
We present a comparative characterization of the performance of a single-crystal and a polycrystalline diamond pixel-detector employing the standard CMS pixel readout chips. Measurements were carried out at the Fermilab Test Beam Facility, FTBF, using protons of momentum 120 GeV/c tracked by a high-resolution pixel telescope. Particular attention was directed to the study of the charge-collection, the charge-sharing among adjacent pixels and the achievable position resolution. The performance of the single-crystal detector was excellent and comparable to the best available silicon pixel-detectors. The measured average detection-efficiency was near unity, ε = 0.99860±0.00006, and the position-resolution for shared hits was about 6 μm. On the other hand, the performance of the polycrystalline detector was hampered by its lower charge collection distance and the readout chip threshold. A new readout chip, capable of operating at much lower threshold (around 1 ke−), would be required to fully exploit the potential performance of the polycrystalline diamond pixel-detector.
2012
Test-beam studies of diamond sensors for SLHC
DOI: 10.5167/uzh-143399
2017
Corrigendum to: Search for dijet resonances in proton–proton collisions at $\sqrt{s} = 13$ TeV and constraints on dark matter and other models [Phys. Lett. B 769 (2017) 520–542]
A search is presented for narrow resonances decaying to dijet final states in proton–proton collisions at s√=13TeV using data corresponding to an integrated luminosity of 12.9 $fb{−1}$. The dijet mass spectrum is well described by a smooth parameterization and no significant evidence for the production of new particles is observed. Upper limits at 95% confidence level are reported on the production cross section for narrow resonances with masses above 0.6 TeV. In the context of specific models, the limits exclude string resonances with masses below 7.4 TeV, scalar diquarks below 6.9 TeV, axigluons and colorons below 5.5 TeV, excited quarks below 5.4 TeV, color-octet scalars below 3.0 TeV, W′ bosons below 2.7 TeV, Z′ bosons below 2.1 TeV and between 2.3 and 2.6 TeV, and RS gravitons below 1.9 TeV. These extend previous limits in the dijet channel. Vector and axial-vector mediators in a simplified model of interactions between quarks and dark matter are excluded below 2.0 TeV. The first limits in the dijet channel on dark matter mediators are presented as functions of dark matter mass and are compared to the exclusions of dark matter in direct detection experiments.
DOI: 10.5167/uzh-140761
2017
A search for new phenomena in pp collisions at $\sqrt {s} = 13$ TeV in final states with missing transverse momentum and at least one jet using the $α_{T}$ variable
DOI: 10.5167/uzh-142438
2017
Search for Diboson Resonances with CMS and Pixel Barrel Detector Calibration and Upgrade
DOI: 10.48550/arxiv.2202.04499
2022
Lightweight Jet Reconstruction and Identification as an Object Detection Task
We apply object detection techniques based on deep convolutional blocks to end-to-end jet identification and reconstruction tasks encountered at the CERN Large Hadron Collider (LHC). Collision events produced at the LHC and represented as an image composed of calorimeter and tracker cells are given as an input to a Single Shot Detection network. The algorithm, named PFJet-SSD performs simultaneous localization, classification and regression tasks to cluster jets and reconstruct their features. This all-in-one single feed-forward pass gives advantages in terms of execution time and an improved accuracy w.r.t. traditional rule-based methods. A further gain is obtained from network slimming, homogeneous quantization, and optimized runtime for meeting memory and latency constraints of a typical real-time processing environment. We experiment with 8-bit and ternary quantization, benchmarking their accuracy and inference latency against a single-precision floating-point. We show that the ternary network closely matches the performance of its full-precision equivalent and outperforms the state-of-the-art rule-based algorithm. Finally, we report the inference latency on different hardware platforms and discuss future applications.
DOI: 10.48550/arxiv.2203.16255
2022
Physics Community Needs, Tools, and Resources for Machine Learning
Machine learning (ML) is becoming an increasingly important component of cutting-edge physics research, but its computational requirements present significant challenges. In this white paper, we discuss the needs of the physics community regarding ML across latency and throughput regimes, the tools and resources that offer the possibility of addressing these needs, and how these can be best utilized and accessed in the coming years.
DOI: 10.1038/s42256-022-00486-4
2022
Author Correction: Autoencoders on field-programmable gate arrays for real-time, unsupervised new physics detection at 40 MHz at the Large Hadron Collider
2022
Lightweight Jet Reconstruction and Identification as an Object Detection Task
DOI: 10.48550/arxiv.2205.07690
2022
Real-time semantic segmentation on FPGAs for autonomous vehicles with hls4ml
In this paper, we investigate how field programmable gate arrays can serve as hardware accelerators for real-time semantic segmentation tasks relevant for autonomous driving. Considering compressed versions of the ENet convolutional neural network architecture, we demonstrate a fully-on-chip deployment with a latency of 4.9 ms per image, using less than 30% of the available resources on a Xilinx ZCU102 evaluation board. The latency is reduced to 3 ms per image when increasing the batch size to ten, corresponding to the use case where the autonomous vehicle receives inputs from multiple cameras simultaneously. We show, through aggressive filter reduction and heterogeneous quantization-aware training, and an optimized implementation of convolutional layers, that the power consumption and resource utilization can be significantly reduced while maintaining accuracy on the Cityscapes dataset.
DOI: 10.2172/1592124
2019
Accelerated Machine Learning as a Service for Particle Physics Computing
Accelerated Machine Learning as a Service for Particle Physics Computing: • Amount and complexity of high-energy physics data increases dramatically from 2025 onward • Traditional algorithms will require too much CPU time • Machine learning can solve combinatorially-scaling problems in constant time, but must be fast enough
DOI: 10.5072/zenodo.458983
2019
New-Physics agnostic searches for New Physics
DOI: 10.2172/1630707
2019
hls4ml: Deploying Deep Learning on FPGAs for L1 trigger and Data Acquisition
neural network to demonstrate state-of-the-art performance for top quark jet tagging at the LHC, and apply a ResNet-50 model with transfer learning for neutrino event classification. Using Project Brainwave b y Microsoft to accelerate the ResNet-50 image classification model, we achieve average inference times of 60 (10) milliseconds with our experimental physics software framework using Brainwave as a cloud (edge or on-premises) service, representing an improvement by a factor of approximately 30 (175) in model inference latency over traditional CPU inference in current experimental hardware. A single FPGA service accessed by many CPUs achieves a throughput of 600--700 inferences per second using an image batch of one, comparable to large batch-size GPU throughput and significantly better than small batch-size GPU throughput. Deployed as an edge or cloud service for the particle physics computing model, coprocessor accelerators can have a higher duty cycle and are potentially much more cost-effective.
DOI: 10.2172/1633738
2019
Interaction Network for Jet Characterization at the LHC
Deep learning plays a significant role in jet tagging. Interaction network / message passing network are parameter efficient. The proposed network out-performs some other deep learning approaches. There is promising direction for future taggers and other problems.
DOI: 10.5167/uzh-146421
2018
Measurements of $t\overline{t}$ cross sections in association with b jets and inclusive jets and their ratio using dilepton final states in pp collisions at $\sqrt{s}$ = 13 TeV
The cross sections for the production of $t\overline{t}b\overline{b}$ and $t\overline{t}jj$ events and their ratio $\sigma_{t\overline{t}b\overline{b}} / \sigma_{t\overline{t}jj}$ are measured using data corresponding to an integrated luminosity of 2.3 $fb^{−1}$collected in pp collisions at $\sqrt{s}$ = 13TeV with the CMS detector at the LHC. Events with two leptons (e or μ) and at least four reconstructed jets, including at least two identified as b quark jets, in the final state are selected. In the full phase space, the measured ratio is $0.022 \pm 0.003(stat) \pm 0.006(syst)$, the cross section $\sigma_{t\overline{t}b\overline{b}}$ is $4.0 \pm 0.6(stat) \pm 1.3(syst)pb$ and $\sigma_{t\overline{t}jj}$ is $184 \pm 6(stat) \pm 33(syst)pb$. The measurements are compared with the standard model expectations obtained from a powheg simulation at next-to-leading-order interfaced with pythia.
2019
FPGA-Accelerated Machine Learning Inference as a Service for Particle Physics Computing
2019
Interaction Network for Jet Characterization at the LHC [Slides]
DOI: 10.5281/zenodo.3895029
2019
Accelerated Machine Learning as a Service for Particle Physics Computing
DOI: 10.5281/zenodo.3598989
2019
hls4ml: Deploying Deep Learning on FPGAs for L1 trigger and Data Acquisition [PowerPoint]
2021
hls4ml: An Open-Source Codesign Workflow to Empower Scientific Low-Power Machine Learning Devices
Accessible machine learning algorithms, software, and diagnostic tools for energy-efficient devices and systems are extremely valuable across a broad range of application domains. In scientific domains, real-time near-sensor processing can drastically improve experimental design and accelerate scientific discoveries. To support domain scientists, we have developed hls4ml, an open-source software-hardware codesign workflow to interpret and translate machine learning algorithms for implementation with both FPGA and ASIC technologies. We expand on previous hls4ml work by extending capabilities and techniques towards low-power implementations and increased usability: new Python APIs, quantization-aware pruning, end-to-end FPGA workflows, long pipeline kernels for low power, and new device backends include an ASIC workflow. Taken together, these and continued efforts in hls4ml will arm a new generation of domain scientists with accessible, efficient, and powerful tools for machine-learning-accelerated discovery.
DOI: 10.5281/zenodo.4883651
2021
Jet Single Shot Detection
Data for training Jet Single Shot Detection network for simultaneous localization and classification of jets.
2021
arXiv : Jet Single Shot Detection
2021
arXiv : Improving Variational Autoencoders for New Physics Detection at the LHC with Normalizing Flows
We investigate how to improve new physics detection strategies exploiting variational autoencoders and normalizing flows for anomaly detection at the Large Hadron Collider. As a working example, we consider the DarkMachines challenge dataset. We show how different design choices (e.g., event representations, anomaly score definitions, network architectures) affect the result on specific benchmark new physics models. Once a baseline is established, we discuss how to improve the anomaly detection accuracy by exploiting normalizing flow layers in the latent space of the variational autoencoder.
DOI: 10.21468/scipost.report.3866
2021
Report on 2105.14027v2
We describe the outcome of a data challenge conducted as part of the Dark Machines (https://www.darkmachines.org)initiative and the Les Houches 2019 workshop on Physics at TeV colliders.The challenged aims to detect signals of new physics at the Large Hadron Collider (LHC) using unsupervised machine learning algorithms.First, we propose how an anomaly score could be implemented to define model-independent signal regions in LHC searches.We define and describe a large benchmark dataset, consisting of > 1 billion simulated LHC events corresponding to 10 fb -1 of proton-proton collisions at a center-of-mass energy of 13 TeV.We then review a wide range of anomaly detection and density estimation algorithms, developed in the context of the data challenge, and we measure their performance in a set of realistic analysis environments.We draw a number of useful conclusions that will aid the development of unsupervised new physics searches during the third run of the LHC, and provide our benchmark dataset for future studies at https://www.phenoMLdata.org.
2021
Autoencoders on FPGAs for real-time, unsupervised new physics detection at 40 MHz at the Large Hadron Collider
DOI: 10.48550/arxiv.2110.08508
2021
Improving Variational Autoencoders for New Physics Detection at the LHC with Normalizing Flows
We investigate how to improve new physics detection strategies exploiting variational autoencoders and normalizing flows for anomaly detection at the Large Hadron Collider. As a working example, we consider the DarkMachines challenge dataset. We show how different design choices (e.g., event representations, anomaly score definitions, network architectures) affect the result on specific benchmark new physics models. Once a baseline is established, we discuss how to improve the anomaly detection accuracy by exploiting normalizing flow layers in the latent space of the variational autoencoder.
DOI: 10.48550/arxiv.2107.02157
2021
LHC physics dataset for unsupervised New Physics detection at 40 MHz
In particle detectors at the Large Hadron Collider, tens of terabytes of data are produced every second from proton-proton collisions occurring at a rate of 40 megahertz. This data rate is reduced to a sustainable level by a real-time event filter processing system which decides whether each collision event should be kept for further analysis or be discarded. We introduce a dataset of proton collision events which emulates a typical data stream collected by such a real-time processing system, pre-filtered by requiring the presence of at least one electron or muon. This dataset could be used to develop novel event selection strategies and assess their sensitivity to new phenomena. In particular, by publishing this dataset we intend to stimulate a community-based effort towards the design of novel algorithms for performing unsupervised New Physics detection, customized to fit the bandwidth, latency and computational resource constraints of the real-time event selection system of a typical particle detector.
DOI: 10.48550/arxiv.2110.13041
2021
Applications and Techniques for Fast Machine Learning in Science
In this community review report, we discuss applications and techniques for fast machine learning (ML) in science -- the concept of integrating power ML methods into the real-time experimental data processing loop to accelerate scientific discovery. The material for the report builds on two workshops held by the Fast ML for Science community and covers three main areas: applications for fast ML across a number of scientific domains; techniques for training and implementing performant and resource-efficient ML algorithms; and computing architectures, platforms, and technologies for deploying these algorithms. We also present overlapping challenges across the multiple scientific domains where common solutions can be found. This community report is intended to give plenty of examples and inspiration for scientific discovery through integrated and accelerated ML solutions. This is followed by a high-level overview and organization of technical advances, including an abundance of pointers to source material, which can enable these breakthroughs.
DOI: 10.5281/zenodo.5070454
2021
Unsupervised New Physics detection at 40 MHz: Black Box Dataset
Unsupervised New Physics detection at 40 MHz data challenge Signal Benchmark Dataset consisting of the signal+background Black Box datasets, containing collision events (simulation of LHC 13 TeV proton-proton collisions) pre-filtered by a requirement of a muon or electron with 23 GeV transverse momentum. Data format description available on the data challenge web page: https://mpp-hep.github.io/ADC2021/
DOI: 10.5281/zenodo.5046428
2021
Unsupervised New Physics detection at 40 MHz: Training Dataset
Unsupervised New Physics detection at 40 MHz data challenge Training dataset, consisting of a cocktail of Standard Model collision events (simulation of LHC 13 TeV proton-proton collisions) pre-filtered by a requirement of a muon or electron with 23 GeV transverse momentum. Data format description available on the data challenge web page: https://mpp-hep.github.io/ADC2021/
DOI: 10.5281/zenodo.5046445
2021
Unsupervised New Physics detection at 40 MHz: A -&gt; 4 leptons Signal Benchmark Dataset
Unsupervised New Physics detection at 40 MHz data challenge Signal Benchmark Dataset consisting of A -&gt; 4 leptons decays produced in collision events (simulation of LHC 13 TeV proton-proton collisions) pre-filtered by a requirement of a muon or electron with 23 GeV transverse momentum. Data format description available on the data challenge web page: https://mpp-hep.github.io/ADC2021/
DOI: 10.5281/zenodo.5046388
2021
Unsupervised New Physics detection at 40 MHz: Training Dataset
Unsupervised New Physics detection at 40 MHz data challenge Training dataset, consisting of a cocktail of Standard Model collision events (simulation of LHC 13 TeV proton-proton collisions) pre-filtered by a requirement of a muon or electron with 23 GeV transverse momentum. Data format description available on the data challenge web page: https://mpp-hep.github.io/ADC2021/
DOI: 10.5281/zenodo.7152599
2021
Unsupervised New Physics detection at 40 MHz: LQ -&gt; b tau Signal Benchmark Dataset
Unsupervised New Physics detection at 40 MHz data challenge Signal Benchmark Dataset consisting of Leptoquarks -&gt; b tau decays produced in collision events (simulation of LHC 13 TeV proton-proton collisions) pre-filtered by a requirement of a muon or electron with 23 GeV transverse momentum. Data format description available on the data challenge web page: https://mpp-hep.github.io/ADC2021/
DOI: 10.5281/zenodo.5055453
2021
Unsupervised New Physics detection at 40 MHz: LQ -&gt; b tau Signal Benchmark Dataset
Unsupervised New Physics detection at 40 MHz data challenge Signal Benchmark Dataset consisting of Leptoquarks -&gt; b tau decays produced in collision events (simulation of LHC 13 TeV proton-proton collisions) pre-filtered by a requirement of a muon or electron with 23 GeV transverse momentum. Data format description available on the data challenge web page: https://mpp-hep.github.io/ADC2021/
DOI: 10.5281/zenodo.7152614
2021
Unsupervised New Physics detection at 40 MHz: h^0 -&gt; tau tau Signal Benchmark Dataset
Unsupervised New Physics detection at 40 MHz data challenge Signal Benchmark Dataset consisting of h^0 -&gt; tau tau decays produced in collision events (simulation of LHC 13 TeV proton-proton collisions) pre-filtered by a requirement of a muon or electron with 23 GeV transverse momentum. Data format description available on the data challenge web page: https://mpp-hep.github.io/ADC2021/
DOI: 10.5281/zenodo.5061687
2021
Unsupervised New Physics detection at 40 MHz: h+ -&gt; tau nu Signal Benchmark Dataset
Unsupervised New Physics detection at 40 MHz data challenge Signal Benchmark Dataset consisting of h+ -&gt; tau nu decays produced in collision events (simulation of LHC 13 TeV proton-proton collisions) pre-filtered by a requirement of a muon or electron with 23 GeV transverse momentum. Data format description available on the data challenge web page: https://mpp-hep.github.io/ADC2021/
DOI: 10.5281/zenodo.5061632
2021
Unsupervised New Physics detection at 40 MHz: h^0 -&gt; tau tau Signal Benchmark Dataset
Unsupervised New Physics detection at 40 MHz data challenge Signal Benchmark Dataset consisting of h^0 -&gt; tau tau decays produced in collision events (simulation of LHC 13 TeV proton-proton collisions) pre-filtered by a requirement of a muon or electron with 23 GeV transverse momentum. Data format description available on the data challenge web page: https://mpp-hep.github.io/ADC2021/
DOI: 10.5281/zenodo.7152617
2021
Unsupervised New Physics detection at 40 MHz: h+ -&gt; tau nu Signal Benchmark Dataset
Unsupervised New Physics detection at 40 MHz data challenge Signal Benchmark Dataset consisting of h+ -&gt; tau nu decays produced in collision events (simulation of LHC 13 TeV proton-proton collisions) pre-filtered by a requirement of a muon or electron with 23 GeV transverse momentum. Data format description available on the data challenge web page: https://mpp-hep.github.io/ADC2021/
DOI: 10.5281/zenodo.5070455
2021
Unsupervised New Physics detection at 40 MHz: Black Box Dataset
Unsupervised New Physics detection at 40 MHz data challenge Signal Benchmark Dataset consisting of the signal+background Black Box datasets, containing collision events (simulation of LHC 13 TeV proton-proton collisions) pre-filtered by a requirement of a muon or electron with 23 GeV transverse momentum. Data format description available on the data challenge web page: https://mpp-hep.github.io/ADC2021/
DOI: 10.5281/zenodo.7152590
2021
Unsupervised New Physics detection at 40 MHz: A -&gt; 4 leptons Signal Benchmark Dataset
Unsupervised New Physics detection at 40 MHz data challenge Signal Benchmark Dataset consisting of A -&gt; 4 leptons decays produced in collision events (simulation of LHC 13 TeV proton-proton collisions) pre-filtered by a requirement of a muon or electron with 23 GeV transverse momentum. Data format description available on the data challenge web page: https://mpp-hep.github.io/ADC2021/
DOI: 10.5281/zenodo.5072068
2021
Unsupervised New Physics detection at 40 MHz: Black Box Dataset
Unsupervised New Physics detection at 40 MHz data challenge Signal Benchmark Dataset consisting of the signal+background Black Box datasets, containing collision events (simulation of LHC 13 TeV proton-proton collisions) pre-filtered by a requirement of a muon or electron with 23 GeV transverse momentum. Data format description available on the data challenge web page: https://mpp-hep.github.io/ADC2021/