ϟ

M. Pierini

Here are all the papers by M. Pierini that you can download and read on OA.mg.
M. Pierini’s last known institution is . Download M. Pierini PDFs here.

Claim this Profile →
DOI: 10.1088/1126-6708/2008/03/049
2008
Cited 389 times
Model-independent constraints on Δ<i>F</i>= 2 operators and the scale of new physics
We update the constraints on new-physics contributions to Delta F=2 processes from the generalized unitarity triangle analysis, including the most recent experimental developments. Based on these constraints, we derive upper bounds on the coefficients of the most general Delta F=2 effective Hamiltonian. These upper bounds can be translated into lower bounds on the scale of new physics that contributes to these low-energy effective interactions. We point out that, due to the enhancement in the renormalization group evolution and in the matrix elements, the coefficients of non-standard operators are much more constrained than the coefficient of the operator present in the Standard Model. Therefore, the scale of new physics in models that generate new Delta F=2 operators, such as next-to-minimal flavour violation, has to be much higher than the scale of minimal flavour violation, and it most probably lies beyond the reach of direct searches at the LHC.
DOI: 10.1007/jhep01(2014)164
2014
Cited 294 times
First look at the physics case of TLEP
A bstract The discovery by the ATLAS and CMS experiments of a new boson with mass around 125 GeV and with measured properties compatible with those of a Standard-Model Higgs boson, coupled with the absence of discoveries of phenomena beyond the Standard Model at the TeV scale, has triggered interest in ideas for future Higgs factories. A new circular e + e − collider hosted in a 80 to 100 km tunnel, TLEP, is among the most attractive solutions proposed so far. It has a clean experimental environment, produces high luminosity for top-quark, Higgs boson, W and Z studies, accommodates multiple detectors, and can reach energies up to the $$ \mathrm{t}\overline{\mathrm{t}} $$ threshold and beyond. It will enable measurements of the Higgs boson properties and of Electroweak Symmetry-Breaking (EWSB) parameters with unequalled precision, offering exploration of physics beyond the Standard Model in the multi-TeV range. Moreover, being the natural precursor of the VHE-LHC, a 100 TeV hadron machine in the same tunnel, it builds up a long-term vision for particle physics. Altogether, the combination of TLEP and the VHE-LHC offers, for a great cost effectiveness, the best precision and the best search reach of all options presently on the market. This paper presents a first appraisal of the salient features of the TLEP physics potential, to serve as a baseline for a more extensive design study.
DOI: 10.1088/1748-0221/13/07/p07027
2018
Cited 269 times
Fast inference of deep neural networks in FPGAs for particle physics
Recent results at the Large Hadron Collider (LHC) have pointed to enhanced physics capabilities through the improvement of the real-time event processing techniques. Machine learning methods are ubiquitous and have proven to be very powerful in LHC physics, and particle physics as a whole. However, exploration of the use of such techniques in low-latency, low-power FPGA hardware has only just begun. FPGA-based trigger and data acquisition (DAQ) systems have extremely low, sub-microsecond latency requirements that are unique to particle physics. We present a case study for neural network inference in FPGAs focusing on a classifier for jet substructure which would enable, among many other physics scenarios, searches for new dark sector particles and novel measurements of the Higgs boson. While we focus on a specific example, the lessons are far-reaching. We develop a package based on High-Level Synthesis (HLS) called hls4ml to build machine learning models in FPGAs. The use of HLS increases accessibility across a broad user community and allows for a drastic decrease in firmware development time. We map out FPGA resource usage and latency versus neural network hyperparameters to identify the problems in particle physics that would benefit from performing neural network inference with FPGAs. For our example jet substructure model, we fit well within the available resources of modern FPGAs with a latency on the scale of 100 ns.
DOI: 10.1007/jhep05(2019)036
2019
Cited 132 times
Variational autoencoders for new physics mining at the Large Hadron Collider
A bstract Using variational autoencoders trained on known physics processes, we develop a one-sided threshold test to isolate previously unseen processes as outlier events. Since the autoencoder training does not depend on any specific new physics signature, the proposed procedure doesn’t make specific assumptions on the nature of new physics. An event selection based on this algorithm would be complementary to classic LHC searches, typically based on model-dependent hypothesis testing. Such an algorithm would deliver a list of anomalous events, that the experimental collaborations could further scrutinize and even release as a catalog, similarly to what is typically done in other scientific domains. Event topologies repeating in this dataset could inspire new-physics model building and new experimental searches. Running in the trigger system of the LHC experiments, such an application could identify anomalous events that would be otherwise lost, extending the scientific reach of the LHC.
DOI: 10.1038/s42256-021-00356-5
2021
Cited 98 times
Automatic heterogeneous quantization of deep neural networks for low-latency inference on the edge for particle detectors
Although the quest for more accurate solutions is pushing deep learning research towards larger and more complex algorithms, edge devices demand efficient inference and therefore reduction in model size, latency and energy consumption. One technique to limit model size is quantization, which implies using fewer bits to represent weights and biases. Such an approach usually results in a decline in performance. Here, we introduce a method for designing optimally heterogeneously quantized versions of deep neural network models for minimum-energy, high-accuracy, nanosecond inference and fully automated deployment on chip. With a per-layer, per-parameter type automatic quantization procedure, sampling from a wide range of quantizers, model energy consumption and size are minimized while high accuracy is maintained. This is crucial for the event selection procedure in proton–proton collisions at the CERN Large Hadron Collider, where resources are strictly limited and a latency of $${\mathcal{O}}(1)\,\upmu{\rm{s}}$$ is required. Nanosecond inference and a resource consumption reduced by a factor of 50 when implemented on field-programmable gate array hardware are achieved. With edge computing on custom hardware, real-time inference with deep neural networks can reach the nanosecond timescale. An important application in this regime is event processing at particle collision detectors like those at the Large Hadron Collider (LHC). To ensure high performance as well as reduced resource consumption, a method is developed, and made available as an extension of the Keras library, to automatically design optimal quantization of the different layers in a deep neural network.
DOI: 10.1088/1361-6633/ac36b9
2021
Cited 78 times
The LHC Olympics 2020 a community challenge for anomaly detection in high energy physics
A new paradigm for data-driven, model-agnostic new physics searches at colliders is emerging, and aims to leverage recent breakthroughs in anomaly detection and machine learning. In order to develop and benchmark new anomaly detection methods within this framework, it is essential to have standard datasets. To this end, we have created the LHC Olympics 2020, a community challenge accompanied by a set of simulated collider events. Participants in these Olympics have developed their methods using an R&D dataset and then tested them on black boxes: datasets with an unknown anomaly (or not). Methods made use of modern machine learning tools and were based on unsupervised learning (autoencoders, generative adversarial networks, normalizing flows), weakly supervised learning, and semi-supervised learning. This paper will review the LHC Olympics 2020 challenge, including an overview of the competition, a description of methods deployed in the competition, lessons learned from the experience, and implications for data analyses with future datasets as well as future colliders.
DOI: 10.21468/scipostphys.12.1.043
2022
Cited 50 times
The Dark Machines Anomaly Score Challenge: Benchmark Data and Model Independent Event Classification for the Large Hadron Collider
We describe the outcome of a data challenge conducted as part of the Dark Machines Initiative and the Les Houches 2019 workshop on Physics at TeV colliders. The challenged aims at detecting signals of new physics at the LHC using unsupervised machine learning algorithms. First, we propose how an anomaly score could be implemented to define model-independent signal regions in LHC searches. We define and describe a large benchmark dataset, consisting of >1 Billion simulated LHC events corresponding to $10~\rm{fb}^{-1}$ of proton-proton collisions at a center-of-mass energy of 13 TeV. We then review a wide range of anomaly detection and density estimation algorithms, developed in the context of the data challenge, and we measure their performance in a set of realistic analysis environments. We draw a number of useful conclusions that will aid the development of unsupervised new physics searches during the third run of the LHC, and provide our benchmark dataset for future studies at https://www.phenoMLdata.org. Code to reproduce the analysis is provided at https://github.com/bostdiek/DarkMachines-UnsupervisedChallenge.
DOI: 10.1103/physrevlett.129.271801
2022
Cited 47 times
Impact of the Recent Measurements of the Top-Quark and <mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" display="inline"><mml:mi>W</mml:mi></mml:math> -Boson Masses on Electroweak Precision Fits
We assess the impact of the very recent measurement of the top-quark mass by the CMS Collaboration on the fit of electroweak data in the standard model and beyond, with particular emphasis on the prediction for the mass of the W boson. We then compare this prediction with the average of the corresponding experimental measurements including the new measurement by the CDF Collaboration, and discuss its compatibility in the standard model, in new physics models with oblique corrections, and in the dimension-six standard model effective field theory. Finally, we present the updated global fit to electroweak precision data in these models.
DOI: 10.1007/s12210-023-01137-5
2023
Cited 22 times
New UTfit analysis of the unitarity triangle in the Cabibbo–Kobayashi–Maskawa scheme
Flavour mixing and CP violation as measured in weak decays and mixing of neutral mesons are a fundamental tool to test the Standard Model and to search for new physics. New analyses performed at the LHC experiment open an unprecedented insight into the Cabibbo–Kobayashi–Maskawa metrology and new evidence for rare decays. Important progress has also been achieved in theoretical calculations of several hadronic quantities with a remarkable reduction of the uncertainties. This improvement is essential since previous studies of the Unitarity Triangle did show that possible contributions from new physics, if any, must be tiny and could easily be hidden by theoretical and experimental errors. Thanks to the experimental and theoretical advances, the Cabibbo–Kobayashi–Maskawa picture provides very precise Standard Model predictions through global analyses. We present here the results of the latest global Standard Model analysis performed by the UTfit collaboration including all the most updated inputs from experiments, lattice Quantum Chromo-Dynamics and phenomenological calculations.
DOI: 10.1088/1126-6708/2005/07/028
2005
Cited 182 times
The 2004 UTfit collaboration report on the status of the unitarity triangle in the standard model
Using the latest determinations of several theoretical and experimental parameters, we update the Unitarity Triangle analysis in the Standard Model. The basic experimental constraints come from the measurements of |Vub/Vcb|, Δmd, the lower limit on Δms, K, and the measurement of the phase of the Bd–dmixing amplitude through the time-dependent CP asymmetry in B0→J/ψK0 decays. In addition, we consider the direct determination of α, γ, 2β+γ and cos 2β from the measurements of new CP-violating quantities, recently performed at the B factories. We also discuss the opportunities offered by improving the precision of the various physical quantities entering in the determination of the Unitarity Triangle parameters. The results and the plots presented in this paper can also be found at the URL http://www.utfit.org , where they are continuously updated with the newest experimental and theoretical results.
DOI: 10.1016/j.physrep.2010.05.003
2010
Cited 178 times
Flavor physics in the quark sector
In the past decade, one of the major challenges of particle physics has been to gain an in-depth understanding of the role of quark flavor. In this time frame, measurements and the theoretical interpretation of their results have advanced tremendously. A much broader understanding of flavor particles has been achieved; apart from their masses and quantum numbers, there now exist detailed measurements of the characteristics of their interactions allowing stringent tests of Standard Model predictions. Among the most interesting phenomena of flavor physics is the violation of the CP symmetry that has been subtle and difficult to explore. In the past, observations of CP violation were confined to neutral K mesons, but since the early 1990s, a large number of CP-violating processes have been studied in detail in neutral B mesons. In parallel, measurements of the couplings of the heavy quarks and the dynamics for their decays in large samples of K,D, and B mesons have been greatly improved in accuracy and the results are being used as probes in the search for deviations from the Standard Model. In the near future, there will be a transition from the current to a new generation of experiments; thus a review of the status of quark flavor physics is timely. This report is the result of the work of physicists attending the 5th CKM workshop, hosted by the University of Rome "La Sapienza", September 9–13, 2008. It summarizes the results of the current generation of experiments that are about to be completed and it confronts these results with the theoretical understanding of the field which has greatly improved in the past decade.
DOI: 10.1140/epjc/s10052-008-0716-1
2008
Cited 146 times
B, D and K decays
The present report documents the results of Working Group 2: B, D and K decays, of the workshop on Flavor in the Era of the LHC, held at CERN from November 2005 through March 2007. With the advent of the LHC, we will be able to probe New Physics (NP) up to energy scales almost one order of magnitude larger than it has been possible with present accelerator facilities. While direct detection of new particles will be the main avenue to establish the presence of NP at the LHC, indirect searches will provide precious complementary information, since most probably it will not be possible to measure the full spectrum of new particles and their couplings through direct production. In particular, precision measurements and computations in the realm of flavor physics are expected to play a key role in constraining the unknown parameters of the Lagrangian of any NP model emerging from direct searches at the LHC. The aim of Working Group 2 was twofold: on the one hand, to provide a coherent up-to-date picture of the status of flavor physics before the start of the LHC; on the other hand, to initiate activities on the path towards integrating information on NP from high-p T and flavor data. This report is organized as follows: in Sect. 1, we give an overview of NP models, focusing on a few examples that have been discussed in some detail during the workshop, with a short description of the available computational tools for flavor observables in NP models. Section 2 contains a concise discussion of the main theoretical problem in flavor physics: the evaluation of the relevant hadronic matrix elements for weak decays. Section 3 contains a detailed discussion of NP effects in a set of flavor observables that we identified as "benchmark channels" for NP searches. The experimental prospects for flavor physics at future facilities are discussed in Sect. 4. Finally, Sect. 5 contains some assessments on the work done at the workshop and the prospects for future developments.
DOI: 10.1007/jhep12(2016)135
2016
Cited 127 times
Electroweak precision observables and Higgs-boson signal strengths in the Standard Model and beyond: present and future
We present results from a state-of-the-art fit of electroweak precision observables and Higgs-boson signal-strength measurements performed using 7 and 8 TeV data from the Large Hadron Collider. Based on the HEPfit package, our study updates the traditional fit of electroweak precision observables and extends it to include Higgs-boson measurements. As a result we obtain constraints on new physics corrections to both electroweak observables and Higgs-boson couplings. We present the projected accuracy of the fit taking into account the expected sensitivities at future colliders.
DOI: 10.1103/physrevd.82.013003
2010
Cited 112 times
Higgs boson look-alikes at the LHC
The discovery of a Higgs particle is possible in a variety of search channels at the LHC. However, the true identity of any putative Higgs boson will, at first, remain ambiguous until one has experimentally excluded other possible assignments of quantum numbers and couplings. We quantify the degree to which one can discriminate a standard model Higgs boson from ``look-alikes'' at, or close to, the moment of discovery at the LHC. We focus on the fully-reconstructible golden decay mode to a pair of $Z$ bosons and a four-lepton final state. Considering both on-shell and off-shell $Z$'s, we show how to utilize the full decay information from the events, including the distributions and correlations of the five relevant angular variables. We demonstrate how the finite phase space acceptance of any LHC detector sculpts the decay distributions, a feature neglected in previous studies. We use likelihood ratios to discriminate a standard model Higgs from look-alikes with other spins or nonstandard parity, $CP$, or form factors. For a resonance mass of $200\text{ }\text{ }\mathrm{GeV}/{c}^{2}$, we achieve a median discrimination significance of $3\ensuremath{\sigma}$ with as few as 19 events, and even better discrimination for the off-shell decays of a $145\text{ }\text{ }\mathrm{GeV}/{c}^{2}$ resonance.
DOI: 10.1140/epjc/s10052-020-8251-9
2020
Cited 97 times
Calorimetry with deep learning: particle simulation and reconstruction for collider physics
Abstract Using detailed simulations of calorimeter showers as training data, we investigate the use of deep learning algorithms for the simulation and reconstruction of single isolated particles produced in high-energy physics collisions. We train neural networks on single-particle shower data at the calorimeter-cell level, and show significant improvements for simulation and reconstruction when using these networks compared to methods which rely on currently-used state-of-the-art algorithms. We define two models: an end-to-end reconstruction network which performs simultaneous particle identification and energy regression of particles when given calorimeter shower data, and a generative network which can provide reasonable modeling of calorimeter showers for different particle types at specified angles and energies. We investigate the optimization of our models with hyperparameter scans. Furthermore, we demonstrate the applicability of the reconstruction model to shower inputs from other detector geometries, specifically ATLAS-like and CMS-like geometries. These networks can serve as fast and computationally light methods for particle shower simulation and reconstruction for current and future experiments at particle colliders.
DOI: 10.1140/epjc/s10052-020-7904-z
2020
Cited 97 times
HEPfit: a code for the combination of indirect and direct constraints on high energy physics models
Abstract is a flexible open-source tool which, given the Standard Model or any of its extensions, allows to (i) fit the model parameters to a given set of experimental observables; (ii) obtain predictions for observables. can be used either in Monte Carlo mode, to perform a Bayesian Markov Chain Monte Carlo analysis of a given model, or as a library, to obtain predictions of observables for a given point in the parameter space of the model, allowing to be used in any statistical framework. In the present version, around a thousand observables have been implemented in the Standard Model and in several new physics scenarios. In this paper, we describe the general structure of the code as well as models and observables implemented in the current release.
DOI: 10.1140/epjc/s10052-020-7608-4
2020
Cited 94 times
JEDI-net: a jet identification algorithm based on interaction networks
Abstract We investigate the performance of a jet identification algorithm based on interaction networks (JEDI-net) to identify all-hadronic decays of high-momentum heavy particles produced at the LHC and distinguish them from ordinary jets originating from the hadronization of quarks and gluons. The jet dynamics are described as a set of one-to-one interactions between the jet constituents. Based on a representation learned from these interactions, the jet is associated to one of the considered categories. Unlike other architectures, the JEDI-net models achieve their performance without special handling of the sparse input jet representation, extensive pre-processing, particle ordering, or specific assumptions regarding the underlying detector geometry. The presented models give better results with less model parameters, offering interesting prospects for LHC applications.
DOI: 10.1140/epjc/s10052-019-7113-9
2019
Cited 91 times
Learning representations of irregular particle-detector geometry with distance-weighted graph networks
We explore the use of graph networks to deal with irregular-geometry detectors in the context of particle reconstruction. Thanks to their representation-learning capabilities, graph networks can exploit the full detector granularity, while natively managing the event sparsity and arbitrarily complex detector geometries. We introduce two distance-weighted graph network architectures, dubbed GarNet and GravNet layers, and apply them to a typical particle reconstruction task. The performance of the new architectures is evaluated on a data set of simulated particle interactions on a toy model of a highly granular calorimeter, loosely inspired by the endcap calorimeter to be installed in the CMS detector for the High-Luminosity LHC phase. We study the clustering of energy depositions, which is the basis for calorimetric particle reconstruction, and provide a quantitative comparison to alternative approaches. The proposed algorithms provide an interesting alternative to existing methods, offering equally performing or less resource-demanding solutions with less underlying assumptions on the detector geometry and, consequently, the possibility to generalize to other detectors.
DOI: 10.1088/2632-2153/aba042
2020
Cited 60 times
Compressing deep neural networks on FPGAs to binary and ternary precision with <tt>hls4ml</tt>
We present the implementation of binary and ternary neural networks in the hls4ml library, designed to automatically convert deep neural network models to digital circuits with field-programmable gate arrays (FPGA) firmware. Starting from benchmark models trained with floating point precision, we investigate different strategies to reduce the network's resource consumption by reducing the numerical precision of the network parameters to binary or ternary. We discuss the trade-off between model accuracy and resource consumption. In addition, we show how to balance between latency and accuracy by retaining full precision on a selected subset of network components. As an example, we consider two multiclass classification tasks: handwritten digit recognition with the MNIST data set and jet identification with simulated proton-proton collisions at the CERN Large Hadron Collider. The binary and ternary implementation has similar performance to the higher precision implementation while using drastically fewer FPGA resources.
DOI: 10.1140/epjp/s13360-021-01109-4
2021
Cited 55 times
Adversarially Learned Anomaly Detection on CMS open data: re-discovering the top quark
Abstract We apply an Adversarially Learned Anomaly Detection (ALAD) algorithm to the problem of detecting new physics processes in proton–proton collisions at the Large Hadron Collider. Anomaly detection based on ALAD matches performances reached by Variational Autoencoders, with a substantial improvement in some cases. Training the ALAD algorithm on 4.4 fb $$^{-1}$$ <mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML"> <mml:msup> <mml:mrow /> <mml:mrow> <mml:mo>-</mml:mo> <mml:mn>1</mml:mn> </mml:mrow> </mml:msup> </mml:math> of 8 TeV CMS Open Data, we show how a data-driven anomaly detection and characterization would work in real life, re-discovering the top quark by identifying the main features of the $$t \bar{t}$$ <mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML"> <mml:mrow> <mml:mi>t</mml:mi> <mml:mover> <mml:mrow> <mml:mi>t</mml:mi> </mml:mrow> <mml:mrow> <mml:mo>¯</mml:mo> </mml:mrow> </mml:mover> </mml:mrow> </mml:math> experimental signature at the LHC.
DOI: 10.1140/epjc/s10052-021-08853-y
2021
Cited 54 times
Learning multivariate new physics
Abstract We discuss a method that employs a multilayer perceptron to detect deviations from a reference model in large multivariate datasets. Our data analysis strategy does not rely on any prior assumption on the nature of the deviation. It is designed to be sensitive to small discrepancies that arise in datasets dominated by the reference model. The main conceptual building blocks were introduced in D’Agnolo and Wulzer (Phys Rev D 99 (1), 015014. 10.1103/PhysRevD.99.015014 . arXiv:1806.02350 [hep-ph], 2019). Here we make decisive progress in the algorithm implementation and we demonstrate its applicability to problems in high energy physics. We show that the method is sensitive to putative new physics signals in di-muon final states at the LHC. We also compare our performances on toy problems with the ones of alternative methods proposed in the literature.
DOI: 10.1088/2632-2153/ac0ea1
2021
Cited 53 times
Fast convolutional neural networks on FPGAs with hls4ml
Abstract We introduce an automated tool for deploying ultra low-latency, low-power deep neural networks with convolutional layers on field-programmable gate arrays (FPGAs). By extending the hls4ml library, we demonstrate an inference latency of 5 µ s using convolutional architectures, targeting microsecond latency applications like those at the CERN Large Hadron Collider. Considering benchmark models trained on the Street View House Numbers Dataset, we demonstrate various methods for model compression in order to fit the computational constraints of a typical FPGA device used in trigger and data acquisition systems of particle detectors. In particular, we discuss pruning and quantization-aware training, and demonstrate how resource utilization can be significantly reduced with little to no loss in model accuracy. We show that the FPGA critical resource consumption can be reduced by 97% with zero loss in model accuracy, and by 99% when tolerating a 6% accuracy degradation.
DOI: 10.1140/epjc/s10052-021-09158-w
2021
Cited 46 times
MLPF: efficient machine-learned particle-flow reconstruction using graph neural networks
In general-purpose particle detectors, the particle-flow algorithm may be used to reconstruct a comprehensive particle-level view of the event by combining information from the calorimeters and the trackers, significantly improving the detector resolution for jets and the missing transverse momentum. In view of the planned high-luminosity upgrade of the CERN Large Hadron Collider (LHC), it is necessary to revisit existing reconstruction algorithms and ensure that both the physics and computational performance are sufficient in an environment with many simultaneous proton-proton interactions (pileup). Machine learning may offer a prospect for computationally efficient event reconstruction that is well-suited to heterogeneous computing platforms, while significantly improving the reconstruction quality over rule-based algorithms for granular detectors. We introduce MLPF, a novel, end-to-end trainable, machine-learned particle-flow algorithm based on parallelizable, computationally efficient, and scalable graph neural networks optimized using a multi-task objective on simulated events. We report the physics and computational performance of the MLPF algorithm on a Monte Carlo dataset of top quark-antiquark pairs produced in proton-proton collisions in conditions similar to those expected for the high-luminosity LHC. The MLPF algorithm improves the physics response with respect to a rule-based benchmark algorithm and demonstrates computationally scalable particle-flow reconstruction in a high-pileup environment.
DOI: 10.3389/fdata.2020.598927
2021
Cited 41 times
Distance-Weighted Graph Neural Networks on FPGAs for Real-Time Particle Reconstruction in High Energy Physics
Graph neural networks have been shown to achieve excellent performance for several crucial tasks in particle physics, such as charged particle tracking, jet tagging, and clustering. An important domain for the application of these networks is the FGPA-based first layer of real-time data filtering at the CERN Large Hadron Collider, which has strict latency and resource constraints. We discuss how to design distance-weighted graph networks that can be executed with a latency of less than 1$\mu\mathrm{s}$ on an FPGA. To do so, we consider a representative task associated to particle reconstruction and identification in a next-generation calorimeter operating at a particle collider. We use a graph network architecture developed for such purposes, and apply additional simplifications to match the computing constraints of Level-1 trigger systems, including weight quantization. Using the $\mathtt{hls4ml}$ library, we convert the compressed models into firmware to be implemented on an FPGA. Performance of the synthesized models is presented both in terms of inference accuracy and resource usage.
DOI: 10.1103/physrevd.106.033003
2022
Cited 36 times
Global analysis of electroweak data in the Standard Model
We perform a global fit of electroweak data within the Standard Model, using state-of-the art experimental and theoretical results, including a determination of the electromagnetic coupling at the electroweak scale based on recent lattice calculations. In addition to the posteriors for all parameters and observables obtained from the global fit, we present indirect determinations for all parameters and predictions for all observables. Furthermore, we present full predictions, obtained using only the experimental information on Standard Model parameters, and a fully indirect determination of Standard Model parameters using only experimental information on electroweak data. Finally, we discuss in detail the compatibility of experimental data with the Standard Model and find a global p-value of 0.5.
DOI: 10.21468/scipostphys.12.1.037
2022
Cited 30 times
Publishing statistical models: Getting the most out of particle physics experiments
The statistical models used to derive the results of experimental analyses are of incredible scientific value and are essential information for analysis preservation and reuse. In this paper, we make the scientific case for systematically publishing the full statistical models and discuss the technical developments that make this practical. By means of a variety of physics cases - including parton distribution functions, Higgs boson measurements, effective field theory interpretations, direct searches for new physics, heavy flavor physics, direct dark matter detection, world averages, and beyond the Standard Model global fits - we illustrate how detailed information on the statistical modelling can enhance the short- and long-term impact of experimental results.
DOI: 10.1103/physrevd.107.076017
2023
Cited 14 times
Evaluating generative models in high energy physics
There has been a recent explosion in research into machine-learning-based generative modeling to tackle computational challenges for simulations in high energy physics (HEP). In order to use such alternative simulators in practice, we need well-defined metrics to compare different generative models and evaluate their discrepancy from the true distributions. We present the first systematic review and investigation into evaluation metrics and their sensitivity to failure modes of generative models, using the framework of two-sample goodness-of-fit testing, and their relevance and viability for HEP. Inspired by previous work in both physics and computer vision, we propose two new metrics, the Fr\'echet and kernel physics distances (FPD and KPD, respectively) and perform a variety of experiments measuring their performance on simple Gaussian-distributed and simulated high energy jet datasets. We find FPD, in particular, to be the most sensitive metric to all alternative jet distributions tested and recommend its adoption, along with the KPD and Wasserstein distances between individual feature distributions, for evaluating generative models in HEP. We finally demonstrate the efficacy of these proposed metrics in evaluating and comparing a novel attention-based generative adversarial particle transformer to the state-of-the-art message-passing generative adversarial network jet simulation model. The code for our proposed metrics is provided in the open source jetnet python library.
DOI: 10.1016/s0370-2693(01)00700-6
2001
Cited 155 times
Charming penguins strike back
By using the recent experimental measurements of B→ππ and B→Kπ branching ratios, we find that factorization is unable to reproduce the observed BRs even taking into account the uncertainties of the input parameters. Charming and GIM penguins allow to reconcile the theoretical predictions with the data. Because of these large effects, we conclude, however, that it is not possible, with the present theoretical and experimental accuracy, to determine the CP violation angle γ from these decays. Contrary to factorization, we predict large asymmetries for several of the particle–antiparticle BRs, in particular BR(B+→K+π0), BR(Bd→K+π−) and BR(Bd→π+π−). This opens new perspectives for the study of CP violation in B systems.
DOI: 10.1088/1126-6708/2006/10/081
2006
Cited 123 times
The unitarity triangle fit in the standard model and hadronic parameters from lattice QCD: a reappraisal after the measurements of Δ<i>m</i><sub><i>s</i></sub>and<i>BR</i>(<i>B</i>→τν<sub>τ</sub>)
The recent measurements of the B_s meson mixing amplitude by CDF and of the leptonic branching fraction BR(B to tau nu) by Belle call for an upgraded analysis of the Unitarity Triangle in the Standard Model. Besides improving the previous constraints on the parameters of the CKM matrix, these new measurements, combined with the recent determinations of the angles alpha, beta and gamma from non-leptonic decays, allow, in the Standard Model, a quite accurate extraction of the values of the hadronic matrix elements relevant for K-Kbar and B_{s,d}-B_{s,d}bar mixing and of the leptonic decay constant f_B. These values, obtained ``experimentally'', can then be compared with the theoretical predictions, mainly from lattice QCD. In this paper we upgrade the UT fit, we determine from the data the kaon B-parameter B_Khat, the B^0 mixing amplitude parameters f_Bs B^{1/2}_Bs and xi, the decay constant f_B, and make a comparison of the obtained values with lattice predictions. We also discuss the different determinations of V_ub and show that current data do not favour the value measured in inclusive decays.
DOI: 10.1103/physrevlett.95.221804
2005
Cited 114 times
Effect of Penguin Operators in the<mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" display="inline"><mml:msup><mml:mi>B</mml:mi><mml:mn>0</mml:mn></mml:msup><mml:mo>→</mml:mo><mml:mi>J</mml:mi><mml:mo>/</mml:mo><mml:mi>ψ</mml:mi><mml:msup><mml:mi>K</mml:mi><mml:mn>0</mml:mn></mml:msup></mml:math><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" display="inline"><mml:mi>C</mml:mi><mml:mi>P</mml:mi></mml:math>Asymmetry
Performing a fit to the available experimental data, we quantify the effect of long-distance contributions from penguin contractions in B0 --> J/psiK0 decays. We estimate the deviation of the measured S(CP) term of the time-dependent CP asymmetry from sin2beta induced by these contributions and by the penguin operators. We find deltaS is identically equal to S(CP)(J/psiK) - sin2beta = 0.000 +/- 0.012 ([-0.025, 0.024]@95% probability), an uncertainty much larger than previous estimates and comparable to the present systematic error quoted by the experiments at the B factories.
DOI: 10.1103/physrevlett.97.151803
2006
Cited 98 times
Constraints on New Physics from the Quark Mixing Unitarity Triangle
The status of the unitary triangle beyond the standard model including the most recent results on Deltam[s] on dilepton asymmetries and on width differences is presented. Even allowing for general new physics loop contributions the unitarity triangle must be very close to the standard model result. With the new measurements from the Fermilab Tevatron, we obtain for the first time a significant constraint on new physics in the Bs sector. We present the allowed ranges of new physics contributions to DeltaF=2 processes and of the time-dependent CP asymmetry in Bs-->J/psivarphi decays.
DOI: 10.1016/j.physletb.2007.08.055
2007
Cited 91 times
<mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" altimg="si1.gif" overflow="scroll"><mml:mi>D</mml:mi><mml:mtext>–</mml:mtext><mml:mover accent="true"><mml:mi>D</mml:mi><mml:mo>¯</mml:mo></mml:mover></mml:math> mixing and new physics: General considerations and constraints on the MSSM
Combining the recent experimental evidence of D–D¯ mixing, we extract model-independent information on the mixing amplitude and on its CP-violating phase. Using this information, we present new constraints on the flavour structure of up-type squark mass matrices in supersymmetric extensions of the Standard Model.
DOI: 10.1016/j.physletb.2010.02.063
2010
Cited 88 times
An improved Standard Model prediction of <mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" altimg="si1.gif" overflow="scroll"><mml:mi mathvariant="italic">BR</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>B</mml:mi><mml:mo>→</mml:mo><mml:mi>τ</mml:mi><mml:mi>ν</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math> and its implications for New Physics
The recently measured B -> tau nu branching ratio allows to test the Standard Model by probing virtual effects of new heavy particles, such as a charged Higgs boson. The accuracy of the test is currently limited by the experimental error on BR(B -> tau nu) and by the uncertainty on the parameters fB and |Vub|. The redundancy of the Unitarity Triangle fit allows to reduce the error on these parameters and thus to perform a more precise test of the Standard Model. Using the current experimental inputs, we obtain BR(B -> tau nu)_SM = (0.84 +- 0.11)x10^{-4}, to be compared with BR(B -> tau nu)_exp = (1.73 +- 0.34)x10^{-4}. The Standard Model prediction can be modified by New Physics effects in the decay amplitude as well as in the Unitarity Triangle fit. We discuss how to disentangle the two possible contributions in the case of minimal flavour violation at large tan beta and generic loop-mediated New Physics. We also consider two specific models with minimal flavour violation: the Type-II Two Higgs Doublet Model and the Minimal Supersymmetric Standard Model.
DOI: 10.1140/epjc/s10052-013-2370-5
2013
Cited 76 times
The light stop window
We show that a right-handed stop in the 200-400 GeV mass range, together with a nearly degenerate neutralino and, possibly, a gluino below 1.5 TeV, follows from reasonable assumptions, is consistent with present data, and offers interesting discovery prospects at the LHC. Triggering on an extra jet produced in association with stops allows the experimental search for stops even when their mass difference with neutralinos is very small and the decay products are too soft for direct observation. Using a razor analysis, we are able to set stop bounds that are stronger than those published by ATLAS and CMS.
DOI: 10.1007/jhep02(2012)075
2012
Cited 65 times
Interpreting LHC SUSY searches in the phenomenological MSSM
We interpret within the phenomenological MSSM (pMSSM) the results of SUSY searches published by the CMS collaboration based on the first ~1 fb^-1 of data taken during the 2011 LHC run at 7 TeV. The pMSSM is a 19-dimensional parametrization of the MSSM that captures most of its phenomenological features. It encompasses, and goes beyond, a broad range of more constrained SUSY models. Performing a global Bayesian analysis, we obtain posterior probability densities of parameters, masses and derived observables. In contrast to constraints derived for particular SUSY breaking schemes, such as the CMSSM, our results provide more generic conclusions on how the current data constrain the MSSM.
DOI: 10.1109/icmla.2019.00270
2019
Cited 56 times
Anomaly Detection with Conditional Variational Autoencoders
Exploiting the rapid advances in probabilistic inference, in particular variational Bayes and variational autoencoders (VAEs), for anomaly detection (AD) tasks remains an open research question. Previous works argued that training VAE models only with inliers is insufficient and the framework should be significantly modified in order to discriminate the anomalous instances. In this work, we exploit the deep conditional variational autoencoder (CVAE) and we define an original loss function together with a metric that targets hierarchically structured data AD. Our motivating application is a real world problem: monitoring the trigger system which is a basic component of many particle physics experiments at the CERN Large Hadron Collider (LHC). In the experiments we show the superior performance of this method for classical machine learning (ML) benchmarks and for our application.
DOI: 10.1140/epjp/i2019-12710-3
2019
Cited 50 times
Pileup mitigation at the Large Hadron Collider with graph neural networks
At the Large Hadron Collider, the high-transverse-momentum events studied by experimental collaborations occur in coincidence with parasitic low-transverse-momentum collisions, usually referred to as pileup. Pileup mitigation is a key ingredient of the online and offline event reconstruction as pileup affects the reconstruction accuracy of many physics observables. We present a classifier based on Graph Neural Networks, trained to retain particles coming from high-transverse-momentum collisions, while rejecting those coming from pileup collisions. This model is designed as a refinement of the PUPPI algorithm (D. Bertolini et al., JHEP 10, 059 (2014)), employed in many LHC data analyses since 2015. Thanks to an extended basis of input information and the learning capabilities of the considered network architecture, we show an improvement in pileup-rejection performances with respect to state-of-the-art solutions.
DOI: 10.1103/physrevd.102.012010
2020
Cited 47 times
Interaction networks for the identification of boosted <mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" display="inline"><mml:mi>H</mml:mi><mml:mo stretchy="false">→</mml:mo><mml:mi>b</mml:mi><mml:mover accent="true"><mml:mi>b</mml:mi><mml:mo stretchy="false">¯</mml:mo></mml:mover></mml:math> decays
We develop an algorithm based on an interaction network to identify high-transverse-momentum Higgs bosons decaying to bottom quark-antiquark pairs and distinguish them from ordinary jets that reflect the configurations of quarks and gluons at short distances. The algorithm's inputs are features of the reconstructed charged particles in a jet and the secondary vertices associated with them. Describing the jet shower as a combination of particle-to-particle and particle-to-vertex interactions, the model is trained to learn a jet representation on which the classification problem is optimized. The algorithm is trained on simulated samples of realistic LHC collisions, released by the CMS Collaboration on the CERN Open Data Portal. The interaction network achieves a drastic improvement in the identification performance with respect to state-of-the-art algorithms.
DOI: 10.1007/s41781-019-0027-2
2019
Cited 43 times
FPGA-Accelerated Machine Learning Inference as a Service for Particle Physics Computing
Large-scale particle physics experiments face challenging demands for high-throughput computing resources both now and in the future. New heterogeneous computing paradigms on dedicated hardware with increased parallelization, such as Field Programmable Gate Arrays (FPGAs), offer exciting solutions with large potential gains. The growing applications of machine learning algorithms in particle physics for simulation, reconstruction, and analysis are naturally deployed on such platforms. We demonstrate that the acceleration of machine learning inference as a web service represents a heterogeneous computing solution for particle physics experiments that potentially requires minimal modification to the current computing model. As examples, we retrain the ResNet-50 convolutional neural network to demonstrate state-of-the-art performance for top quark jet tagging at the LHC and apply a ResNet-50 model with transfer learning for neutrino event classification. Using Project Brainwave by Microsoft to accelerate the ResNet-50 image classification model, we achieve average inference times of 60 (10) ms with our experimental physics software framework using Brainwave as a cloud (edge or on-premises) service, representing an improvement by a factor of approximately 30 (175) in model inference latency over traditional CPU inference in current experimental hardware. A single FPGA service accessed by many CPUs achieves a throughput of 600–700 inferences per second using an image batch of one, comparable to large batch-size GPU throughput and significantly better than small batch-size GPU throughput. Deployed as an edge or cloud service for the particle physics computing model, coprocessor accelerators can have a higher duty cycle and are potentially much more cost-effective.
DOI: 10.1088/1748-0221/15/05/p05026
2020
Cited 41 times
Fast inference of Boosted Decision Trees in FPGAs for particle physics
We describe the implementation of Boosted Decision Trees in the hls4ml library, which allows the translation of a trained model into FPGA firmware through an automated conversion process. Thanks to its fully on-chip implementation, hls4ml performs inference of Boosted Decision Tree models with extremely low latency. With a typical latency less than 100 ns, this solution is suitable for FPGA-based real-time processing, such as in the Level-1 Trigger system of a collider experiment. These developments open up prospects for physicists to deploy BDTs in FPGAs for identifying the origin of jets, better reconstructing the energies of muons, and enabling better selection of rare signal processes.
DOI: 10.3389/fdata.2022.787421
2022
Cited 23 times
Applications and Techniques for Fast Machine Learning in Science
In this community review report, we discuss applications and techniques for fast machine learning (ML) in science-the concept of integrating powerful ML methods into the real-time experimental data processing loop to accelerate scientific discovery. The material for the report builds on two workshops held by the Fast ML for Science community and covers three main areas: applications for fast ML across a number of scientific domains; techniques for training and implementing performant and resource-efficient ML algorithms; and computing architectures, platforms, and technologies for deploying these algorithms. We also present overlapping challenges across the multiple scientific domains where common solutions can be found. This community report is intended to give plenty of examples and inspiration for scientific discovery through integrated and accelerated ML solutions. This is followed by a high-level overview and organization of technical advances, including an abundance of pointers to source material, which can enable these breakthroughs.
DOI: 10.1038/s42256-022-00441-3
2022
Cited 22 times
Autoencoders on field-programmable gate arrays for real-time, unsupervised new physics detection at 40 MHz at the Large Hadron Collider
To study the physics of fundamental particles and their interactions, the Large Hadron Collider was constructed at CERN, where protons collide to create new particles measured by detectors. Collisions occur at a frequency of 40 MHz, and with an event size of roughly 1 MB it is impossible to read out and store the generated amount of data from the detector and therefore a multi-tiered, real-time filtering system is required. In this paper, we show how to adapt and deploy deep-learning-based autoencoders for the unsupervised detection of new physics signatures in the challenging environment of a real-time event selection system at the Large Hadron Collider. The first-stage filter, implemented on custom electronics, decides within a few microseconds whether an event should be kept or discarded. At this stage, the rate is reduced from 40 MHz to about 100 kHz. We demonstrate the deployment of an unsupervised selection algorithm on this custom electronics, running in as little as 80 ns and enhancing the signal-over-background ratio by three orders of magnitude. This work enables the practical deployment of these networks during the next data-taking campaign of the Large Hadron Collider. The Large Hadron Collider produces 40 million collision events per second, most of which need to be discarded by a real-time filtering system. Unsupervised deep learning algorithms are developed and deployed on custom electronics to search for rare events indicating new physics, rather than for specific events led by theory.
DOI: 10.1140/epjc/s10052-022-10226-y
2022
Cited 22 times
Learning new physics from an imperfect machine
We show how to deal with uncertainties on the Standard Model predictions in an agnostic new physics search strategy that exploits artificial neural networks. Our approach builds directly on the specific Maximum Likelihood ratio treatment of uncertainties as nuisance parameters for hypothesis testing that is routinely employed in high-energy physics. After presenting the conceptual foundations of our method, we first illustrate all aspects of its implementation and extensively study its performances on a toy one-dimensional problem. We then show how to implement it in a multivariate setup by studying the impact of two typical sources of experimental uncertainties in two-body final states at the LHC.
DOI: 10.1088/2632-2153/ad07f7
2023
Cited 7 times
Unravelling physics beyond the standard model with classical and quantum anomaly detection
Abstract Much hope for finding new physics phenomena at microscopic scale relies on the observations obtained from High Energy Physics experiments, like the ones performed at the Large Hadron Collider (LHC). However, current experiments do not indicate clear signs of new physics that could guide the development of additional Beyond Standard Model (BSM) theories. Identifying signatures of new physics out of the enormous amount of data produced at the LHC falls into the class of anomaly detection and constitutes one of the greatest computational challenges. In this article, we propose a novel strategy to perform anomaly detection in a supervised learning setting, based on the artificial creation of anomalies through a random process. For the resulting supervised learning problem, we successfully apply classical and quantum support vector classifiers (CSVC and QSVC respectively) to identify the artificial anomalies among the SM events. Even more promising, we find that employing an SVC trained to identify the artificial anomalies, it is possible to identify realistic BSM events with high accuracy. In parallel, we also explore the potential of quantum algorithms for improving the classification accuracy and provide plausible conditions for the best exploitation of this novel computational paradigm.
DOI: 10.1023/a:1015568724369
2002
Cited 94 times
2007
Cited 84 times
SuperB: A High-Luminosity Asymmetric e+ e- Super Flavor Factory. Conceptual Design Report
The physics objectives of SuperB, an asymmetric electron-positron collider with a luminosity above 10^36/cm^2/s are described, together with the conceptual design of a novel low emittance design that achieves this performance with wallplug power comparable to that of the current B Factories, and an upgraded detector capable of doing the physics in the SuperB environment.
DOI: 10.1186/1754-0410-3-6
2009
Cited 75 times
First evidence of new physics in b ↔ stransitions
We combine all the available experimental information on Bs mixing, including the very recent tagged analyses of Bs to J/Psi phi by the CDF and D0 collaborations. We find that the phase of the Bs mixing amplitude deviates more than 3 sigma from the Standard Model prediction. While no single measurement has a 3 sigma significance yet, all the constraints show a remarkable agreement with the combined result. This is a first evidence of physics beyond the Standard Model. This result disfavours New Physics models with Minimal Flavour Violation with the same significance.
DOI: 10.1109/tns.2021.3087100
2021
Cited 27 times
A Reconfigurable Neural Network ASIC for Detector Front-End Data Compression at the HL-LHC
Despite advances in the programmable logic capabilities of modern trigger systems, a significant bottleneck remains in the amount of data to be transported from the detector to off-detector logic where trigger decisions are made. We demonstrate that a neural network autoencoder model can be implemented in a radiation tolerant ASIC to perform lossy data compression alleviating the data transmission problem while preserving critical information of the detector energy profile. For our application, we consider the high-granularity calorimeter from the CMS experiment at the CERN Large Hadron Collider. The advantage of the machine learning approach is in the flexibility and configurability of the algorithm. By changing the neural network weights, a unique data compression algorithm can be deployed for each sensor in different detector regions, and changing detector or collider conditions. To meet area, performance, and power constraints, we perform a quantization-aware training to create an optimized neural network hardware implementation. The design is achieved through the use of high-level synthesis tools and the hls4ml framework, and was processed through synthesis and physical layout flows based on a LP CMOS 65 nm technology node. The flow anticipates 200 Mrad of ionizing radiation to select gates, and reports a total area of 3.6 mm^2 and consumes 95 mW of power. The simulated energy consumption per inference is 2.4 nJ. This is the first radiation tolerant on-detector ASIC implementation of a neural network that has been designed for particle physics applications.
DOI: 10.1007/jhep02(2022)074
2022
Cited 15 times
Autoencoders for semivisible jet detection
A bstract The production of dark matter particles from confining dark sectors may lead to many novel experimental signatures. Depending on the details of the theory, dark quark production in proton-proton collisions could result in semivisible jets of particles: collimated sprays of dark hadrons of which only some are detectable by particle collider experiments. The experimental signature is characterised by the presence of reconstructed missing momentum collinear with the visible components of the jets. This complex topology is sensitive to detector inefficiencies and mis-reconstruction that generate artificial missing momentum. With this work, we propose a signal-agnostic strategy to reject ordinary jets and identify semivisible jets via anomaly detection techniques. A deep neural autoencoder network with jet substructure variables as input proves highly useful for analyzing anomalous jets. The study focuses on the semivisible jet signature; however, the technique can apply to any new physics model that predicts signatures with anomalous jets from non-SM particles.
DOI: 10.3389/fdata.2022.803685
2022
Cited 15 times
Improving Variational Autoencoders for New Physics Detection at the LHC With Normalizing Flows
We investigate how to improve new physics detection strategies exploiting variational autoencoders and normalizing flows for anomaly detection at the Large Hadron Collider. As a working example, we consider the DarkMachines challenge dataset. We show how different design choices (e.g., event representations, anomaly score definitions, network architectures) affect the result on specific benchmark new physics models. Once a baseline is established, we discuss how to improve the anomaly detection accuracy by exploiting normalizing flow layers in the latent space of the variational autoencoder.
DOI: 10.1103/physrevd.107.076017
2022
Cited 14 times
Evaluating generative models in high energy physics
There has been a recent explosion in research into machine-learning-based generative modeling to tackle computational challenges for simulations in high energy physics (HEP). In order to use such alternative simulators in practice, we need well defined metrics to compare different generative models and evaluate their discrepancy from the true distributions. We present the first systematic review and investigation into evaluation metrics and their sensitivity to failure modes of generative models, using the framework of two-sample goodness-of-fit testing, and their relevance and viability for HEP. Inspired by previous work in both physics and computer vision, we propose two new metrics, the Fr\'echet and kernel physics distances (FPD and KPD), and perform a variety of experiments measuring their performance on simple Gaussian-distributed, and simulated high energy jet datasets. We find FPD, in particular, to be the most sensitive metric to all alternative jet distributions tested and recommend its adoption, along with the KPD and Wasserstein distances between individual feature distributions, for evaluating generative models in HEP. We finally demonstrate the efficacy of these proposed metrics in evaluating and comparing a novel attention-based generative adversarial particle transformer to the state-of-the-art message-passing generative adversarial network jet simulation model.
DOI: 10.48550/arxiv.2301.10780
2023
Cited 6 times
Quantum anomaly detection in the latent space of proton collision events at the LHC
We propose a new strategy for anomaly detection at the LHC based on unsupervised quantum machine learning algorithms. To accommodate the constraints on the problem size dictated by the limitations of current quantum hardware we develop a classical convolutional autoencoder. The designed quantum anomaly detection models, namely an unsupervised kernel machine and two clustering algorithms, are trained to find new-physics events in the latent representation of LHC data produced by the autoencoder. The performance of the quantum algorithms is benchmarked against classical counterparts on different new-physics scenarios and its dependence on the dimensionality of the latent space and the size of the training dataset is studied. For kernel-based anomaly detection, we identify a regime where the quantum model significantly outperforms its classical counterpart. An instance of the kernel machine is implemented on a quantum computer to verify its suitability for available hardware. We demonstrate that the observed consistent performance advantage is related to the inherent quantum properties of the circuit used.
DOI: 10.1088/1742-6596/2438/1/012100
2023
Cited 5 times
Machine Learning for Particle Flow Reconstruction at CMS
We provide details on the implementation of a machine-learning based particle flow algorithm for CMS. The standard particle flow algorithm reconstructs stable particles based on calorimeter clusters and tracks to provide a global event reconstruction that exploits the combined information of multiple detector subsystems, leading to strong improvements for quantities such as jets and missing transverse energy. We have studied a possible evolution of particle flow towards heterogeneous computing platforms such as GPUs using a graph neural network. The machine-learned PF model reconstructs particle candidates based on the full list of tracks and calorimeter clusters in the event. For validation, we determine the physics performance directly in the CMS software framework when the proposed algorithm is interfaced with the offline reconstruction of jets and missing transverse energy. We also report the computational performance of the algorithm, which scales approximately linearly in runtime and memory usage with the input size.
DOI: 10.1016/j.nuclphysb.2005.06.035
2005
Cited 64 times
Upper bounds on rare K and B decays from minimal flavour violation
We study the branching ratios of rare K and B decays in models with minimal flavour violation, using the presently available information from the universal unitarity triangle analysis and from the measurements of Br(B -> X_s gamma), Br(B -> X_s l^+l^-) and Br(K^+ -> pi^+ nu nubar). We find the following upper bounds: Br(K^+ -> pi^+ nu nubar)< 11.9 10^{-11}, Br(K_L -> pi^0 nu nubar)< 4.6 10^{-11}, Br(K_L -> mu mubar)_{SD}< 1.4 10^{-9}, Br(B -> X_s nu nubar)< 5.2 10^{-5}, Br(B -> X_d nu nubar)< 2.2 10^{-6}, Br(B_s -> mu mubar)< 7.4 10^{-9}, Br(B_d -> mu mubar)< 2.2 10^{-10} at 95 % probability. We analyze in detail various possible scenarios with positive or negative interference of Standard Model and New Physics contributions, and show how an improvement of experimental data corresponding to the projected 2010 B factory integrated luminosities will allow to disentangle and test these different possibilities. Finally, anticipating that subsequently the leading role in constraining this kind of new physics will be taken over by the rare decays K^+ -> pi^+ nu nubar, K_L -> pi^0 nu nubar and B_{s,d} -> mu mubar, that are dominated by the Z^0 -penguin function C, we also present plots for several branching ratios as functions of C . We point out an interesting triple correlation between K^+ -> pi^+ nu nubar, B -> X_s gamma and B -> X_s l^+l^- present in MFV models.
DOI: 10.1088/1742-6596/1525/1/012081
2020
Cited 27 times
Particle Generative Adversarial Networks for full-event simulation at the LHC and their application to pileup description
We investigate how a Generative Adversarial Network could be used to generate a list of particle four-momenta from LHC proton collisions, allowing one to define a generative model that could abstract from the irregularities of typical detector geometries. As an example of application, we show how such an architecture could be used as a generator of LHC parasitic collisions (pileup). We present two approaches to generate the events: unconditional generator and generator conditioned on missing transverse energy. We assess generation performances in a realistic LHC data-analysis environment, with a pileup mitigation algorithm applied.
DOI: 10.1038/s41597-022-01187-8
2022
Cited 13 times
LHC physics dataset for unsupervised New Physics detection at 40 MHz
Abstract In the particle detectors at the Large Hadron Collider, hundreds of millions of proton-proton collisions are produced every second. If one could store the whole data stream produced in these collisions, tens of terabytes of data would be written to disk every second. The general-purpose experiments ATLAS and CMS reduce this overwhelming data volume to a sustainable level, by deciding in real-time whether each collision event should be kept for further analysis or be discarded. We introduce a dataset of proton collision events that emulates a typical data stream collected by such a real-time processing system, pre-filtered by requiring the presence of at least one electron or muon. This dataset could be used to develop novel event selection strategies and assess their sensitivity to new phenomena. In particular, we intend to stimulate a community-based effort towards the design of novel algorithms for performing unsupervised new physics detection, customized to fit the bandwidth, latency and computational resource constraints of the real-time event selection system of a typical particle detector.
DOI: 10.22323/1.398.0500
2022
Cited 12 times
Unitarity Triangle global fits beyond the Standard Model: UTfit 2021 new physics update
Flavour physics represents a unique test bench for the Standard Model (SM).New analyses performed at the LHC experiments are now providing unprecedented insights into CKM metrology and new evidences for rare decays.The CKM picture can provide very precise SM predictions through global analyses.We present here the results of the latest results from UTfit new physics analysis: the Unitarity Triangle (UT) analysis can be used to constrain the parameter space in possible new physics (NP) scenarios.All of the available experimental and theoretical information on Δ = 2 processes is reinterpreted including a model-independent NP parametrisation.We determine the allowed NP contributions in the kaon, , , and sectors and, in various NP scenarios, we translate them into bounds for the NP scale as a function of NP couplings.
DOI: 10.1088/1126-6708/2006/03/080
2006
Cited 49 times
The UTfit collaboration report on the status of the unitarity triangle beyond the Standard Model I. Model-independent analysis and minimal flavour violation
Starting from a (new physics independent) tree level determination of and , we perform the Unitarity Triangle analysis in general extensions of the Standard Model with arbitrary new physics contributions to loop-mediated processes. Using a simple parameterization, we determine the allowed ranges of non-standard contributions to |ΔF| = 2 processes. Remarkably, the recent measurements from B factories allow us to determine with good precision the shape of the Unitarity Triangle even in the presence of new physics, and to derive stringent constraints on non-standard contributions to |ΔF| = 2 processes. Since the present experimental constraints favour models with Minimal Flavour Violation, we present the determination of the Universal Unitarity Triangle that can be defined in this class of extensions of the Standard Model. Finally, we perform a combined fit of the Unitarity Triangle and of new physics contributions in Minimal Flavour Violation, reaching a sensitivity to a new physics scale of about 5 TeV. We also extrapolate all these analyses into a ``year 2010'' scenario for experimental and theoretical inputs in the flavour sector. All the results presented in this paper are also available at the URL http://www.utfit.org, where they are continuously updated.
DOI: 10.1103/physrevd.78.075008
2008
Cited 47 times
Missing energy look-alikes with<mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" display="inline"><mml:mn>100</mml:mn><mml:mtext> </mml:mtext><mml:mtext> </mml:mtext><mml:msup><mml:mi>pb</mml:mi><mml:mrow><mml:mo>−</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msup></mml:math>at the CERN LHC
A missing energy discovery is possible at the LHC with the first $100\text{ }\text{ }{\mathrm{pb}}^{\ensuremath{-}1}$ of understood data. We present a realistic strategy to rapidly narrow the list of candidate theories at, or close to, the moment of discovery. The strategy is based on robust ratios of inclusive counts of simple physics objects. We study specific cases showing discrimination of look-alike models in simulated data sets that are at least 10 to 100 times smaller than used in previous studies. We discriminate supersymmetry models from nonsupersymmetric look-alikes with only $100\text{ }\text{ }{\mathrm{pb}}^{\ensuremath{-}1}$ of simulated data, using combinations of observables that trace back to differences in spin.
DOI: 10.48550/arxiv.1402.1664
2014
Cited 34 times
The UTfit Collaboration Average of $D$ meson mixing data: Winter 2014
We update the analysis of $D$ meson mixing including the latest experimental results as of January 2014. We derive constraints on the parameters $M_{12}$, $\Gamma_{12}$ and $\Phi_{12}$ that describe $D$ meson mixing using all available data, allowing for CP violation. We also provide posterior distributions for observable parameters appearing in $D$ physics.
DOI: 10.1088/2632-2153/ac7c56
2022
Cited 10 times
Particle-based fast jet simulation at the LHC with variational autoencoders
We study how to use deep variational autoencoders (VAEs) for a fast simulation of jets of particles at the Large Hadron Collider. We represent jets as a list of constituents, characterized by their momenta. Starting from a simulation of the jet before detector effects, we train a deep VAE to return the corresponding list of constituents after detection. Doing so, we bypass both the time-consuming detector simulation and the collision reconstruction steps of a traditional processing chain, speeding up significantly the events generation workflow. Through model optimization and hyperparameter tuning, we achieve state-of-the-art precision on the jet four-momentum, while providing an accurate description of the constituents momenta, and an inference time comparable to that of a rule-based fast simulation.
DOI: 10.1103/physrevd.74.051301
2006
Cited 43 times
New bounds on the Cabibbo-Kobayashi-Maskawa matrix from<mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" display="inline"><mml:mi>B</mml:mi><mml:mo>→</mml:mo><mml:mi>K</mml:mi><mml:mi>π</mml:mi><mml:mi>π</mml:mi></mml:math>Dalitz plot analyses
We present a new technique to extract information on the unitarity triangle from the study of $B\ensuremath{\rightarrow}K\ensuremath{\pi}\ensuremath{\pi}$ Dalitz plots. Using the sensitivity of Dalitz analyses to the absolute values and the phases of decay amplitudes and isospin symmetry, we obtain a new constraint on the elements of the CKM matrix. We discuss in detail the role of electroweak penguin contributions and outline future prospects.
DOI: 10.1007/s41781-019-0028-1
2019
Cited 21 times
Topology Classification with Deep Learning to Improve Real-Time Event Selection at the LHC
We show how an event topology classification based on deep learning could be used to improve the purity of data samples selected in real time at the Large Hadron Collider. We consider different data representations, on which different kinds of multi-class classifiers are trained. Both raw data and high-level features are utilized. In the considered examples, a filter based on the classifier’s score can be trained to retain $$\sim 99\%$$ of the interesting events and reduce the false-positive rate by more than one order of magnitude. By operating such a filter as part of the online event selection infrastructure of the LHC experiments, one could benefit from a more flexible and inclusive selection strategy while reducing the amount of downstream resources wasted in processing false positives. The saved resources could translate into a reduction of the detector operation cost or into an effective increase of storage and processing capabilities, which could be reinvested to extend the physics reach of the LHC experiments.
2020
Cited 18 times
Automatic deep heterogeneous quantization of Deep Neural Networks for ultra low-area, low-latency inference on the edge at particle colliders
While the quest for more accurate solutions is pushing deep learning research towards larger and more complex algorithms, edge devices demand efficient inference i.e. reduction in model size, latency and energy consumption. A technique to limit model size is quantization, i.e. using fewer bits to represent weights and biases. Such an approach usually results in a decline in performance. Here, we introduce a novel method for designing optimally heterogeneously quantized versions of deep neural network models for minimum-energy, high-accuracy, nanosecond inference and fully automated deployment on chip. With a per-layer, per-parameter type automatic quantization procedure, sampling from a wide range of quantizers, model energy consumption and size are minimized while high accuracy is maintained. This is crucial for the event selection procedure in proton-proton collisions at the CERN Large Hadron Collider, where resources are strictly limited and a latency of ${\mathcal O}(1)~\mu$s is required. Nanosecond inference and a resource consumption reduced by a factor of $50$ when implemented on FPGA hardware is achieved.
2021
Cited 14 times
The LHC Olympics 2020: A Community Challenge for Anomaly Detection in High Energy Physics
A new paradigm for data-driven, model-agnostic new physics searches at colliders is emerging, and aims to leverage recent breakthroughs in anomaly detection and machine learning. In order to develop and benchmark new anomaly detection methods within this framework, it is essential to have standard datasets. To this end, we have created the LHC Olympics 2020, a community challenge accompanied by a set of simulated collider events. Participants in these Olympics have developed their methods using an R&D dataset and then tested them on black boxes: datasets with an unknown anomaly (or not). This paper will review the LHC Olympics 2020 challenge, including an overview of the competition, a description of methods deployed in the competition, lessons learned from the experience, and implications for data analyses with future datasets as well as future colliders.
DOI: 10.1140/epjc/s10052-022-10830-y
2022
Cited 9 times
Learning new physics efficiently with nonparametric methods
Abstract We present a machine learning approach for model-independent new physics searches. The corresponding algorithm is powered by recent large-scale implementations of kernel methods, nonparametric learning algorithms that can approximate any continuous function given enough data. Based on the original proposal by D’Agnolo and Wulzer (Phys Rev D 99(1):015014, 2019, arXiv:1806.02350 [hep-ph]), the model evaluates the compatibility between experimental data and a reference model, by implementing a hypothesis testing procedure based on the likelihood ratio. Model-independence is enforced by avoiding any prior assumption about the presence or shape of new physics components in the measurements. We show that our approach has dramatic advantages compared to neural network implementations in terms of training times and computational resources, while maintaining comparable performances. In particular, we conduct our tests on higher dimensional datasets, a step forward with respect to previous studies.
DOI: 10.1103/physrevlett.100.031802
2008
Cited 32 times
<mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" display="inline"><mml:msub><mml:mi>B</mml:mi><mml:mi>s</mml:mi></mml:msub><mml:mo>→</mml:mo><mml:msup><mml:mi>K</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mo>*</mml:mo><mml:mo stretchy="false">)</mml:mo><mml:mn>0</mml:mn></mml:mrow></mml:msup><mml:msup><mml:mover accent="true"><mml:mi>K</mml:mi><mml:mo>¯</mml:mo></mml:mover><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mo>*</mml:mo><mml:mo stretchy="false">)</mml:mo><…
We point out that time-dependent $CP$ asymmetries in ${B}_{s}\ensuremath{\rightarrow}{K}^{*0}{\overline{K}}^{*0}$ decays probe the presence of new physics in $b\ensuremath{\rightarrow}s$ transitions with an unprecedented theoretical accuracy. We show that, contrary to the case of ${B}_{d}\ensuremath{\rightarrow}\ensuremath{\phi}{K}_{S}$, it is possible to obtain a model-independent prediction for the coefficient $S({B}_{s}\ensuremath{\rightarrow}{K}^{*0}{\overline{K}}^{*0})$ in the standard model. We give an estimate of the experimental precision achievable with the next generation of $B$ physics experiments. We also discuss how this approach can be extended to the case of ${B}_{s}\ensuremath{\rightarrow}{\overline{K}}^{*0}{K}^{0}$, ${B}_{s}\ensuremath{\rightarrow}{K}^{*0}{\overline{K}}^{0}$, and ${B}_{s}\ensuremath{\rightarrow}{K}^{0}{\overline{K}}^{0}$ decays and the different experimental challenges for these channels.
DOI: 10.1007/jhep03(2014)123
2014
Cited 24 times
The UTfit collaboration average of D meson mixing data: Winter 2014
We update the analysis of D meson mixing including the latest experimental results as of January 2014. We derive constraints on the parameters M 12, Γ12 and Φ12 that describe D meson mixing using all available data, allowing for CP violation. We also provide posterior distributions for observable parameters appearing in D physics.
DOI: 10.1016/j.nuclphysbps.2015.09.361
2016
Cited 20 times
Update of the electroweak precision fit, interplay with Higgs-boson signal strengths and model-independent constraints on new physics
We present updated global fits of the Standard Model and beyond to electroweak precision data, taking into account recent progress in theoretical calculations and experimental measurements. From the fits, we derive model-independent constraints on new physics by introducing oblique and epsilon parameters, and modified Zbb‾ and HVV couplings. Furthermore, we also perform fits of the scale factors of the Higgs-boson couplings to observed signal strengths of the Higgs boson.
2020
Cited 16 times
Ultra Low-latency, Low-area Inference Accelerators using Heterogeneous Deep Quantization with QKeras and hls4ml
In this paper, we introduce the QKeras library, an extension of the Keras library allowing for the creation of heterogeneously quantized versions of deep neural network models, through drop-in replacement of Keras layers. These models are trained quantization-aware, where the user can trade off model area or energy consumption by accuracy. We demonstrate how the reduction of numerical precision, through quantization-aware training, significantly reduces resource consumption while retaining high accuracy when implemented on FPGA hardware. Together with the hls4ml library, this allows for a fully automated deployment of quantized Keras models on chip, crucial for ultra low-latency inference. As a benchmark problem, we consider a classification task for the triggering of events in proton-proton collisions at the CERN Large Hadron Collider, where a latency of ${\mathcal O}(1)~\mu$s is required.
DOI: 10.1140/epjc/s10052-020-8230-1
2020
Cited 15 times
The DNNLikelihood: enhancing likelihood distribution with Deep Learning
We introduce the DNNLikelihood, a novel framework to easily encode, through Deep Neural Networks (DNN), the full experimental information contained in complicated likelihood functions (LFs). We show how to efficiently parametrise the LF, treated as a multivariate function of parameters and nuisance parameters with high dimensionality, as an interpolating function in the form of a DNN predictor. We do not use any Gaussian approximation or dimensionality reduction, such as marginalisation or profiling over nuisance parameters, so that the full experimental information is retained. The procedure applies to both binned and unbinned LFs, and allows for an efficient distribution to multiple software platforms, e.g. through the framework-independent ONNX model format. The distributed DNNLikelihood can be used for different use cases, such as re-sampling through Markov Chain Monte Carlo techniques, possibly with custom priors, combination with other LFs, when the correlations among parameters are known, and re-interpretation within different statistical approaches, i.e. Bayesian vs frequentist. We discuss the accuracy of our proposal and its relations with other approximation techniques and likelihood distribution frameworks. As an example, we apply our procedure to a pseudo-experiment corresponding to a realistic LHC search for new physics already considered in the literature.
DOI: 10.48550/arxiv.2012.01563
2020
Cited 15 times
Accelerated Charged Particle Tracking with Graph Neural Networks on FPGAs
We develop and study FPGA implementations of algorithms for charged particle tracking based on graph neural networks. The two complementary FPGA designs are based on OpenCL, a framework for writing programs that execute across heterogeneous platforms, and hls4ml, a high-level-synthesis-based compiler for neural network to firmware conversion. We evaluate and compare the resource usage, latency, and tracking performance of our implementations based on a benchmark dataset. We find a considerable speedup over CPU-based execution is possible, potentially enabling such algorithms to be used effectively in future computing workflows and the FPGA-based Level-1 trigger at the CERN Large Hadron Collider.
DOI: 10.1142/12200
2020
Cited 15 times
Artificial Intelligence for High Energy Physics
la diffusion de documents scientifiques de niveau recherche, publiés ou non, émanant des établissements d'enseignement et de recherche français ou étrangers, des laboratoires publics ou privés.
DOI: 10.1109/asap52443.2021.00025
2021
Cited 13 times
Accelerating Recurrent Neural Networks for Gravitational Wave Experiments
This paper presents novel reconfigurable architectures for reducing the latency of recurrent neural networks (RNNs) that are used for detecting gravitational waves. Gravitational interferometers such as the LIGO detectors capture cosmic events such as black hole mergers which happen at unknown times and of varying durations, producing time-series data. We have developed a new architecture capable of accelerating RNN inference for analyzing time-series data from LIGO detectors. This architecture is based on optimizing the initiation intervals (II) in a multi-layer LSTM (Long Short-Term Memory) network, by identifying appropriate reuse factors for each layer. A customizable template for this architecture has been designed, which enables the generation of low-latency FPGA designs with efficient resource utilization using high-level synthesis tools. The proposed approach has been evaluated based on two LSTM models, targeting a ZYNQ 7045 FPGA and a U250 FPGA. Experimental results show that with balanced II, the number of DSPs can be reduced up to 42% while achieving the same IIs. When compared to other FPGA-based LSTM designs, our design can achieve about 4.92 to 12.4 times lower latency.
DOI: 10.48550/arxiv.2106.11535
2021
Cited 13 times
Particle Cloud Generation with Message Passing Generative Adversarial Networks
In high energy physics (HEP), jets are collections of correlated particles produced ubiquitously in particle collisions such as those at the CERN Large Hadron Collider (LHC). Machine learning (ML)-based generative models, such as generative adversarial networks (GANs), have the potential to significantly accelerate LHC jet simulations. However, despite jets having a natural representation as a set of particles in momentum-space, a.k.a. a particle cloud, there exist no generative models applied to such a dataset. In this work, we introduce a new particle cloud dataset (JetNet), and apply to it existing point cloud GANs. Results are evaluated using (1) 1-Wasserstein distances between high- and low-level feature distributions, (2) a newly developed Fr\'{e}chet ParticleNet Distance, and (3) the coverage and (4) minimum matching distance metrics. Existing GANs are found to be inadequate for physics applications, hence we develop a new message passing GAN (MPGAN), which outperforms existing point cloud GANs on virtually every metric and shows promise for use in HEP. We propose JetNet as a novel point-cloud-style dataset for the ML community to experiment with, and set MPGAN as a benchmark to improve upon for future generative models. Additionally, to facilitate research and improve accessibility and reproducibility in this area, we release the open-source JetNet Python package with interfaces for particle cloud datasets, implementations for evaluation and loss metrics, and more tools for ML in HEP development.
DOI: 10.1140/epjp/s13360-024-05028-y
2024
Autoencoders for real-time SUEP detection
Abstract Confining dark sectors with pseudo-conformal dynamics can produce Soft Unclustered Energy Patterns (SUEP), at the Large Hadron Collider: the production of dark quarks in proton–proton collisions leading to a dark shower and the high-multiplicity production of dark hadrons. The final experimental signature is spherically symmetric energy deposits by an anomalously large number of soft Standard Model particles with a transverse energy of O(100) $$\,\text {MeV}$$ <mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML"> <mml:mrow> <mml:mspace /> <mml:mtext>MeV</mml:mtext> </mml:mrow> </mml:math> . Assuming Yukawa-like couplings of the scalar portal state, the dominant production mode is gluon fusion, and the dominant background comes from multi-jet QCD events. We have developed a deep learning-based Anomaly Detection technique to reject QCD jets and identify any anomalous signature, including SUEP, in real-time in the High-Level Trigger system of experiments like the Compact Muon Solenoid at the Large Hadron Collider. A deep convolutional neural autoencoder network has been trained using QCD events by taking transverse energy deposits in the inner tracker, electromagnetic calorimeter, and hadron calorimeter sub-detectors as 3-channel image data. Due to the sparse nature of the data, only $$\sim $$ <mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML"> <mml:mo>∼</mml:mo> </mml:math> 0.5% of the total $$\sim $$ <mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML"> <mml:mo>∼</mml:mo> </mml:math> $${300}\,\textrm{k}$$ <mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML"> <mml:mrow> <mml:mn>300</mml:mn> <mml:mspace /> <mml:mtext>k</mml:mtext> </mml:mrow> </mml:math> image pixels have nonzero values. To tackle this challenge, a nonstandard loss function, the inverse of the so-called Dice Loss, is exploited. The trained autoencoder with learned spatial features of QCD jets can detect 40% of the SUEP events, with a QCD event mistagging rate as low as 2%. The model inference time has been measured using the $$^{\texttt {TM}}$$ <mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML"> <mml:msup> <mml:mrow /> <mml:mi>TM</mml:mi> </mml:msup> </mml:math> processor and found to be $$\sim $$ <mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML"> <mml:mo>∼</mml:mo> </mml:math> $${20}\textrm{ms}$$ <mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML"> <mml:mrow> <mml:mn>20</mml:mn> <mml:mtext>ms</mml:mtext> </mml:mrow> </mml:math> , which perfectly satisfies the High-Level Trigger system’s latency of $$\mathcal {O}(10^2)~\textrm{ms}$$ <mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML"> <mml:mrow> <mml:mi>O</mml:mi> <mml:mo>(</mml:mo> <mml:msup> <mml:mn>10</mml:mn> <mml:mn>2</mml:mn> </mml:msup> <mml:mo>)</mml:mo> <mml:mspace /> <mml:mtext>ms</mml:mtext> </mml:mrow> </mml:math> . Given the virtue of the unsupervised learning of the autoencoders, the trained model can be applied to any new physics model that predicts an experimental signature anomalous to QCD jets.
DOI: 10.22323/1.457.0007
2024
Overview and theoretical prospects for CKM matrix and CP violation from the UTfit Collaboration
Unitarity Triangle updateChecking the usual tensions.. sin2 exp = 0.689 ± 0.019 sin2 UTfit = 0.739 ± 0.027 Vub exp = (3.75 ± 0.26) • 10 -3 Vub UTfit = (3.70 ± 0.09) • 10 -3 Vcb exp = (41.32± 0.73) • 10 -3 Vcb UTfit = (42.21± 0.51) • 10 -3 x = exclusive * = inclusive 17 Unitarity Triangle update ing the usual tensions.. 689 ± 0.019 .739± 0.027 Vub exp = (3.75 ± 0.26) • 10 -3 Vub UTfit = (3.70 ± 0.09) • 10 -3 Vcb exp = (41.32± 0.73) • 10 -3 Vcb UTfit = (42.21± 0.51) • 10 -3 x = exclusive * = inclusive Marcella Bona Unitarit Checking the usual tensions.. sin2 exp = 0.689 ± 0.019 sin2 UTfit = 0.739 ± 0.027 Vub exp = (3.75 ± 0 Vub UTfit = (3.70 ± 0 Vcb exp = (41.32± 0.73) • 10 -3 Vcb UTfit = (42.21± 0.51) • 10 -3 x = exclusive * = inclusive 17 Marcella Bona Unitarity Triangle update Checking the usual tensions.. sin2 exp = 0.689 ± 0.019 sin2 UTfit = 0.739 ± 0.027 Vub exp = (3.75 ± 0.26) • 10 -3 Vub UTfit = (3.70 ± 0.09) • 10 -3 Vcb exp = (41.32± 0.73) • 10 -3 Vcb UTfit = (42.21± 0.51) • 10 -3 x = exclusive * = inclusive Marcella Bona Unitarity Checking the usual tensions..sin2 exp = 0.689 ± 0.019 sin2 UTfit = 0.739 ± 0.027 Vub exp = (3.75 ± 0 Vub UTfit = (3.70 ± 0 Vcb exp = (41.32± 0.73) • 10 -3 Vcb UTfit = (42.21± 0.51) • 10 -3 x = exclusive * = inclusive
DOI: 10.1016/j.physletb.2009.03.011
2009
Cited 26 times
Searching for new physics with <mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" altimg="si1.gif" overflow="scroll"><mml:mi>B</mml:mi><mml:mo>→</mml:mo><mml:mi>K</mml:mi><mml:mi>π</mml:mi></mml:math> decays
We propose a method to quantify the Standard Model uncertainty in B→Kπ decays using the experimental data, assuming that power counting provides a reasonable estimate of the subleading terms in the 1/mb expansion. Using this method, we show that present B→Kπ data are compatible with the Standard Model. We analyze the pattern of subleading terms required to reproduce the B→Kπ data and argue that anomalously large subleading terms are not needed. Finally, we find that SKSπ0 is fairly insensitive to hadronic uncertainties and obtain the Standard Model estimate SKSπ0=0.74±0.04.
DOI: 10.2172/1415022
2017
Cited 17 times
Physics at a 100 TeV pp Collider: Standard Model Processes
This report summarises the properties of Standard Model processes at the 100 TeV pp collider. We document the production rates and typical distributions for a number of benchmark Standard Model processes, and discuss new dynamical phenomena arising at the highest energies available at this collider. We discuss the intrinsic physics interest in the measurement of these Standard Model processes, as well as their role as backgrounds for New Physics searches.
DOI: 10.1007/s41781-021-00060-4
2021
Cited 11 times
Analysis-Specific Fast Simulation at the LHC with Deep Learning
We present a fast-simulation application based on a deep neural network, designed to create large analysis-specific datasets. Taking as an example the generation of W + jet events produced in s= 13 TeV proton-proton collisions, we train a neural network to model detector resolution effects as a transfer function acting on an analysis-specific set of relevant features, computed at generation level, i.e., in absence of detector effects. Based on this model, we propose a novel fast-simulation workflow that starts from a large amount of generator-level events to deliver large analysis-specific samples. The adoption of this approach would result in about an order-of-magnitude reduction in computing and storage requirements for the collision simulation workflow. This strategy could help the high energy physics community to face the computing challenges of the future High-Luminosity LHC.
DOI: 10.1051/epjconf/202125103072
2021
Cited 11 times
Multi-particle reconstruction in the High Granularity Calorimeter using object condensation and graph neural networks
The high-luminosity upgrade of the LHC will come with unprecedented physics and computing challenges. One of these challenges is the accurate reconstruction of particles in events with up to 200 simultaneous protonproton interactions. The planned CMS High Granularity Calorimeter offers fine spatial resolution for this purpose, with more than 6 million channels, but also poses unique challenges to reconstruction algorithms aiming to reconstruct individual particle showers. In this contribution, we propose an end-to-end machine-learning method that performs clustering, classification, and energy and position regression in one step while staying within memory and computational constraints. We employ GravNet, a graph neural network, and an object condensation loss function to achieve this task. Additionally, we propose a method to relate truth showers to reconstructed showers by maximising the energy weighted intersection over union using maximal weight matching. Our results show the efficiency of our method and highlight a promising research direction to be investigated further.
DOI: 10.1140/epjc/s10052-022-10665-7
2022
Cited 7 times
End-to-end multi-particle reconstruction in high occupancy imaging calorimeters with graph neural networks
Abstract We present an end-to-end reconstruction algorithm to build particle candidates from detector hits in next-generation granular calorimeters similar to that foreseen for the high-luminosity upgrade of the CMS detector. The algorithm exploits a distance-weighted graph neural network, trained with object condensation, a graph segmentation technique. Through a single-shot approach, the reconstruction task is paired with energy regression. We describe the reconstruction performance in terms of efficiency as well as in terms of energy resolution. In addition, we show the jet reconstruction performance of our method and discuss its inference computational cost. To our knowledge, this work is the first-ever example of single-shot calorimetric reconstruction of $${\mathcal {O}}(1000)$$ <mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML"> <mml:mrow> <mml:mi>O</mml:mi> <mml:mo>(</mml:mo> <mml:mn>1000</mml:mn> <mml:mo>)</mml:mo> </mml:mrow> </mml:math> particles in high-luminosity conditions with 200 pileup.
DOI: 10.1088/1742-6596/2438/1/012090
2023
GNN-based end-to-end reconstruction in the CMS Phase 2 High-Granularity Calorimeter
Abstract We present the current stage of research progress towards a one-pass, completely Machine Learning (ML) based imaging calorimeter reconstruction. The model used is based on Graph Neural Networks (GNNs) and directly analyzes the hits in each HGCAL endcap. The ML algorithm is trained to predict clusters of hits originating from the same incident particle by labeling the hits with the same cluster index. We impose simple criteria to assess whether the hits associated as a cluster by the prediction are matched to those hits resulting from any particular individual incident particles. The algorithm is studied by simulating two tau leptons in each of the two HGCAL endcaps, where each tau may decay according to its measured standard model branching probabilities. The simulation includes the material interaction of the tau decay products which may create additional particles incident upon the calorimeter. Using this varied multiparticle environment we can investigate the application of this reconstruction technique and begin to characterize energy containment and performance.
DOI: 10.1145/3640464
2024
LL-GNN: Low Latency Graph Neural Networks on FPGAs for High Energy Physics
This work presents a novel reconfigurable architecture for Low Latency Graph Neural Network (LL-GNN) designs for particle detectors, delivering unprecedented low latency performance. Incorporating FPGA-based GNNs into particle detectors presents a unique challenge since it requires sub-microsecond latency to deploy the networks for online event selection with a data rate of hundreds of terabytes per second in the Level-1 triggers at the CERN Large Hadron Collider experiments. This article proposes a novel outer-product based matrix multiplication approach, which is enhanced by exploiting the structured adjacency matrix and a column-major data layout. In addition, we propose a custom code transformation for the matrix multiplication operations, which leverages the structured sparsity patterns and binary features of adjacency matrices to reduce latency and improve hardware efficiency. Moreover, a fusion step is introduced to further reduce the end-to-end design latency by eliminating unnecessary boundaries. Furthermore, a GNN-specific algorithm-hardware co-design approach is presented which not only finds a design with a much better latency but also finds a high accuracy design under given latency constraints. To facilitate this, a customizable template for this low latency GNN hardware architecture has been designed and open-sourced, which enables the generation of low-latency FPGA designs with efficient resource utilization using a high-level synthesis tool. Evaluation results show that our FPGA implementation is up to 9.0 times faster and achieves up to 13.1 times higher power efficiency than a GPU implementation. Compared to the previous FPGA implementations, this work achieves 6.51 to 16.7 times lower latency. Moreover, the latency of our FPGA design is sufficiently low to enable deployment of GNNs in a sub-microsecond, real-time collider trigger system, enabling it to benefit from improved accuracy. The proposed LL-GNN design advances the next generation of trigger systems by enabling sophisticated algorithms to process experimental data efficiently.
DOI: 10.48550/arxiv.2402.01876
2024
Sets are all you need: Ultrafast jet classification on FPGAs for HL-LHC
We study various machine learning based algorithms for performing accurate jet flavor classification on field-programmable gate arrays and demonstrate how latency and resource consumption scale with the input size and choice of algorithm. These architectures provide an initial design for models that could be used for tagging at the CERN LHC during its high-luminosity phase. The high-luminosity upgrade will lead to a five-fold increase in its instantaneous luminosity for proton-proton collisions and, in turn, higher data volume and complexity, such as the availability of jet constituents. Through quantization-aware training and efficient hardware implementations, we show that O(100) ns inference of complex architectures such as deep sets and interaction networks is feasible at a low computational resource cost.
DOI: 10.1051/epjconf/202429509036
2024
Symbolic Regression on FPGAs for Fast Machine Learning Inference
The high-energy physics community is investigating the potential of deploying machine-learning-based solutions on Field-Programmable Gate Arrays (FPGAs) to enhance physics sensitivity while still meeting data processing time constraints. In this contribution, we introduce a novel end-to-end procedure that utilizes a machine learning technique called symbolic regression (SR). It searches the equation space to discover algebraic relations approximating a dataset. We use PySR (a software to uncover these expressions based on an evolutionary algorithm) and extend the functionality of hls4ml (a package for machine learning inference in FPGAs) to support PySR-generated expressions for resource-constrained production environments. Deep learning models often optimize the top metric by pinning the network size because the vast hyperparameter space prevents an extensive search for neural architecture. Conversely, SR selects a set of models on the Pareto front, which allows for optimizing the performance-resource trade-off directly. By embedding symbolic forms, our implementation can dramatically reduce the computational resources needed to perform critical tasks. We validate our method on a physics benchmark: the multiclass classification of jets produced in simulated proton-proton collisions at the CERN Large Hadron Collider. We show that our approach can approximate a 3-layer neural network using an inference model that achieves up to a 13-fold decrease in execution time, down to 5 ns, while still preserving more than 90% approximation accuracy.
DOI: 10.23731/cyrm-2017-003.1
2017
Cited 15 times
Physics at a 100 TeV pp Collider: Standard Model Processes
This report summarises the properties of Standard Model processes at the 100 TeV pp collider. We document the production rates and typical distributions for a number of benchmark Standard Model processes, and discuss new dynamical phenomena arising at the highest energies available at this collider. We discuss the intrinsic physics interest in the measurement of these Standard Model processes, as well as their role as backgrounds for New Physics searches.
DOI: 10.1051/epjconf/201921406025
2019
Cited 14 times
Large-Scale Distributed Training Applied to Generative Adversarial Networks for Calorimeter Simulation
In recent years, several studies have demonstrated the benefit of using deep learning to solve typical tasks related to high energy physics data taking and analysis. In particular, generative adversarial networks are a good candidate to supplement the simulation of the detector response in a collider environment. Training of neural network models has been made tractable with the improvement of optimization methods and the advent of GP-GPU well adapted to tackle the highly-parallelizable task of training neural nets. Despite these advancements, training of large models over large data sets can take days to weeks. Even more so, finding the best model architecture and settings can take many expensive trials. To get the best out of this new technology, it is important to scale up the available network-training resources and, consequently, to provide tools for optimal large-scale distributed training. In this context, our development of a new training workflow, which scales on multi-node/multi-GPU architectures with an eye to deployment on high performance computing machines is described. We describe the integration of hyper parameter optimization with a distributed training framework using Message Passing Interface, for models defined in keras [12] or pytorch [13]. We present results on the speedup of training generative adversarial networks trained on a data set composed of the energy deposition from electron, photons, charged and neutral hadrons in a fine grained digital calorimeter.
DOI: 10.1007/jhep04(2016)155
2016
Cited 14 times
Combination of Run-1 exotic searches in diboson final states at the LHC
We perform a statistical combination of the ATLAS and CMS results for the search of a heavy resonance decaying to a pair of vector bosons with the $$ \sqrt{s}=8 $$ TeV datasets collected at the LHC. We take into account six searches in hadronic and semileptonic final states carried out by the two collaborations. We consider only public information provided by ATLAS and CMS in the HEPDATA database and in papers published in refereed journals. We interpret the combined results within the context of a few benchmark new physics models, such as models predicting the existence of a W′ or a bulk Randall-Sundrum spin-2 resonance, for which we present exclusion limits, significances, p-values and best-fit cross sections. A heavy diboson resonance with a production cross section of ∼4-5 fb and mass between 1.9 and 2.0 TeV is the exotic scenario most consistent with the experimental results. Models in which a heavy resonance decays preferentially to a WW final state are disfavoured.
DOI: 10.22323/1.398.0512
2022
Cited 6 times
Unitarity Triangle global fits testing the Standard Model: UTfit 2021 Standard Model update
Flavour physics represents a unique test bench for the Standard Model (SM).New analyses performed at the LHC experiments are now providing unprecedented insights into CKM metrology and new evidences for rare decays.The CKM picture can provide very precise SM predictions through global analyses.We present here the results of the latest global SM analysis performed by the UTfit collaboration including all the most updated inputs from experiments, lattice QCD and phenomenological calculations.
DOI: 10.1016/j.physletb.2006.12.043
2007
Cited 20 times
Hunting the CKM weak phase with time-integrated Dalitz analyses of decays
We present a new technique to extract information on the unitarity triangle from the study of Bs→Kππ Dalitz plot. Using isospin symmetry and the possibility to access the decay amplitudes from Dalitz analyses, we propose a new strategy to extract the weak phase γ from Bs→Kππ.
DOI: 10.1016/j.nuclphysbps.2013.06.015
2013
Cited 15 times
Standard Model updates and new physics analysis with the Unitarity Triangle fit
We present the summer 2012 update of the Unitarity Triangle (UT) analysis performed by the UTfit Collaboration within the Standard Model (SM) and beyond. The increased accuracy on several of the fundamental constraints is now enhancing some of the tensions amongst and within the constraint themselves. In particular, the long standing tension between exclusive and inclusive determinations of the Vub and Vcb CKM matrix elements is now playing a major role. Then we present the generalisation the UT analysis to investigate new physics (NP) effects, updating the constraints on NP contributions to ΔF=2 processes. In the NP analysis, both CKM and NP parameters are fitted simultaneously to obtain the possible NP effects in any specific sector. Finally, based on the NP constraints, we derive upper bounds on the coefficients of the most general ΔF=2 effective Hamiltonian. These upper bounds can be translated into lower bounds on the scale of NP that contributes to these low-energy effective interactions.
DOI: 10.48550/arxiv.1901.05282
2019
Cited 13 times
LHC analysis-specific datasets with Generative Adversarial Networks
Using generative adversarial networks (GANs), we investigate the possibility of creating large amounts of analysis-specific simulated LHC events at limited computing cost. This kind of generative model is analysis specific in the sense that it directly generates the high-level features used in the last stage of a given physics analyses, learning the N-dimensional distribution of relevant features in the context of a specific analysis selection. We apply this idea to the generation of muon four-momenta in $Z \to \mu\mu$ events at the LHC. We highlight how use-case specific issues emerge when the distributions of the considered quantities exhibit particular features. We show how substantial performance improvements and convergence speed-up can be obtained by including regression terms in the loss function of the generator. We develop an objective criterion to assess the geenrator performance in a quantitative way. With further development, a generalization of this approach could substantially reduce the needed amount of centrally produced fully simulated events in large particle physics experiments.
DOI: 10.1016/j.nuclphysbps.2015.09.128
2016
Cited 12 times
Global Bayesian Analysis of the Higgs-boson Couplings
We present preliminary results of a bayesian fit to the Wilson coefficients of the Standard Model gauge invariant dimension-6 operators involving one or more Higgs fields, using data on electroweak precision observables and Higgs boson signal strengths.
DOI: 10.1088/2632-2153/ac5435
2022
Cited 5 times
Source-agnostic gravitational-wave detection with recurrent autoencoders
Abstract We present an application of anomaly detection techniques based on deep recurrent autoencoders (AEs) to the problem of detecting gravitational wave (GW) signals in laser interferometers. Trained on noise data, this class of algorithms could detect signals using an unsupervised strategy, i.e. without targeting a specific kind of source. We develop a custom architecture to analyze the data from two interferometers. We compare the obtained performance to that obtained with other AE architectures and with a convolutional classifier. The unsupervised nature of the proposed strategy comes with a cost in terms of accuracy, when compared to more traditional supervised techniques. On the other hand, there is a qualitative gain in generalizing the experimental sensitivity beyond the ensemble of pre-computed signal templates. The recurrent AE outperforms other AEs based on different architectures. The class of recurrent AEs presented in this paper could complement the search strategy employed for GW detection and extend the discovery reach of the ongoing detection campaigns.
DOI: 10.1103/physrevd.76.014015
2007
Cited 19 times
Improved determination of the CKM angle<mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" display="inline"><mml:mi>α</mml:mi></mml:math>from<mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" display="inline"><mml:mi>B</mml:mi><mml:mo>→</mml:mo><mml:mi>π</mml:mi><mml:mi>π</mml:mi></mml:math>decays
Motivated by a recent paper that compares the results of the analysis of the CKM angle $\ensuremath{\alpha}$ in the frequentist and in the Bayesian approaches, we have reconsidered the information on the hadronic amplitudes, which helps in constraining the value of $\ensuremath{\alpha}$ in the standard model. We find that the Bayesian method gives consistent results irrespective of the parametrization of the hadronic amplitudes and that the results of the frequentist and Bayesian approaches are equivalent when comparing meaningful probability ranges or confidence levels. We also find that from $B\ensuremath{\rightarrow}\ensuremath{\pi}\ensuremath{\pi}$ decays alone the 95% probability region for $\ensuremath{\alpha}$ is the interval [80\ifmmode^\circ\else\textdegree\fi{}, 170\ifmmode^\circ\else\textdegree\fi{}], well consistent with recent analyses of the unitarity triangle where, by using all the available experimental and theoretical information, one gets $\ensuremath{\alpha}=(93\ifmmode\pm\else\textpm\fi{}4)\ifmmode^\circ\else\textdegree\fi{}$. Last but not least, by using simple arguments on the hadronic matrix elements, we show that the unphysical region $\ensuremath{\alpha}\ensuremath{\sim}0$, present in several experimental analyses, can be eliminated.
DOI: 10.48550/arxiv.1102.0392
2011
Cited 14 times
Theoretical uncertainty in sin 2beta: An update
The source of theoretical uncertainty in the extraction of sin 2beta from the measurement of the golden channel Bd -&gt; J/psi K0 is briefly reviewed. An updated estimate of this uncertainty based on SU(3) flavour symmetry and the measurement of the decay Bd -&gt; J/psi pi0 is also presented.
DOI: 10.22323/1.282.0690
2017
Cited 12 times
Electroweak precision constraints at present and future colliders
We revisit the global fit to electroweak precision observables in the Standard Model and present model-independent bounds on several general new physics scenarios. We present a projection of the fit based on the expected experimental improvements at future $e^+ e^-$ colliders, and compare the constraining power of some of the different experiments that have been proposed. All results have been obtained with the HEPfit code.
DOI: 10.1007/s41781-018-0020-1
2019
Cited 11 times
Detector Monitoring with Artificial Neural Networks at the CMS Experiment at the CERN Large Hadron Collider
Reliable data quality monitoring is a key asset in delivering collision data suitable for physics analysis in any modern large-scale high energy physics experiment. This paper focuses on the use of artificial neural networks for supervised and semi-supervised problems related to the identification of anomalies in the data collected by the CMS muon detectors. We use deep neural networks to analyze LHC collision data, represented as images organized geographically. We train a classifier capable of detecting the known anomalous behaviors with unprecedented efficiency and explore the usage of convolutional autoencoders to extend anomaly detection capabilities to unforeseen failure modes. A generalization of this strategy could pave the way to the automation of the data quality assessment process for present and future high energy physics experiments.
2021
Cited 8 times
arXiv : The Dark Machines Anomaly Score Challenge: Benchmark Data and Model Independent Event Classification for the Large Hadron Collider
We describe the outcome of a data challenge conducted as part of the Dark Machines Initiative and the Les Houches 2019 workshop on Physics at TeV colliders. The challenged aims at detecting signals of new physics at the LHC using unsupervised machine learning algorithms. First, we propose how an anomaly score could be implemented to define model-independent signal regions in LHC searches. We define and describe a large benchmark dataset, consisting of >1 Billion simulated LHC events corresponding to $10~\rm{fb}^{-1}$ of proton-proton collisions at a center-of-mass energy of 13 TeV. We then review a wide range of anomaly detection and density estimation algorithms, developed in the context of the data challenge, and we measure their performance in a set of realistic analysis environments. We draw a number of useful conclusions that will aid the development of unsupervised new physics searches during the third run of the LHC, and provide our benchmark dataset for future studies at this https URL. Code to reproduce the analysis is provided at this https URL.