A
bstract
A number of recent applications of jet substructure, in particular searches for light new particles, require substructure observables that are decorrelated with the jet mass. In this paper ...we introduce the Convolved SubStructure (CSS) approach, which uses a theoretical understanding of the observable to decorrelate the complete shape of its distribution. This decorrelation is performed by convolution with a shape function whose parameters and mass dependence are derived analytically. We consider in detail the case of the
D
2
observable and perform an illustrative case study using a search for a light hadronically decaying
Z
′. We find that the CSS approach completely decorrelates the
D
2
observable over a wide range of masses. Our approach highlights the importance of improving the theoretical understanding of jet substructure observables to exploit increasingly subtle features for performance.
A
bstract
As machine learning algorithms become increasingly sophisticated to exploit subtle features of the data, they often become more dependent on simulations. This paper presents a new approach ...called
weakly supervised classification
in which class proportions are the only input into the machine learning algorithm. Using one of the most challenging binary classification tasks in high energy physics — quark versus gluon tagging — we show that weakly supervised classification can match the performance of fully supervised algorithms. Furthermore, by design, the new algorithm is insensitive to any mis-modeling of discriminating features in the data by the simulation. Weakly supervised classification is a general procedure that can be applied to a wide variety of learning problems to boost performance and robustness when detailed simulations are not reliable or not available.
Collider data must be corrected for detector effects ("unfolded") to be compared with many theoretical calculations and measurements from other experiments. Unfolding is traditionally done for ...individual, binned observables without including all information relevant for characterizing the detector response. We introduce OmniFold, an unfolding method that iteratively reweights a simulated dataset, using machine learning to capitalize on all available information. Our approach is unbinned, works for arbitrarily high-dimensional data, and naturally incorporates information from the full phase space. We illustrate this technique on a realistic jet substructure example from the Large Hadron Collider and compare it to standard binned unfolding methods. This new paradigm enables the simultaneous measurement of all observables, including those not yet invented at the time of the analysis.
Simulating quantum field theories is a flagship application of quantum computing. However, calculating experimentally relevant high energy scattering amplitudes entirely on a quantum computer is ...prohibitively difficult. It is well known that such high energy scattering processes can be factored into pieces that can be computed using well established perturbative techniques, and pieces which currently have to be simulated using classical Markov chain algorithms. These classical Markov chain simulation approaches work well to capture many of the salient features, but cannot capture all quantum effects. To exploit quantum resources in the most efficient way, we introduce a new paradigm for quantum algorithms in field theories. This approach uses quantum computers only for those parts of the problem which are not computable using existing techniques. In particular, we develop a polynomial time quantum final state shower that accurately models the effects of intermediate spin states similar to those present in high energy electroweak showers with a global evolution variable. The algorithm is explicitly demonstrated for a simplified quantum field theory on a quantum computer.
Quantum anomaly detection for collider physics Alvi, Sulaiman; Bauer, Christian W.; Nachman, Benjamin
The journal of high energy physics,
02/2023, Volume:
2023, Issue:
2
Journal Article
Peer reviewed
Open access
A
bstract
We explore the use of Quantum Machine Learning (QML) for anomaly detection at the Large Hadron Collider (LHC). In particular, we explore a semi-supervised approach in the four-lepton final ...state where simulations are reliable enough for a direct background prediction. This is a representative task where classification needs to be performed using small training datasets — a regime that has been suggested for a quantum advantage. We find that Classical Machine Learning (CML) benchmarks outperform standard QML algorithms and are able to automatically identify the presence of anomalous events injected into otherwise background-only datasets.
A
bstract
The likelihood ratio is a crucial quantity for statistical inference in science that enables hypothesis testing, construction of confidence intervals, reweighting of distributions, and ...more. Many modern scientific applications, however, make use of data- or simulation-driven models for which computing the likelihood ratio can be very difficult or even impossible. By applying the so-called “likelihood ratio trick,” approximations of the likelihood ratio may be computed using clever parametrizations of neural network-based classifiers. A number of different neural network setups can be defined to satisfy this procedure, each with varying performance in approximating the likelihood ratio when using finite training data. We present a series of empirical studies detailing the performance of several common loss functionals and parametrizations of the classifier output in approximating the likelihood ratio of two univariate and multivariate Gaussian distributions as well as simulated high-energy particle physics datasets.
A
bstract
Complete anomaly detection strategies that are both signal sensitive and compatible with background estimation have largely focused on resonant signals. Non-resonant new physics scenarios ...are relatively under-explored and may arise from off-shell effects or final states with significant missing energy. In this paper, we extend a class of weakly supervised anomaly detection strategies developed for resonant physics to the non-resonant case. Machine learning models are trained to reweight, generate, or morph the background, extrapolated from a control region. A classifier is then trained in a signal region to distinguish the estimated background from the data. The new methods are demonstrated using a semi-visible jet signature as a benchmark signal model, and are shown to automatically identify the anomalous events without specifying the signal ahead of time.
A
bstract
High-multiplicity all-hadronic final states are an important, but difficult final state for searching for physics beyond the Standard Model. A powerful search method is to look for large ...jets with accidental substructure due to multiple hard partons falling within a single jet. One way for estimating the background in this search is to exploit an approximate factorization in quantum chromodynamics whereby the jet mass distribution is determined only by its kinematic properties. Traditionally, this approach has been executed using histograms constructed in a background-rich region. We propose a new approach based on Generative Adversarial Networks (GANs). These neural network approaches are naturally unbinned and can be readily conditioned on multiple jet properties. In addition to using vanilla GANs for this purpose, a modification to the traditional WGAN approach has been investigated where weight clipping is replaced by drawing weights from a naturally compact set (in this case, the circle). Both the vanilla and modified WGAN approaches significantly outperform the histogram method, especially when modeling the dependence on features not used in the histogram construction. These results can be useful for enhancing the sensitivity of LHC searches to high-multiplicity final states involving many quarks and gluons and serve as a useful benchmark where GANs may have immediate benefit to the HEP community.
Pileup Mitigation with Machine Learning (PUMML) Komiske, Patrick T.; Metodiev, Eric M.; Nachman, Benjamin ...
The journal of high energy physics,
12/2017, Volume:
2017, Issue:
12
Journal Article
Peer reviewed
Open access
A
bstract
Pileup involves the contamination of the energy distribution arising from the primary collision of interest (leading vertex) by radiation from soft collisions (pileup). We develop a new ...technique for removing this contamination using machine learning and convolutional neural networks. The network takes as input the energy distribution of charged leading vertex particles, charged pileup particles, and all neutral particles and outputs the energy distribution of particles coming from leading vertex alone. The PUMML algorithm performs remarkably well at eliminating pileup distortion on a wide range of simple and complex jet observables. We test the robustness of the algorithm in a number of ways and discuss how the network can be trained directly on data.
Run 5 of the high-luminosity LHC era (and beyond) may provide new opportunities to search for physics beyond the standard model at interaction point 2. In particular, taking advantage of the existing ...ALICE detector and infrastructure provides an opportunity to search for displaced decays of beyond standard model long-lived particles. While this proposal may well be preempted by ongoing ALICE physics goals, examination of its potential new physics reach provides a compelling comparison with respect to other long-lived particle proposals. In particular, full event reconstruction and particle identification could be possible by making use of the existing L3 magnet and ALICE time projection chamber. For several well-motivated portals, the reach competes with or exceeds the sensitivity of MATHUSLA and SHiP, provided that a total integrated luminosity of approximately 100 fb−1 could be delivered to interaction point 2.