Skip to main content
eScholarship
Open Access Publications from the University of California

Neural Posterior Unfolding

(2024)

Differential cross section measurements are the currency of scientific exchange in particle and nuclear physics. The key challenge for these analyses is the correction for detector distortions known as deconvolution or unfolding. In the case of binned cross section measurements, there are many tools for regularized matrix inversion where the matrix governs the detector response going from pre- to post-detector observables. In this paper, we show how normalizing flows and neural posterior estimation can be used for unfolding. This approach has many potential advantages, including implicit regularization from the neural networks and fast inference from amortized training. We demonstrate this approach using simple Gaussian examples as well as a simulated jet substructure measurement at the Large Hadron Collider.

Multidimensional Deconvolution with Profiling

(2024)

In many experimental contexts, it is necessary to statistically remove the impact of instrumental effects in order to physically interpret measurements. This task has been extensively studied in particle physics, where the deconvolution task is called unfolding. A number of recent methods have shown how to perform high-dimensional, unbinned unfolding using machine learning. However, one of the assumptions in all of these methods is that the detector response is correctly modeled in the Monte Carlo simulation. In practice, the detector response depends on a number of nuisance parameters that can be constrained with data. We propose a new algorithm called Profile OmniFold, which works in a similar iterative manner as the OmniFold algorithm while being able to simultaneously profile the nuisance parameters. We illustrate the method with a Gaussian example as a proof of concept highlighting its promising capabilities.

Cover page of Constraints on compact objects from the Dark Energy Survey 5-yr supernova sample

Constraints on compact objects from the Dark Energy Survey 5-yr supernova sample

(2024)

ABSTRACT: Gravitational lensing magnification of Type Ia supernovae (SNe Ia) allows information to be obtained about the distribution of matter on small scales. In this paper, we derive limits on the fraction $\alpha$ of the total matter density in compact objects (which comprise stars, stellar remnants, small stellar groupings, and primordial black holes) of mass M > 0.03 ${\rm M}_{\odot }$ over cosmological distances. Using 1532 SNe Ia from the Dark Energy Survey Year 5 sample (DES-SN5YR) combined with a Bayesian prior for the absolute magnitude M, we obtain α < 0.12 at the 95 per cent confidence level after marginalization over cosmological parameters, lensing due to large-scale structure, and intrinsic non-Gaussianity. Similar results are obtained using priors from the cosmic microwave background, baryon acoustic oscillations, and galaxy weak lensing, indicating our results do not depend on the background cosmology. We argue our constraints are likely to be conservative (in the sense of the values we quote being higher than the truth), but discuss scenarios in which they could be weakened by systematics of the order of $\Delta \alpha \sim 0.04$.

Search for heavy right-handed Majorana neutrinos in the decay of top quarks produced in proton-proton collisions at s=13  TeV with the ATLAS detector

(2024)

A search for heavy right-handed Majorana neutrinos is performed with the ATLAS detector at the CERN Large Hadron Collider, using the 140  fb−1 of proton–proton collision data at s=13  TeV collected during Run 2. This search targets tt¯ production, in which both top quarks decay into a bottom quark and a W boson, where one of the W bosons decays hadronically and the other decays into an electron or muon and a heavy neutral lepton. The heavy neutral lepton is identified through a decay into an electron or muon and another W boson, resulting in a pair of same-charge same-flavor leptons in the final state. This paper presents the first search for heavy neutral leptons in the mass range of 15–75 GeV using tt¯ events. No significant excess is observed over the background expectation, and upper limits are placed on the signal cross sections. Assuming a benchmark scenario of the phenomenological type-I seesaw model, these cross section limits are then translated into upper limits on the mixing parameters of the heavy Majorana neutrino with Standard Model neutrinos. © 2024 CERN, for the ATLAS Collaboration 2024 CERN

Cover page of Sequential Kalman tuning of the t-preconditioned Crank-Nicolson algorithm: efficient, adaptive and gradient-free inference for Bayesian inverse problems

Sequential Kalman tuning of the t-preconditioned Crank-Nicolson algorithm: efficient, adaptive and gradient-free inference for Bayesian inverse problems

(2024)

Abstract: Ensemble Kalman Inversion (EKI) has been proposed as an efficient method for the approximate solution of Bayesian inverse problems with expensive forward models. However, when applied to the Bayesian inverse problem EKI is only exact in the regime of Gaussian target measures and linear forward models. In this work we propose embedding EKI and Flow Annealed Kalman Inversion, its normalizing flow (NF) preconditioned variant, within a Bayesian annealing scheme as part of an adaptive implementation of the t-preconditioned Crank-Nicolson (tpCN) sampler. The tpCN sampler differs from standard pCN in that its proposal is reversible with respect to the multivariate t-distribution. The more flexible tail behaviour allows for better adaptation to sampling from non-Gaussian targets. Within our Sequential Kalman Tuning (SKT) adaptation scheme, EKI is used to initialize and precondition the tpCN sampler for each annealed target. The subsequent tpCN iterations ensure particles are correctly distributed according to each annealed target, avoiding the accumulation of errors that would otherwise impact EKI. We demonstrate the performance of SKT for tpCN on three challenging numerical benchmarks, showing significant improvements in the rate of convergence compared to adaptation within standard SMC with importance weighted resampling at each temperature level, and compared to similar adaptive implementations of standard pCN. The SKT scheme applied to tpCN offers an efficient, practical solution for solving the Bayesian inverse problem when gradients of the forward model are not available. Code implementing the SKT schemes for tpCN is available at https://github.com/RichardGrumitt/KalmanMC.

Cover page of Calibrating Bayesian generative machine learning for Bayesiamplification

Calibrating Bayesian generative machine learning for Bayesiamplification

(2024)

Abstract: Recently, combinations of generative and Bayesian deep learning have been introduced in particle physics for both fast detector simulation and inference tasks. These neural networks aim to quantify the uncertainty on the generated distribution originating from limited training statistics. The interpretation of a distribution-wide uncertainty however remains ill-defined. We show a clear scheme for quantifying the calibration of Bayesian generative machine learning models. For a Continuous Normalizing Flow applied to a low-dimensional toy example, we evaluate the calibration of Bayesian uncertainties from either a mean-field Gaussian weight posterior, or Monte Carlo sampling network weights, to gauge their behaviour on unsteady distribution edges. Well calibrated uncertainties can then be used to roughly estimate the number of uncorrelated truth samples that are equivalent to the generated sample and clearly indicate data amplification for smooth features of the distribution.

Software Performance of the ATLAS Track Reconstruction for LHC Run 3

(2024)

Charged particle reconstruction in the presence of many simultaneous proton–proton (pp) collisions in the LHC is a challenging task for the ATLAS experiment’s reconstruction software due to the combinatorial complexity. This paper describes the major changes made to adapt the software to reconstruct high-activity collisions with an average of 50 or more simultaneous pp interactions per bunch crossing (pile-up) promptly using the available computing resources. The performance of the key components of the track reconstruction chain and its dependence on pile-up are evaluated, and the improvement achieved compared to the previous software version is quantified. For events with an average of 60pp collisions per bunch crossing, the updated track reconstruction is twice as fast as the previous version, without significant reduction in reconstruction efficiency and while reducing the rate of combinatorial fake tracks by more than a factor two.

White paper on light sterile neutrino searches and related phenomenology

(2024)

Moment extraction using an unfolding protocol without binning

(2024)

Deconvolving ("unfolding") detector distortions is a critical step in the comparison of cross-section measurements with theoretical predictions in particle and nuclear physics. However, most existing approaches require histogram binning while many theoretical predictions are at the level of statistical moments. We develop a new approach to directly unfold distribution moments as a function of another observable without having to first discretize the data. Our moment unfolding technique uses machine learning and is inspired by Boltzmann weight factors and generative adversarial networks (GANs). We demonstrate the performance of this approach using jet substructure measurements in collider physics. With this illustrative example, we find that our moment unfolding protocol is more precise than bin-based approaches and is as or more precise than completely unbinned methods.

Cover page of Deep Generative Models for Fast Photon Shower Simulation in ATLAS

Deep Generative Models for Fast Photon Shower Simulation in ATLAS

(2024)

The need for large-scale production of highly accurate simulated event samples for the extensive physics programme of the ATLAS experiment at the Large Hadron Collider motivates the development of new simulation techniques. Building on the recent success of deep learning algorithms, variational autoencoders and generative adversarial networks are investigated for modelling the response of the central region of the ATLAS electromagnetic calorimeter to photons of various energies. The properties of synthesised showers are compared with showers from a full detector simulation using geant4. Both variational autoencoders and generative adversarial networks are capable of quickly simulating electromagnetic showers with correct total energies and stochasticity, though the modelling of some shower shape distributions requires more refinement. This feasibility study demonstrates the potential of using such algorithms for ATLAS fast calorimeter simulation in the future and shows a possible way to complement current simulation techniques.