Skip to main content
eScholarship
Open Access Publications from the University of California

Total Cost of Ownership and Evaluation of Google Cloud Resources for the ATLAS Experiment at the LHC

(2025)

Abstract: The ATLAS Google Project was established as part of an ongoing evaluation of the use of commercial clouds by the ATLAS Collaboration, in anticipation of the potential future adoption of such resources by WLCG grid sites to fulfil or complement their computing pledges. Seamless integration of Google cloud resources into the worldwide ATLAS distributed computing infrastructure was achieved at large scale and for an extended period of time, and hence cloud resources are shown to be an effective mechanism to provide additional, flexible computing capacity to ATLAS. For the first time a total cost of ownership analysis has been performed, to identify the dominant cost drivers and explore effective mechanisms for cost control. Network usage significantly impacts the costs of certain ATLAS workflows, underscoring the importance of implementing such mechanisms. Resource bursting has been successfully demonstrated, whilst exposing the true cost of this type of activity. A follow-up to the project is underway to investigate methods for improving the integration of cloud resources in data-intensive distributed computing environments and reducing costs related to network connectivity, which represents the primary expense when extensively utilising cloud resources.

Search for Dark Matter Produced in Association with a Dark Higgs Boson in the bb¯ Final State Using pp Collisions at s=13 TeV with the ATLAS Detector

(2025)

A search is performed for dark matter particles produced in association with a resonantly produced pair of (Formula presented)-quarks with (Formula presented) using (Formula presented) of proton-proton collisions at a center-of-mass energy of 13 TeV recorded by the ATLAS detector at the LHC. This signature is expected in extensions of the standard model predicting the production of dark matter particles, in particular those containing a dark Higgs boson (Formula presented) that decays into (Formula presented). The highly boosted (Formula presented) topology is reconstructed using jet reclustering and a new identification algorithm. This search places stringent constraints across regions of the dark Higgs model parameter space that satisfy the observed relic density, excluding dark Higgs bosons with masses between 30 and 150 GeV in benchmark scenarios with (Formula presented) mediator masses up to 4.8 TeV at 95% confidence level.

Cover page of I/O in Machine Learning Applications on HPC Systems: A 360-degree Survey

I/O in Machine Learning Applications on HPC Systems: A 360-degree Survey

(2025)

Growing interest in Artificial Intelligence (AI) has resulted in a surge in demand for faster methods of Machine Learning (ML) model training and inference. This demand for speed has prompted the use of high performance computing (HPC) systems that excel in managing distributed workloads. Because data is the main fuel for AI applications, the performance of the storage and I/O subsystem of HPC systems is critical. In the past, HPC applications accessed large portions of data written by simulations or experiments or ingested data for visualizations or analysis tasks. ML workloads perform small reads spread across a large number of random files. This shift of I/O access patterns poses several challenges to modern parallel storage systems. In this paper, we survey I/O in ML applications on HPC systems, and target literature within a 6-year time window from 2019 to 2024. We define the scope of the survey, provide an overview of the common phases of ML, review available profilers and benchmarks, examine the I/O patterns encountered during offline data preparation, training, and inference, and explore I/O optimizations utilized in modern ML frameworks and proposed in recent literature. Lastly, we seek to expose research gaps that could spawn further R&D.

Cover page of Data Readiness for AI: A 360-Degree Survey

Data Readiness for AI: A 360-Degree Survey

(2025)

Artificial Intelligence (AI) applications critically depend on data. Poor quality data produces inaccurate and ineffective AI models that may lead to incorrect or unsafe use. Evaluation of data readiness is a crucial step in improving the quality and appropriateness of data usage for AI. R&D efforts have been spent on improving data quality. However, standardized metrics for evaluating data readiness for use in AI training are still evolving. In this study, we perform a comprehensive survey of metrics used to verify data readiness for AI training. This survey examines more than 140 papers published by ACM Digital Library, IEEE Xplore, journals such as Nature, Springer, and Science Direct, and online articles published by prominent AI experts. This survey aims to propose a taxonomy of data readiness for AI (DRAI) metrics for structured and unstructured datasets. We anticipate that this taxonomy will lead to new standards for DRAI metrics that would be used for enhancing the quality, accuracy, and fairness of AI training and inference.

Software and computing for Run 3 of the ATLAS experiment at the LHC

(2025)

Abstract: The ATLAS experiment has developed extensive software and distributed computing systems for Run 3 of the LHC. These systems are described in detail, including software infrastructure and workflows, distributed data and workload management, database infrastructure, and validation. The use of these systems to prepare the data for physics analysis and assess its quality are described, along with the software tools used for data analysis itself. An outlook for the development of these projects towards Run 4 is also provided.

Configuration, Performance, and Commissioning of the ATLAS b-jet Triggers for the 2022 and 2023 LHC data-taking periods

(2025)

Abstract: In 2022 and 2023, the Large Hadron Collider produced approximately two billion hadronic interactions each second from bunches of protons that collide at a rate of 40 MHz. The ATLAS trigger system is used to reduce this rate to a few kHz for recording. Selections based on hadronic jets, their energy, and event topology reduce the rate to 𝒪(10) kHz while maintaining high efficiencies for important signatures resulting in b-quarks, but to reach the desired recording rate of hundreds of Hz, additional real-time selections based on the identification of jets containing b-hadrons (b-jets) are employed to achieve low thresholds on the jet transverse momentum at the High-Level Trigger. The configuration, commissioning, and performance of the real-time ATLAS b-jet identification algorithms for the early LHC Run 3 collision data are presented. These recent developments provide substantial gains in signal efficiency for critical signatures; for the Standard Model production of Higgs boson pairs, a 50% improvement in selection efficiency is observed in final states with four b-quarks or two b-quarks and two hadronically decaying τ-leptons.

Measurement of photonuclear jet production in ultraperipheral Pb+Pb collisions at sNN=5.02 TeV with the ATLAS detector

(2025)

In ultrarelativistic heavy ion collisions at the LHC, each nucleus acts a sources of high-energy real photons that can scatter off the opposing nucleus in ultraperipheral photonuclear ((Formula presented)) collisions. Hard scattering processes initiated by the photons in such collisions provide a novel method for probing nuclear parton distributions in a kinematic region not easily accessible to other measurements. ATLAS has measured production of dijet and multijet final states in ultraperipheral (Formula presented) collisions at (Formula presented) using a dataset recorded in 2018 with an integrated luminosity of (Formula presented). Photonuclear final states are selected by requiring a rapidity gap in the photon direction; this selects events where one of the outgoing nuclei remains intact. Jets are reconstructed using the anti-(Formula presented) algorithm with radius parameter, (Formula presented). Triple-differential cross sections, unfolded for detector response, are measured and presented using two sets of kinematic variables. The first set consists of the total transverse momentum ((Formula presented)), rapidity, and mass of the jet system. The second set uses (Formula presented) and particle-level nuclear and photon parton momentum fractions, (Formula presented) and (Formula presented), respectively. The results are compared with leading-order perturbative QCD calculations of photonuclear jet production cross sections, where all leading order predictions using existing fits fall below the data in the shadowing region. More detailed theoretical comparisons will allow these results to strongly constrain nuclear parton distributions, and these data provide results from the LHC directly comparable to early physics results at the planned Electron-Ion Collider.

Search for light neutral particles decaying promptly into collimated pairs of electrons or muons in pp collisions at s = 13 TeV with the ATLAS detector

(2025)

A search for a dark photon, a new light neutral particle, which decays promptly into collimated pairs of electrons or muons is presented. The search targets dark photons resulting from the exotic decay of the Standard Model Higgs boson, assuming its production via the dominant gluon-gluon fusion mode. The analysis is based on 140fb-1 of data collected with the ATLAS detector at the Large Hadron Collider from proton-proton collisions at a center-of-mass energy of 13 TeV. Events with collimated pairs of electrons or muons are analysed and background contributions are estimated using data-driven techniques. No significant excess in the data above the Standard Model background is observed. Upper limits are set at 95% confidence level on the branching ratio of the Higgs boson decay into dark photons between 0.001% and 5%, depending on the assumed dark photon mass and signal model.

Measurement of the associated production of a top-antitop-quark pair and a Higgs boson decaying into a bb¯ pair in pp collisions at s=13 TeV using the ATLAS detector at the LHC

(2025)

Abstract: This paper reports the measurement of Higgs boson production in association with a $$t\bar{t}$$ t t ¯ pair in the $$H\rightarrow b\bar{b}$$ H → b b ¯ decay channel. The analysis uses 140 fb $$^{-1}$$ - 1 of 13 $$\text {TeV}$$ TeV proton–proton collision data collected with the ATLAS detector at the Large Hadron Collider. The final states with one or two electrons or muons are employed. An excess of events over the expected background is found with an observed (expected) significance of 4.6 (5.4) standard deviations. The $$t\bar{t}H$$ t t ¯ H cross-section is $$\sigma _{t\bar{t}H} = 411~^{+101}_{-92}~\text {fb} = 411 \pm 54(\text {stat.})~^{+85}_{-75}(\text {syst.})~\text {fb}$$ σ t t ¯ H = 411 - 92 + 101 fb = 411 ± 54 ( stat. ) - 75 + 85 ( syst. ) fb for a Higgs boson mass of 125.09 $$\text {GeV}$$ GeV , consistent with the prediction of the Standard Model of $$507^{+35}_{-50}$$ 507 - 50 + 35 fb. The cross-section is also measured differentially in bins of the Higgs boson transverse momentum within the simplified template cross-section framework.

Search for Magnetic Monopole Pair Production in Ultraperipheral Pb+Pb Collisions at sNN=5.36 TeV with the ATLAS Detector at the LHC

(2025)

This Letter presents a search for highly ionizing magnetic monopoles in 262  μb−1 of ultraperipheral Pb+Pb collision data at sNN=5.36  TeV collected by the ATLAS detector at the LHC. A new methodology that exploits the properties of clusters of hits reconstructed in the innermost silicon detector layers is introduced to study highly ionizing particles in heavy-ion data. No significant excess above the background, which is estimated using a data-driven technique, is observed. Using a nonperturbative semiclassical model, upper limits at 95% confidence level are set on the cross section for pair production of monopoles with a single Dirac magnetic charge in the mass range of 20–150 GeV. Depending on the model, monopoles with a single Dirac magnetic charge and mass below 80–120 GeV are excluded. © 2025 CERN, for the ATLAS Collaboration 2025 CERN