Thank you for visiting nature.com. You are using a browser version with limited support for CSS. To obtain
the best experience, we recommend you use a more up to date browser (or turn off compatibility mode in
Internet Explorer). In the meantime, to ensure continued support, we are displaying the site without styles
and JavaScript.
Reconstructing unstable heavy particles is a crucial aspect of many analyses at the Large Hadron Collider (LHC). We introduce SPA-Net, a machine-learning approach to this problem which outperforms existing baseline methods, performs several auxiliary tasks, and leads to significant improvements in three example flagship LHC analyses
Energetic neutrino beams from symmetric muon and anti-muon decays are used to study long-baseline neutrino oscillation, and constrain the Charge-Parity (CP) violation phase in the three-flavour neutrino mixing. Here, the authors provide results based on neutrino oscillation simulations to show that more than five standard deviation sensitivity on CP violation can be obtained from 5-10 years of data taken with the help of DUNE-like detectors.
The standard model of particle physics describes the fundamental constituents of matter and their interactions. We review the status of experimental hints for new physics, which, if confirmed, would require the extension of the standard model with new particles and new interactions.
Axions are hypothetical particles that constitute leading candidates for the identity of dark matter. Here, the authors improve previous exclusion bounds on axion-like particles in the range of 1.4–200 peV, and report direct terrestrial limits on the coupling of protons and neutrons with axion-like dark matter.
An emerging set of proposals seeks to use arrays of optomechanical sensors to detect weak distributed forces, for applications ranging from gravity-based subterranean imaging to dark matter searches. We propose an array of entanglement-enhanced optomechanical sensors to improve the broadband sensitivity of distributed force sensing.
The renormalization group is a key ingredient in methods of improving perturbative computations in particle physics. Here I briefly discuss its role in perturbative quantum chromodynamics and particularly the running of its coupling constant.
Reducing resource usage will improve the environmental impact of high-performance computing — but doing so can clash with the science goals of funders. Computational physicist Peter Skands explains how he approached the conflict.
Ten years after the discovery of the Higgs boson, the ATLAS Collaboration probes its underlying mechanism, the electroweak symmetry breaking, by measuring the scattering of Z bosons, one of the mediators of the weak interactions.
In proton–proton collisions, the CMS Collaboration measures the simultaneous production of three particles, each consisting of a charm quark and a charm antiquark, which yields insights into how the proton’s constituents interact.
The CMS Collaboration finds evidence for the contribution from off-shell Higgs bosons to the production of events with two Z bosons. This provides a measurement of the Higgs boson’s width.
Sarah Malik explains how quantum random walks can be used to model the cascades of quarks and gluons resulting from the proton–proton collisions at the Large Hadron Collider.