Leptoquarks (
Constraining effective field theories with machine learning
An important part of the Large Hadron Collider (LHC) legacy will be precise limits on indirect effects of new physics, framed for instance in terms of an effective field theory. These measurements often involve many theory parameters and observables, which makes them challenging for traditional analysis methods. We discuss the underlying problem of “likelihoodfree” inference and present powerful new analysis techniques that combine physics insights, statistical methods, and the power of machine learning. We have developed MadMiner, a new Python package that makes it straightforward to apply these techniques. In example LHC problems we show that the new approach lets us put stronger constraints on theory parameters than established methods, demonstrating its potential to improve the new physics reach of the LHC legacy measurements. While we present techniques optimized for particle physics, the likelihoodfree inference formulation is much more general, and these ideas are part of a broader movement that is changing scientific inference in fields as diverse as cosmology, genetics, and epidemiology.
more »
« less
 Award ID(s):
 1836650
 NSFPAR ID:
 10256982
 Editor(s):
 Doglioni, C.; Kim, D.; Stewart, G.A.; Silvestris, L.; Jackson, P.; Kamleh, W.
 Date Published:
 Journal Name:
 EPJ Web of Conferences
 Volume:
 245
 ISSN:
 2100014X
 Page Range / eLocation ID:
 06026
 Format(s):
 Medium: X
 Sponsoring Org:
 National Science Foundation
More Like this

Abstract s) are hypothetical particles that appear in various extensions of the Standard Model (SM), that can explain observed differences between SM theory predictions and experimental results. The production of these particles has been widely studied at various experiments, most recently at the Large Hadron Collider (LHC), and stringent bounds have been placed on their masses and couplings, assuming the simplest beyondSM (BSM) hypotheses. However, the limits are significantly weaker for$$\textrm{LQ}$$ $\text{LQ}$ models with family nonuniversal couplings containing enhanced couplings to thirdgeneration fermions. We present a new study on the production of a$$\textrm{LQ}$$ $\text{LQ}$ at the LHC, with preferential couplings to thirdgeneration fermions, considering protonproton collisions at$$\textrm{LQ}$$ $\text{LQ}$ and$$\sqrt{s} = 13 \, \textrm{TeV}$$ $\sqrt{s}=13\phantom{\rule{0ex}{0ex}}\text{TeV}$ . Such a hypothesis is well motivated theoretically and it can explain the recent anomalies in the precision measurements of$$\sqrt{s} = 13.6 \, \textrm{TeV}$$ $\sqrt{s}=13.6\phantom{\rule{0ex}{0ex}}\text{TeV}$ meson decay rates, specifically the$$\textrm{B}$$ $\text{B}$ ratios. Under a simplified model where the$$R_{D^{(*)}}$$ ${R}_{{D}^{(\ast )}}$ masses and couplings are free parameters, we focus on cases where the$$\textrm{LQ}$$ $\text{LQ}$ decays to a$$\textrm{LQ}$$ $\text{LQ}$ lepton and a$$\tau $$ $\tau $ quark, and study how the results are affected by different assumptions about chiral currents and interference effects with other BSM processes with the same final states, such as diagrams with a heavy vector boson,$$\textrm{b}$$ $\text{b}$ . The analysis is performed using machine learning techniques, resulting in an increased discovery reach at the LHC, allowing us to probe new physics phase space which addresses the$$\textrm{Z}^{\prime }$$ ${\text{Z}}^{\prime}$ meson anomalies, for$$\textrm{B}$$ $\text{B}$ masses up to$$\textrm{LQ}$$ $\text{LQ}$ , for the high luminosity LHC scenario.$$5.00\, \textrm{TeV}$$ $5.00\phantom{\rule{0ex}{0ex}}\text{TeV}$ 
Abstract The production of heavy neutral mass resonances, $$\text {Z}^{\prime }$$ Z ′ , has been widely studied theoretically and experimentally. Although the nature, mass, couplings, and associated quantum numbers of this hypothetical particle are yet to be determined, current LHC experimental results have set strong constraints assuming the simplest beyond Standard Model (SM) hypotheses. We present a new feasibility study on the production of a $$\text {Z}^{\prime }$$ Z ′ boson at the LHC, with family nonuniversal couplings, considering proton–proton collisions at $$\sqrt{s} = 13$$ s = 13 and 14 TeV. Such a hypothesis is well motivated theoretically and it can explain observed differences between SM predictions and experimental results, as well as being a useful tool to further probe recent results in searches for new physics considering nonuniversal fermion couplings. We work under two simplified phenomenological frameworks where the $$\textrm{Z}^{\prime }$$ Z ′ masses and couplings to the SM particles are free parameters, and consider final states of the $$\text {Z}^{\prime }$$ Z ′ decaying to a pair of $$\textrm{b}$$ b quarks. The analysis is performed using machine learning techniques to maximize the sensitivity. Despite being a well motivated physics case in its own merit, such scenarios have not been fully considered in ongoing searches at the LHC. We note the proposed search methodology can be a key mode for discovery over a large mass range, including low masses, traditionally considered difficult due to experimental constrains. In addition, the proposed search is complementary to existing strategies.more » « less

Abernethy, Jacob ; Agarwal, Shivani (Ed.)We study a variant of the sparse PCA (principal component analysis) problem in the “hard” regime, where the inference task is possible yet no polynomialtime algorithm is known to exist. Prior work, based on the lowdegree likelihood ratio, has conjectured a precise expression for the best possible (subexponential) runtime throughout the hard regime. Following instead a statistical physics inspired point of view, we show bounds on the depth of free energy wells for various Gibbs measures naturally associated to the problem. These free energy wells imply hitting time lower bounds that corroborate the lowdegree conjecture: we show that a class of natural MCMC (Markov chain Monte Carlo) methods (with worstcase initialization) cannot solve sparse PCA with less than the conjectured runtime. These lower bounds apply to a wide range of values for two tuning parameters: temperature and sparsity misparametrization. Finally, we prove that the Overlap Gap Property (OGP), a structural property that implies failure of certain local search algorithms, holds in a significant part of the hard regime.more » « less

Koyejo, Sanmi ; Mohamed, Shakir (Ed.)Differentially private mechanisms protect privacy by introducing additional randomness into the data. Restricting access to only the privatized data makes it challenging to perform valid statistical inference on parameters underlying the confidential data. Specifically, the likelihood function of the privatized data requires integrating over the large space of confidential databases and is typically intractable. For Bayesian analysis, this results in a posterior distribution that is doubly intractable, rendering traditional MCMC techniques inapplicable. We propose an MCMC framework to perform Bayesian inference from the privatized data, which is applicable to a wide range of statistical models and privacy mechanisms. Our MCMC algorithm augments the model parameters with the unobserved confidential data, and alternately updates each one conditional on the other. For the potentially challenging step of updating the confidential data, we propose a generic approach that exploits the privacy guarantee of the mechanism to ensure efficiency. In particular, we give results on the computational complexity, acceptance rate, and mixing properties of our MCMC. We illustrate the efficacy and applicability of our methods on a na\"iveBayes loglinear model as well as on a linear regression model.more » « less

Koyejo, S. ; Mohamed, S. ; Agarwal, A. ; Belgrave, D. ; Cho, K. ; Oh, A. (Ed.)Differentially private mechanisms protect privacy by introducing additional randomness into the data. Restricting access to only the privatized data makes it challenging to perform valid statistical inference on parameters underlying the confidential data. Specifically, the likelihood function of the privatized data requires integrating over the large space of confidential databases and is typically intractable. For Bayesian analysis, this results in a posterior distribution that is doubly intractable, rendering traditional MCMC techniques inapplicable. We propose an MCMC framework to perform Bayesian inference from the privatized data, which is applicable to a wide range of statistical models and privacy mechanisms. Our MCMC algorithm augments the model parameters with the unobserved confidential data, and alternately updates each one conditional on the other. For the potentially challenging step of updating the confidential data, we propose a generic approach that exploits the privacy guarantee of the mechanism to ensure efficiency. We give results on the computational complexity, acceptance rate, and mixing properties of our MCMC. We illustrate the efficacy and applicability of our methods on a naiveBayes loglinear model and on a linear regression model.more » « less