skip to main content

Title: Bayesian quantum state reconstruction with a learning-based tuned prior

We demonstrate machine-learning-enhanced Bayesian quantum state tomography on near-term intermediate-scale quantum hardware. Our approach to selecting prior distributions leverages pre-trained neural networks incorporating measurement data and en-ables improved inference times over standard prior distributions.

more » « less
Award ID(s):
Author(s) / Creator(s):
; ; ; ; ; ; ;
Publisher / Repository:
Optica Publishing Group
Date Published:
Page Range / eLocation ID:
Medium: X
Denver, Colorado
Sponsoring Org:
National Science Foundation
More Like this
  1. Abstract

    In recent years, the choice of prior distributions for Bayesian logistic regression has received considerable interest. It is widely acknowledged that noninformative, improper priors have to be used with caution because posterior propriety may not always hold. As an alternative, heavy‐tailed priors such as Cauchy prior distributions have been proposed by Gelman et al. (2008). The motivation for using Cauchy prior distributions is that they are proper, and thus, unlike noninformative priors, they are guaranteed to yield proper posterior distributions. The heavy tails of the Cauchy distribution allow the posterior distribution to adapt to the data. Thus Gelman et al. (2008) suggested the use of these prior distributions as a default weakly informative choice, in the absence of prior knowledge about the regression coefficients. While these prior distributions are guaranteed to have proper posterior distributions, Ghosh, Li, and Mitra (2018), showed that the posterior means may not exist, or can be unreasonably large, when they exist, for datasets with separation. In this paper, we provide a short review of the concept of separation, and we discuss how common prior distributions like the Cauchy and normal prior perform in data with separation. Theoretical and empirical results suggest that lighter tailed prior distributions such as normal prior distributions can be a good default choice when there is separation.

    This article is categorized under:

    Statistical Learning and Exploratory Methods of the Data Sciences > Modeling Methods Statistical Models

    more » « less
  2. null (Ed.)
    We study the problems of identity and closeness testing of n-dimensional product distributions. Prior works of Canonne et al. (2017) and Daskalakis and Pan (2017) have established tight sample complexity bounds for non-tolerant testing over a binary alphabet: given two product distributions P and Q over a binary alphabet, distinguish between the cases P = Q and dTV(P;Q) > epsilon . We build on this prior work to give a more comprehensive map of the complexity of testing of product distributions by investigating tolerant testing with respect to several natural distance measures and over an arbitrary alphabet. Our study gives a fine-grained understanding of how the sample complexity of tolerant testing varies with the distance measures for product distributions. In addition, we also extend one of our upper bounds on product distributions to bounded-degree Bayes nets. 
    more » « less
  3. Summary

    Existing Bayesian model selection procedures require the specification of prior distributions on the parameters appearing in every model in the selection set. In practice, this requirement limits the application of Bayesian model selection methodology. To overcome this limitation, we propose a new approach towards Bayesian model selection that uses classical test statistics to compute Bayes factors between possible models. In several test cases, our approach produces results that are similar to previously proposed Bayesian model selection and model averaging techniques in which prior distributions were carefully chosen. In addition to eliminating the requirement to specify complicated prior distributions, this method offers important computational and algorithmic advantages over existing simulation-based methods. Because it is easy to evaluate the operating characteristics of this procedure for a given sample size and specified number of covariates, our method facilitates the selection of hyperparameter values through prior-predictive simulation.

    more » « less
  4. We employ deep kernel learning electronic coarse-graining (DKL-ECG) with approximate Gaussian processes as a flexible and scalable framework for learning heteroscedastic electronic property distributions as a smooth function of coarse-grained (CG) configuration. The appropriateness of the Gaussian prior on predictive CG property distributions is justified as a function of CG model resolution by examining the statistics of target distributions. The certainties of predictive CG distributions are shown to be limited by CG model resolution with DKL-ECG predictive noise converging to the intrinsic physical noise induced by the CG mapping operator for multiple chemistries. Further analysis of the resolution dependence of learned CG property distributions allows for the identification of CG mapping operators that capture CG degrees of freedom with strong electron–phonon coupling. We further demonstrate the ability to construct the exact quantum chemical valence electronic density of states (EDOS), including behavior in the tails of the EDOS, from an entirely CG model by combining iterative Boltzmann inversion and DKL-ECG. DKL-ECG provides a means of learning CG distributions of all-atom properties that are traditionally “lost” in CG model development, introducing a promising methodological alternative to backmapping algorithms commonly employed to recover all-atom property distributions from CG simulations. 
    more » « less
  5. Kohayakawa, Y. ; Miyazawa, F.K. (Ed.)
    In this work we are interested in the problem of testing quantum entanglement. More specifically, we study the separability problem in quantum property testing, where one is given n copies of an unknown mixed quantum state ϱ on Cd⊗Cd , and one wants to test whether ϱ is separable or ϵ -far from all separable states in trace distance. We prove that n=Ω(d2/ϵ2) copies are necessary to test separability, assuming ϵ is not too small, viz. ϵ=Ω(1/d−−√) . We also study completely positive distributions on the grid [d]×[d] , as a classical analogue of separable states. We analogously prove that Ω(d/ϵ2) samples from an unknown distribution p are necessary to decide whether p is completely positive or ϵ -far from all completely positive distributions in total variation distance. 
    more » « less