Quantum annealing is a powerful alternative model of quantum computing, which can succeed in the presence of environmental noise even without error correction. However, despite great effort, no conclusive demonstration of a quantum speedup (relative to state of the art classical algorithms) has been shown for these systems, and rigorous theoretical proofs of a quantum advantage (such as the adiabatic formulation of Grover’s search problem) generally rely on exponential precision in at least some aspects of the system, an unphysical resource guaranteed to be scrambled by experimental uncertainties and random noise. In this work, we propose a new variant of quantum annealing, called RFQA, which can maintain a scalable quantum speedup in the face of noise and modest control precision. Specifically, we consider a modification of flux qubit-based quantum annealing which includes low-frequency oscillations in the directions of the transverse field terms as the system evolves. We show that this method produces a quantum speedup for finding ground states in the Grover problem and quantum random energy model, and thus should be widely applicable to other hard optimization problems which can be formulated as quantum spin glasses. Further, we explore three realistic noise channels and show that the speedup more »
- Award ID(s):
- Publication Date:
- NSF-PAR ID:
- Journal Name:
- Quantum Science and Technology
- Page Range or eLocation-ID:
- Article No. 025013
- IOP Publishing
- Sponsoring Org:
- National Science Foundation
More Like this
Quantum Computing has attracted much research attention because of its potential to achieve fundamental speed and efficiency improvements in various domains. Among different quantum algorithms, Parameterized Quantum Circuits (PQC) for Quantum Machine Learning (QML) show promises to realize quantum advantages on the current Noisy Intermediate-Scale Quantum (NISQ) Machines. Therefore, to facilitate the QML and PQC research, a recent python library called TorchQuantum has been released. It can construct, simulate, and train PQC for machine learning tasks with high speed and convenient debugging supports. Besides quantum for ML, we want to raise the community's attention on the reversed direction: ML for quantum. Specifically, the TorchQuantum library also supports using data-driven ML models to solve problems in quantum system research, such as predicting the impact of quantum noise on circuit fidelity and improving the quantum circuit compilation efficiency. This paper presents a case study of the ML for quantum part in TorchQuantum. Since estimating the noise impact on circuit reliability is an essential step toward understanding and mitigating noise, we propose to leverage classical ML to predict noise impact on circuit fidelity. Inspired by the natural graph representation of quantum circuits, we propose to leverage a graph transformer model to predict themore »
Dense ensembles of spin qubits are valuable for quantum applications, even though their coherence protection remains challenging. Continuous dynamical decoupling can protect ensemble qubits from noise while allowing gate operations, but it is hindered by the additional noise introduced by the driving. Concatenated continuous driving (CCD) techniques can, in principle, mitigate this problem. Here we provide deeper insights into the dynamics under CCD, based on Floquet theory, that lead to optimized state protection by adjusting driving parameters in the CCD scheme to induce mode evolution control. We experimentally demonstrate the improved control by simultaneously addressing a dense nitrogen-vacancy (NV) ensemble with 1010spins. We achieve an experimental 15-fold improvement in coherence time for an arbitrary, unknown state, and a 500-fold improvement for an arbitrary, known state, corresponding to driving the sidebands and the center band of the resulting Mollow triplet, respectively. We can achieve such coherence time gains by optimizing the driving parameters to take into account the noise affecting our system. By extending the generalized Bloch equation approach to the CCD scenario, we identify the noise sources that dominate the decay mechanisms in NV ensembles, confirm our model by experimental results, and identify the driving strengths yielding optimal coherence.more »
The decoherence effects experienced by the qubits of a quantum processor are generally characterized using the amplitude damping time (
T1) and the dephasing time ( T2). Quantum channel models that exist at the time of writing assume that these parameters are fixed and invariant. However, recent experimental studies have shown that they exhibit a time-varying (TV) behaviour. These time-dependant fluctuations of T1and T2, which become even more pronounced in the case of superconducting qubits, imply that conventional static quantum channel models do not capture the noise dynamics experienced by realistic qubits with sufficient precision. In this article, we study how the fluctuations of T1and T2can be included in quantum channel models. We propose the idea of time-varying quantum channel (TVQC) models, and we show how they provide a more realistic portrayal of decoherence effects than static models in some instances. We also discuss the divergence that exists between TVQCs and their static counterparts by means of a metric known as the diamond norm. In many circumstances this divergence can be significant, which indicates that the time-dependent nature of decoherence must be considered, in order to construct models that capture the real nature of quantum devices.
A distributed sensing protocol uses a network of local sensing nodes to estimate a global feature of the network, such as a weighted average of locally detectable parameters. In the noiseless case, continuous-variable (CV) multipartite entanglement shared by the nodes can improve the precision of parameter estimation relative to the precision attainable by a network without shared entanglement; for an entangled protocol, the root mean square estimation error scales like 1/
Mwith the number Mof sensing nodes, the so-called Heisenberg scaling, while for protocols without entanglement, the error scales like . However, in the presence of loss and other noise sources, although multipartite entanglement still has some advantages for sensing displacements and phases, the scaling of the precision with Mis less favorable. In this paper, we show that using CV error correction codes can enhance the robustness of sensing protocols against imperfections and reinstate Heisenberg scaling up to moderate values of M. Furthermore, while previous distributed sensing protocols could measure only a single quadrature, we construct a protocol in which both quadratures can be sensed simultaneously. Our work demonstrates the value of CV error correction codes in realistic sensing scenarios.
We present a new lossy compression algorithm for statistical floating-point data through a representation learning with binary variables. The algorithm finds a set of basis vectors and their binary coefficients that precisely reconstruct the original data. The optimization for the basis vectors is performed classically, while binary coefficients are retrieved through both simulated and quantum annealing for comparison. A bias correction procedure is also presented to estimate and eliminate the error and bias introduced from the inexact reconstruction of the lossy compression for statistical data analyses. The compression algorithm is demonstrated on two different datasets of lattice quantum chromodynamics simulations. The results obtained using simulated annealing show 3–3.5 times better compression performance than the algorithm based on neural-network autoencoder. Calculations using quantum annealing also show promising results, but performance is limited by the integrated control error of the quantum processing unit, which yields large uncertainties in the biases and coupling parameters. Hardware comparison is further studied between the previous generation D-Wave 2000Q and the current D-Wave Advantage system. Our study shows that the Advantage system is more likely to obtain low-energy solutions for the problems than the 2000Q.