No Arabic abstract
Whereas in standard quantum state tomography one estimates an unknown state by performing various measurements with known devices, and whereas in detector tomography one estimates the POVM elements of a measurement device by subjecting to it various known states, we consider here the case of SPAM (state preparation and measurement) tomography where neither the states nor the measurement device are assumed known. For $d$-dimensional systems measured by $d$-outcome detectors, we find there are at most $d^2(d^2-1)$ gauge parameters that can never be determined by any such experiment, irrespective of the number of unknown states and unknown devices. For the case $d=2$ we find new gauge-invariant quantities that can be accessed directly experimentally and that can be used to detect and describe SPAM errors. In particular, we identify conditions whose violations detect the presence of correlations between SPAM errors. From the perspective of SPAM tomography, standard quantum state tomography and detector tomography are protocols that fix the gauge parameters through the assumption that some set of fiducial measurements is known or that some set of fiducial states is known, respectively.
State preparation and measurement (SPAM) errors limit the performance of many gate-based quantum computing architecures, but are partly correctable after a calibration step that requires, for an exact implementation on a register of $n$ qubits, $2^n$ additional characterization experiments, as well as classical post-processing. Here we introduce an approximate but efficient method for SPAM error characterization requiring the {it classical} processing of $2^n ! times 2^n$ real matrices, but only $O(n^2)$ measurements. The technique assumes that multi-qubit measurement errors are dominated by pair correlations, which are estimated with $n(n-1)k/2$ two-qubit experiments, where $k$ is a parameter related to the accuracy. We demonstrate the technique on the IBM and Rigetti online superconducting quantum computers, allowing comparison of their SPAM errors in both magnitude and degree of correlation. We also study the correlations as a function of the registers geometric layout. We find that the pair-correlation model is fairly accurate on linear arrays of superconducting qubits. However qubits arranged in more closely spaced two-dimensional geometries exhibit significant higher-order (such as 3-qubit) SPAM error correlations.
Crosstalk occurs in most quantum computing systems with more than one qubit. It can cause a variety of correlated and nonlocal crosstalk errors that can be especially harmful to fault-tolerant quantum error correction, which generally relies on errors being local and relatively predictable. Mitigating crosstalk errors requires understanding, modeling, and detecting them. In this paper, we introduce a comprehensive framework for crosstalk errors and a protocol for detecting and localizing them. We give a rigorous definition of crosstalk errors that captures a wide range of disparate physical phenomena that have been called crosstalk, and a concrete model for crosstalk-free quantum processors. Errors that violate this model are crosstalk errors. Next, we give an equivalent but purely operational (model-independent) definition of crosstalk errors. Using this definition, we construct a protocol for detecting a large class of crosstalk errors in a multi-qubit processor by finding conditional dependencies between observed experimental probabilities. It is highly efficient, in the sense that the number of unique experiments required scales at most cubically, and very often quadratically, with the number of qubits. We demonstrate the protocol using simulations of 2-qubit and 6-qubit processors.
The central challenge in building a quantum computer is error correction. Unlike classical bits, which are susceptible to only one type of error, quantum bits (qubits) are susceptible to two types of error, corresponding to flips of the qubit state about the $X$- and $Z$-directions. While the Heisenberg Uncertainty Principle precludes simultaneous monitoring of $X$- and $Z$-flips on a single qubit, it is possible to encode quantum information in large arrays of entangled qubits that enable accurate monitoring of all errors in the system, provided the error rate is low. Another crucial requirement is that errors cannot be correlated. Here, we characterize a superconducting multiqubit circuit and find that charge fluctuations are highly correlated on a length scale over 600~$mu$m; moreover, discrete charge jumps are accompanied by a strong transient suppression of qubit energy relaxation time across the millimeter-scale chip. The resulting correlated errors are explained in terms of the charging event and phonon-mediated quasiparticle poisoning associated with absorption of gamma rays and cosmic-ray muons in the qubit substrate. Robust quantum error correction will require the development of mitigation strategies to protect multiqubit arrays from correlated errors due to particle impacts.
The surface code is designed to suppress errors in quantum computing hardware and currently offers the most believable pathway to large-scale quantum computation. The surface code requires a 2-D array of nearest-neighbor coupled qubits that are capable of implementing a universal set of gates with error rates below approximately 1%, requirements compatible with experimental reality. Consequently, a number of authors are attempting to squeeze additional performance out of the surface code. We describe an optimal complexity error suppression algorithm, parallelizable to O(1) given constant computing resources per unit area, and provide evidence that this algorithm exploits correlations in the error models of each gate in an asymptotically optimal manner.
Quantum computing can become scalable through error correction, but logical error rates only decrease with system size when physical errors are sufficiently uncorrelated. During computation, unused high energy levels of the qubits can become excited, creating leakage states that are long-lived and mobile. Particularly for superconducting transmon qubits, this leakage opens a path to errors that are correlated in space and time. Here, we report a reset protocol that returns a qubit to the ground state from all relevant higher level states. We test its performance with the bit-flip stabilizer code, a simplified version of the surface code for quantum error correction. We investigate the accumulation and dynamics of leakage during error correction. Using this protocol, we find lower rates of logical errors and an improved scaling and stability of error suppression with increasing qubit number. This demonstration provides a key step on the path towards scalable quantum computing.