No Arabic abstract
The accumulation of noise in quantum computers is the dominant issue stymieing the push of quantum algorithms beyond their classical counterparts. We do not expect to be able to afford the overhead required for quantum error correction in the next decade, so in the meantime we must rely on low-cost, unscalable error mitigation techniques to bring quantum computing to its full potential. This paper presents a new error mitigation technique based on quantum phase estimation that can also reduce errors in expectation value estimation (e.g., for variational algorithms). The general idea is to apply phase estimation while effectively post-selecting for the system register to be in the starting state, which allows us to catch and discard errors which knock us away from there. We refer to this technique as verified phase estimation (VPE) and show that it can be adapted to function without the use of control qubits in order to simplify the control circuitry for near-term implementations. Using VPE, we demonstrate the estimation of expectation values on numerical simulations of intermediate scale quantum circuits with multiple orders of magnitude improvement over unmitigated estimation at near-term error rates (even after accounting for the additional complexity of phase estimation). Our numerical results suggest that VPE can mitigate against any single errors that might occur; i.e., the error in the estimated expectation values often scale as O(p^2), where p is the probability of an error occurring at any point in the circuit. This property, combined with robustness to sampling noise reveal VPE as a practical technique for mitigating errors in near-term quantum experiments.
Dynamical decoupling (DD) is a widely-used quantum control technique that takes advantage of temporal symmetries in order to partially suppress quantum errors without the need resource-intensive error detection and correction protocols. This and other open-loop error mitigation techniques are critical for quantum information processing in the era of Noisy Intermediate-Scale Quantum technology. However, despite its utility, dynamical decoupling does not address errors which occur at unstructured times during a circuit, including certain commonly-encountered noise mechanisms such as cross-talk and imperfectly calibrated control pulses. Here, we introduce and demonstrate an alternative technique - `quantum measurement emulation (QME) - that effectively emulates the measurement of stabilizer operators via stochastic gate application, leading to a first-order insensitivity to coherent errors. The QME protocol enables error suppression based on the stabilizer code formalism without the need for costly measurements and feedback, and it is particularly well-suited to discrete coherent errors that are challenging for DD to address.
Motivated by estimation of quantum noise models, we study the problem of learning a Pauli channel, or more generally the Pauli error rates of an arbitrary channel. By employing a novel reduction to the Population Recovery problem, we give an extremely simple algorithm that learns the Pauli error rates of an $n$-qubit channel to precision $epsilon$ in $ell_infty$ using just $O(1/epsilon^2) log(n/epsilon)$ applications of the channel. This is optimal up to the logarithmic factors. Our algorithm uses only unentangled state preparation and measurements, and the post-measurement classical runtime is just an $O(1/epsilon)$ factor larger than the measurement data size. It is also impervious to a limited model of measurement noise where heralded measurement failures occur independently with probability $le 1/4$. We then consider the case where the noise channel is close to the identity, meaning that the no-error outcome occurs with probability $1-eta$. In the regime of small $eta$ we extend our algorithm to achieve multiplicative precision $1 pm epsilon$ (i.e., additive precision $epsilon eta$) using just $Obigl(frac{1}{epsilon^2 eta}bigr) log(n/epsilon)$ applications of the channel.
If NISQ-era quantum computers are to perform useful tasks, they will need to employ powerful error mitigation techniques. Quasi-probability methods can permit perfect error compensation at the cost of additional circuit executions, provided that the nature of the error model is fully understood and sufficiently local both spatially and temporally. Unfortunately these conditions are challenging to satisfy. Here we present a method by which the proper compensation strategy can instead be learned ab initio. Our training process uses multiple variants of the primary circuit where all non-Clifford gates are substituted with gates that are efficient to simulate classically. The process yields a configuration that is near-optimal versus noise in the real system with its non-Clifford gate set. Having presented a range of learning strategies, we demonstrate the power of the technique both with real quantum hardware (IBM devices) and exactly-emulated imperfect quantum computers. The systems suffer a range of noise severities and types, including spatially and temporally correlated variants. In all cases the protocol successfully adapts to the noise and mitigates it to a high degree.
A general method to mitigate the effect of errors in quantum circuits is outlined. The method is developed in sight of characteristics that an ideal method should possess and to ameliorate an existing method which only mitigates state preparation and measurement errors. The method is tested on different IBM Q quantum devices, using randomly generated circuits with up to four qubits. A large majority of results show significant error mitigation.
Contemporary quantum computers have relatively high levels of noise, making it difficult to use them to perform useful calculations, even with a large number of qubits. Quantum error correction is expected to eventually enable fault-tolerant quantum computation at large scales, but until then it will be necessary to use alternative strategies to mitigate the impact of errors. We propose a near-term friendly strategy to mitigate errors by entangling and measuring $M$ copies of a noisy state $rho$. This enables us to estimate expectation values with respect to a state with dramatically reduced error, $rho^M/ mathrm{Tr}(rho^M)$, without explicitly preparing it, hence the name virtual distillation. As $M$ increases, this state approaches the closest pure state to $rho$, exponentially quickly. We analyze the effectiveness of virtual distillation and find that it is governed in many regimes by the behavior of this pure state (corresponding to the dominant eigenvector of $rho$). We numerically demonstrate that virtual distillation is capable of suppressing errors by multiple orders of magnitude and explain how this effect is enhanced as the system size grows. Finally, we show that this technique can improve the convergence of randomized quantum algorithms, even in the absence of device noise.