Do you want to publish a course? Click here

Error mitigation via stabilizer measurement emulation

189   0   0.0 ( 0 )
 Added by Ami Greene
 Publication date 2021
  fields Physics
and research's language is English




Ask ChatGPT about the research

Dynamical decoupling (DD) is a widely-used quantum control technique that takes advantage of temporal symmetries in order to partially suppress quantum errors without the need resource-intensive error detection and correction protocols. This and other open-loop error mitigation techniques are critical for quantum information processing in the era of Noisy Intermediate-Scale Quantum technology. However, despite its utility, dynamical decoupling does not address errors which occur at unstructured times during a circuit, including certain commonly-encountered noise mechanisms such as cross-talk and imperfectly calibrated control pulses. Here, we introduce and demonstrate an alternative technique - `quantum measurement emulation (QME) - that effectively emulates the measurement of stabilizer operators via stochastic gate application, leading to a first-order insensitivity to coherent errors. The QME protocol enables error suppression based on the stabilizer code formalism without the need for costly measurements and feedback, and it is particularly well-suited to discrete coherent errors that are challenging for DD to address.



rate research

Read More

The accumulation of noise in quantum computers is the dominant issue stymieing the push of quantum algorithms beyond their classical counterparts. We do not expect to be able to afford the overhead required for quantum error correction in the next decade, so in the meantime we must rely on low-cost, unscalable error mitigation techniques to bring quantum computing to its full potential. This paper presents a new error mitigation technique based on quantum phase estimation that can also reduce errors in expectation value estimation (e.g., for variational algorithms). The general idea is to apply phase estimation while effectively post-selecting for the system register to be in the starting state, which allows us to catch and discard errors which knock us away from there. We refer to this technique as verified phase estimation (VPE) and show that it can be adapted to function without the use of control qubits in order to simplify the control circuitry for near-term implementations. Using VPE, we demonstrate the estimation of expectation values on numerical simulations of intermediate scale quantum circuits with multiple orders of magnitude improvement over unmitigated estimation at near-term error rates (even after accounting for the additional complexity of phase estimation). Our numerical results suggest that VPE can mitigate against any single errors that might occur; i.e., the error in the estimated expectation values often scale as O(p^2), where p is the probability of an error occurring at any point in the circuit. This property, combined with robustness to sampling noise reveal VPE as a practical technique for mitigating errors in near-term quantum experiments.
Variational Quantum Algorithms (VQAs) are a promising application for near-term quantum processors, however the quality of their results is greatly limited by noise. For this reason, various error mitigation techniques have emerged to deal with noise that can be applied to these algorithms. Recent work introduced a technique for mitigating expectation values against correlated measurement errors that can be applied to measurements of 10s of qubits. We apply these techniques to VQAs and demonstrate its effectiveness in improving estimates to the cost function. Moreover, we use the data resulting from this technique to experimentally characterize measurement errors in terms of the device connectivity on devices of up to 20 qubits. These results should be useful for better understanding the near-term potential of VQAs as well as understanding the correlations in measurement errors on large, near-term devices.
We develop a classical bit-flip correction method to mitigate measurement errors on quantum computers. This method can be applied to any operator, any number of qubits, and any realistic bit-flip probability. We first demonstrate the successful performance of this method by correcting the noisy measurements of the ground-state energy of the longitudinal Ising model. We then generalize our results to arbitrary operators and test our method both numerically and experimentally on IBM quantum hardware. As a result, our correction method reduces the measurement error on the quantum hardware by up to one order of magnitude. We finally discuss how to pre-process the method and extend it to other errors sources beyond measurement errors. For local Hamiltonians, the overhead costs are polynomial in the number of qubits, even if multi-qubit correlations are included.
If NISQ-era quantum computers are to perform useful tasks, they will need to employ powerful error mitigation techniques. Quasi-probability methods can permit perfect error compensation at the cost of additional circuit executions, provided that the nature of the error model is fully understood and sufficiently local both spatially and temporally. Unfortunately these conditions are challenging to satisfy. Here we present a method by which the proper compensation strategy can instead be learned ab initio. Our training process uses multiple variants of the primary circuit where all non-Clifford gates are substituted with gates that are efficient to simulate classically. The process yields a configuration that is near-optimal versus noise in the real system with its non-Clifford gate set. Having presented a range of learning strategies, we demonstrate the power of the technique both with real quantum hardware (IBM devices) and exactly-emulated imperfect quantum computers. The systems suffer a range of noise severities and types, including spatially and temporally correlated variants. In all cases the protocol successfully adapts to the noise and mitigates it to a high degree.
Coherent errors are a dominant noise process in many quantum computing architectures. Unlike stochastic errors, these errors can combine constructively and grow into highly detrimental overrotations. To combat this, we introduce a simple technique for suppressing systematic coherent errors in low-density parity-check (LDPC) stabilizer codes, which we call stabilizer slicing. The essential idea is to slice low-weight stabilizers into two equally-weighted Pauli operators and then apply them by rotating in opposite directions, causing their overrotations to interfere destructively on the logical subspace. With access to native gates generated by 3-body Hamiltonians, we can completely eliminate purely coherent overrotation errors, and for overrotation noise of 0.99 unitarity we achieve a 135-fold improvement in the logical error rate of Surface-17. For more conventional 2-body ion trap gates, we observe an 89-fold improvement for Bacon-Shor-13 with purely coherent errors which should be testable in near-term fault-tolerance experiments. This second scheme takes advantage of the prepared gauge degrees of freedom, and to our knowledge is the first example in which the state of the gauge directly affects the robustness of a codes memory. This work demonstrates that coherent noise is preferable to stochastic noise within certain code and gate implementations when the coherence is utilized effectively.
comments
Fetching comments Fetching comments
Sign in to be able to follow your search criteria
mircosoft-partner

هل ترغب بارسال اشعارات عن اخر التحديثات في شمرا-اكاديميا