ﻻ يوجد ملخص باللغة العربية
Distilling precise estimates from noisy intermediate scale quantum (NISQ) data has recently attracted considerable attention. In order to augment digital qubit metrics, such as gate fidelity, we discuss analog error mitigability, i.e. the ability to accurately distill precise observable estimates, as a hybrid quantum-classical computing benchmarking task. Specifically, we characterize single qubit error rates on IBMs Poughkeepsie superconducting quantum hardware, incorporate control-mediated noise dependence into a generalized rescaling protocol, and analyze how noise characteristics influence Richardson extrapolation-based error mitigation. Our results identify regions in the space of Hamiltonian control fields and circuit-depth which are most amenable to reliable noise extrapolation, as well as shedding light on how low-level hardware characterization can be used as a predictive tool for uncertainty quantification in error mitigated NISQ computations.
Estimating the features of noise is the first step in a chain of protocols that will someday lead to fault tolerant quantum computers. The randomised benchmarking (RB) protocol is designed with this exact mindset, estimating the average strength of n
Benchmarking is how the performance of a computing system is determined. Surprisingly, even for classical computers this is not a straightforward process. One must choose the appropriate benchmark and metrics to extract meaningful results. Different
A precise measurement of dephasing over a range of timescales is critical for improving quantum gates beyond the error correction threshold. We present a metrological tool, based on randomized benchmarking, capable of greatly increasing the precision
As interest in quantum computing grows, there is a pressing need for standardized APIs so that algorithm designers, circuit designers, and physicists can be provided a common reference frame for designing, executing, and optimizing experiments. There
Quantum processors with sizes in the 10-100 qubit range are now increasingly common. However, with increased size comes increased complexity for benchmarking. The effectiveness of a given device may vary greatly between different tasks, and will not