Do you want to publish a course? Click here

Characterizing the Stability of NISQ Devices

153   0   0.0 ( 0 )
 Added by Samudra Dasgupta
 Publication date 2020
and research's language is English




Ask ChatGPT about the research

In this study, we focus on the question of stability of NISQ devices. The parameters that define the device stability profile are motivated by the work of DiVincenzo where the requirements for physical implementation of quantum computing are discussed. We develop the metrics and theoretical framework to quantify the DiVincenzo requirements and study the stability of those key metrics. The basis of our assessment is histogram similarity (in time and space). For identical experiments, devices which produce reproducible histograms in time, and similar histograms in space, are considered more reliable. To investigate such reliability concerns robustly, we propose a moment-based distance (MBD) metric. We illustrate our methodology using data collected from IBMs Yorktown device. Two types of assessments are discussed: spatial stability and temporal stability.



rate research

Read More

Noisy, intermediate-scale quantum (NISQ) computing devices offer opportunities to test the principles of quantum computing but are prone to errors arising from various sources of noise. Fluctuations in the noise itself lead to unstable devices that undermine the reproducibility of NISQ results. Here we characterize the reliability of NISQ devices by quantifying the stability of essential performance metrics. Using the Hellinger distance, we quantify the similarity between experimental characterizations of several NISQ devices by comparing gate fidelities, duty cycles, and register addressability across temporal and spatial scales. Our observations collected over 22 months reveal large fluctuations in each metric that underscore the limited scales on which current NISQ devices may be considered reliable.
109 - Yao Tang 2020
In the era of noisy intermediate-scale quantum (NISQ), executing quantum algorithms on actual quantum devices faces unique challenges. One such challenge is that quantum devices in this era have restricted connectivity: quantum gates are allowed to act only on specific pairs of physical qubits. For this reason, a quantum circuit needs to go through a compiling process called qubit routing before it can be executed on a quantum computer. In this study, we propose a CNOT synthesis method called the token reduction method to solve this problem. The token reduction method works for all quantum computers whose architecture is represented by connected graphs. A major difference between our method and the existing ones is that our method synthesizes a circuit to an output qubit mapping that might be different from the input qubit mapping. The final mapping for the synthesis is determined dynamically during the synthesis process. Results showed that our algorithm consistently outperforms the best publicly accessible algorithm for all of the tested quantum architectures.
Variational Quantum Eigensolvers (VQEs) have recently attracted considerable attention. Yet, in practice, they still suffer from the efforts for estimating cost function gradients for large parameter sets or resource-demanding reinforcement strategies. Here, we therefore consider recent advances in weight-agnostic learning and propose a strategy that addresses the trade-off between finding appropriate circuit architectures and parameter tuning. We investigate the use of NEAT-inspired algorithms which evaluate circuits via genetic competition and thus circumvent issues due to exceeding numbers of parameters. Our methods are tested both via simulation and on real quantum hardware and are used to solve the transverse Ising Hamiltonian and the Sherrington-Kirkpatrick spin model.
Quantum machine learning is one of the most promising applications of quantum computing in the Noisy Intermediate-Scale Quantum(NISQ) era. Here we propose a quantum convolutional neural network(QCNN) inspired by convolutional neural networks(CNN), which greatly reduces the computing complexity compared with its classical counterparts, with $O((log_{2}M)^6) $ basic gates and $O(m^2+e)$ variational parameters, where $M$ is the input data size, $m$ is the filter mask size and $e$ is the number of parameters in a Hamiltonian. Our model is robust to certain noise for image recognition tasks and the parameters are independent on the input sizes, making it friendly to near-term quantum devices. We demonstrate QCNN with two explicit examples. First, QCNN is applied to image processing and numerical simulation of three types of spatial filtering, image smoothing, sharpening, and edge detection are performed. Secondly, we demonstrate QCNN in recognizing image, namely, the recognition of handwritten numbers. Compared with previous work, this machine learning model can provide implementable quantum circuits that accurately corresponds to a specific classical convolutional kernel. It provides an efficient avenue to transform CNN to QCNN directly and opens up the prospect of exploiting quantum power to process information in the era of big data.
The analogy between quantum chemistry and light-front quantum field theory, first noted by Kenneth G. Wilson, serves as motivation to develop light-front quantum simulation of quantum field theory. We demonstrate how calculations of hadron structure can be performed on Noisy Intermediate-Scale Quantum devices within the Basis Light-Front Quantization framework. We calculate the light-front wave functions of pions using an effective light-front Hamiltonian in a basis representation on a current quantum processor. We use the Variational Quantum Eigensolver to find the ground state energy and wave function, which is subsequently used to calculate pion mass radius, decay constant, elastic form factor, and charge radius.
comments
Fetching comments Fetching comments
mircosoft-partner

هل ترغب بارسال اشعارات عن اخر التحديثات في شمرا-اكاديميا