Do you want to publish a course? Click here

Optimizing Quantum Annealing Schedules: From Monte Carlo Tree Search to QuantumZero

124   0   0.0 ( 0 )
 Added by Yuqin Chen
 Publication date 2020
  fields Physics
and research's language is English




Ask ChatGPT about the research

Quantum annealing is a practical approach to execute the native instruction set of the adiabatic quantum computation model. The key of running adiabatic algorithms is to maintain a high success probability of evolving the system into the ground state of a problem-encoded Hamiltonian at the end of an annealing schedule. This is typically done by executing the adiabatic algorithm slowly to enforce adiabacity. However, properly optimized annealing schedule can accelerate the computational process. Inspired by the recent success of DeepMinds AlphaZero algorithm that can efficiently explore and find a good winning strategy from a large combinatorial search with a neural-network-assisted Monte Carlo Tree Search (MCTS), we adopt MCTS and propose a neural-network-enabled version, termed QuantumZero (QZero), to automate the design of an optimal annealing schedule in a hybrid quantum-classical framework. The flexibility of having neural networks allows us to apply transfer-learning technique to boost QZeros performance. We find both MCTS and QZero to perform very well in finding excellent annealing schedules even when the annealing time is short in the 3-SAT examples we consider in this study. We also find MCTS and QZero to be more efficient than many other leading reinforcement leanring algorithms for the task of desining annealing schedules. In particular, if there is a need to solve a large set of similar problems using a quantum annealer, QZero is the method of choice when the neural networks are first pre-trained with examples solved in the past.

rate research

Read More

Classical and quantum annealing are two heuristic optimization methods that search for an optimal solution by slowly decreasing thermal or quantum fluctuations. Optimizing annealing schedules is important both for performance and fair comparisons between classical annealing, quantum annealing, and other algorithms. Here we present a heuristic approach for the optimization of annealing schedules for quantum annealing and apply it to 3D Ising spin glass problems. We find that if both classical and quantum annealing schedules are similarly optimized, classical annealing outperforms quantum annealing for these problems when considering the residual energy obtained in slow annealing. However, when performing many repetitions of fast annealing, simulated quantum annealing is seen to outperform classical annealing for our benchmark problems.
165 - Satoshi Morita 2007
New annealing schedules for quantum annealing are proposed based on the adiabatic theorem. These schedules exhibit faster decrease of the excitation probability than a linear schedule. To derive this conclusion, the asymptotic form of the excitation probability for quantum annealing is explicitly obtained in the limit of long annealing time. Its first-order term, which is inversely proportional to the square of the annealing time, is shown to be determined only by the information at the initial and final times. Our annealing schedules make it possible to drop this term, thus leading to a higher order (smaller) excitation probability. We verify these results by solving numerically the time-dependent Schrodinger equation for small size systems
In a typical quantum annealing protocol, the system starts with a transverse field Hamiltonian which is gradually turned off and replaced by a longitudinal Ising Hamiltonian. The ground state of the Ising Hamiltonian encodes the solution to the computational problem of interest, and the state overlap with this ground state gives the success probability of the annealing protocol. The form of the annealing schedule can have a significant impact on the ground state overlap at the end of the anneal, so precise control over these annealing schedules can be a powerful tool for increasing success probabilities of annealing protocols. Here we show how superconducting circuits, in particular capacitively shunted flux qubits (CSFQs), can be used to construct quantum annealing systems by providing tools for mapping circuit flux biases to Pauli coefficients. We use this mapping to find customized annealing schedules: appropriate circuit control biases that yield a desired annealing schedule, while accounting for the physical limitations of the circuitry. We then provide examples and proposals that utilize this capability to improve quantum annealing performance.
Monte Carlo tree search (MCTS) has achieved state-of-the-art results in many domains such as Go and Atari games when combining with deep neural networks (DNNs). When more simulations are executed, MCTS can achieve higher performance but also requires enormous amounts of CPU and GPU resources. However, not all states require a long searching time to identify the best action that the agent can find. For example, in 19x19 Go and NoGo, we found that for more than half of the states, the best action predicted by DNN remains unchanged even after searching 2 minutes. This implies that a significant amount of resources can be saved if we are able to stop the searching earlier when we are confident with the current searching result. In this paper, we propose to achieve this goal by predicting the uncertainty of the current searching status and use the result to decide whether we should stop searching. With our algorithm, called Dynamic Simulation MCTS (DS-MCTS), we can speed up a NoGo agent trained by AlphaZero 2.5 times faster while maintaining a similar winning rate. Also, under the same average simulation count, our method can achieve a 61% winning rate against the original program.
The likelihood-free sequential Approximate Bayesian Computation (ABC) algorithms, are increasingly popular inference tools for complex biological models. Such algorithms proceed by constructing a succession of probability distributions over the parameter space conditional upon the simulated data lying in an $epsilon$--ball around the observed data, for decreasing values of the threshold $epsilon$. While in theory, the distributions (starting from a suitably defined prior) will converge towards the unknown posterior as $epsilon$ tends to zero, the exact sequence of thresholds can impact upon the computational efficiency and success of a particular application. In particular, we show here that the current preferred method of choosing thresholds as a pre-determined quantile of the distances between simulated and observed data from the previous population, can lead to the inferred posterior distribution being very different to the true posterior. Threshold selection thus remains an important challenge. Here we propose an automated and adaptive method that allows us to balance the need to minimise the threshold with computational efficiency. Moreover, our method which centres around predicting the threshold - acceptance rate curve using the unscented transform, enables us to avoid local minima - a problem that has plagued previous threshold schemes.
comments
Fetching comments Fetching comments
Sign in to be able to follow your search criteria
mircosoft-partner

هل ترغب بارسال اشعارات عن اخر التحديثات في شمرا-اكاديميا