ترغب بنشر مسار تعليمي؟ اضغط هنا

Monte Carlo Tensor Network Renormalization

188   0   0.0 ( 0 )
 نشر من قبل Norm Tubman
 تاريخ النشر 2017
  مجال البحث فيزياء
والبحث باللغة English




اسأل ChatGPT حول البحث

Techniques for approximately contracting tensor networks are limited in how efficiently they can make use of parallel computing resources. In this work we demonstrate and characterize a Monte Carlo approach to the tensor network renormalization group method which can be used straightforwardly on modern computing architectures. We demonstrate the efficiency of the technique and show that Monte Carlo tensor network renormalization provides an attractive path to improving the accuracy of a wide class of challenging computations while also providing useful estimates of uncertainty and a statistical guarantee of unbiased results.



قيم البحث

اقرأ أيضاً

279 - A. W. Sandvik , G. Vidal 2007
We show that the formalism of tensor-network states, such as the matrix product states (MPS), can be used as a basis for variational quantum Monte Carlo simulations. Using a stochastic optimization method, we demonstrate the potential of this approac h by explicit MPS calculations for the transverse Ising chain with up to N=256 spins at criticality, using periodic boundary conditions and D*D matrices with D up to 48. The computational cost of our scheme formally scales as ND^3, whereas standard MPS approaches and the related density matrix renromalization group method scale as ND^5 and ND^6, respectively, for periodic systems.
We propose a second renormalization group method to handle the tensor-network states or models. This method reduces dramatically the truncation error of the tensor renormalization group. It allows physical quantities of classical tensor-network model s or tensor-network ground states of quantum systems to be accurately and efficiently determined.
We develop coarse-graining tensor renormalization group algorithms to compute physical properties of two-dimensional lattice models on finite periodic lattices. Two different coarse-graining strategies, one based on the tensor renormalization group a nd the other based on the higher-order tensor renormalization group, are introduced. In order to optimize the tensor-network model globally, a sweeping scheme is proposed to account for the renormalization effect from the environment tensors under the framework of second renormalization group. We demonstrate the algorithms by the classical Ising model on the square lattice and the Kitaev model on the honeycomb lattice, and show that the finite-size algorithms achieve substantially more accurate results than the corresponding infinite-size ones.
The key idea behind the renormalization group (RG) transformation is that properties of physical systems with very different microscopic makeups can be characterized by a few universal parameters. However, finding the optimal RG transformation remain s difficult due to the many possible choices of the weight factors in the RG procedure. Here we show, by identifying the conditional distribution in the restricted Boltzmann machine (RBM) and the weight factor distribution in the RG procedure, an optimal real-space RG transformation can be learned without prior knowledge of the physical system. This neural Monte Carlo RG algorithm allows for direct computation of the RG flow and critical exponents. This scheme naturally generates a transformation that maximizes the real-space mutual information between the coarse-grained region and the environment. Our results establish a solid connection between the RG transformation in physics and the deep architecture in machine learning, paving the way to further interdisciplinary research.
124 - Glen Evenbly 2015
We discuss in detail algorithms for implementing tensor network renormalization (TNR) for the study of classical statistical and quantum many-body systems. Firstly, we recall established techniques for how the partition function of a 2D classical man y-body system or the Euclidean path integral of a 1D quantum system can be represented as a network of tensors, before describing how TNR can be implemented to efficiently contract the network via a sequence of coarse-graining transformations. The efficacy of the TNR approach is then benchmarked for the 2D classical statistical and 1D quantum Ising models; in particular the ability of TNR to maintain a high level of accuracy over sustained coarse-graining transformations, even at a critical point, is demonstrated.
التعليقات
جاري جلب التعليقات جاري جلب التعليقات
سجل دخول لتتمكن من متابعة معايير البحث التي قمت باختيارها
mircosoft-partner

هل ترغب بارسال اشعارات عن اخر التحديثات في شمرا-اكاديميا