Corpus ID: 209531831

Accuracy and Efficiency of Simplified Tensor Network Codes

@article{Yevick2020AccuracyAE,
  title={Accuracy and Efficiency of Simplified Tensor Network Codes},
  author={D. Yevick and Jesse Thompson},
  journal={arXiv: Statistical Mechanics},
  year={2020}
}
We examine in detail the accuracy, efficiency and implementation issues that arise when a simplified code structure is employed to evaluate the partition function of the two-dimensional square Ising model on periodic lattices though repeated tensor contractions. 

Figures from this paper

The Accuracy of Restricted Boltzmann Machine Models of Ising Systems
TLDR
This article examines in detail the influence of hyperparameters such as the learning rate, the number of hidden nodes and the form of the threshold function on Ising spin system calculations. Expand
Electric Field Propagation Through Singular Value Decomposition
We demonstrate that the singular value decomposition algorithm in conjunction with the fast Fourier transform or finite difference procedures provides a straightforward and accurate method forExpand
Variational Autoencoder Analysis of Ising Model Statistical Distributions and Phase Transitions
Variational autoencoders employ an encoding neural network to generate a probabilistic representation of a data set within a low-dimensional space of latent variables followed by a decoding stageExpand

References

SHOWING 1-10 OF 34 REFERENCES
TensorNetwork: A Library for Physics and Machine Learning
TLDR
The use of the TensorNetwork API is demonstrated with applications both physics and machine learning, with details appearing in companion papers. Expand
Renormalization of tensor networks using graph independent local truncations
We introduce an efficient algorithm for reducing bond dimensions in an arbitrary tensor network without changing its geometry. The method is based on a quantitative understanding of localExpand
Generalization of the tensor renormalization group approach to 3-D or higher dimensions
In this paper, a way of generalizing the tensor renormalization group (TRG) is proposed. Mathematically, the connection between patterns of tensor renormalization group and the concept of truncationExpand
Algorithms for tensor network renormalization
We discuss in detail algorithms for implementing tensor network renormalization (TNR) for the study of classical statistical and quantum many-body systems. First, we recall established techniques forExpand
Tensor network states and algorithms in the presence of a global SU(2) symmetry
Tensor network decompositions offer an efficient description of certain many-body states of a lattice system and are the basis of a wealth of numerical simulation algorithms. In a recent paper [Phys.Expand
Fast convergence of imaginary time evolution tensor network algorithms by recycling the environment
We propose an environment recycling scheme to speed up a class of tensor network algorithms that produce an approximation to the ground state of a local Hamiltonian by simulating an evolution inExpand
Tensor Network Renormalization.
We introduce a coarse-graining transformation for tensor networks that can be applied to study both the partition function of a classical statistical system and the Euclidean path integral of aExpand
TensorNetwork for Machine Learning
TLDR
The encoding of image data into a matrix product state form is explained in detail, and how to contract the network in a way that is parallelizable and well-suited to automatic gradients for optimization is described. Expand
Boundary tensor renormalization group
We develop the tensor renormalization group (TRG) algorithm for statistical systems with open boundaries, which allows us to investigate not only the bulk but also the boundary property, such as theExpand
An experimental realization of biased multicanonical sampling
We demonstrate that the properties of functions of stochastic system parameters can be rapidly determined experimentally within any desired region of the system observables through the biasedExpand
...
1
2
3
4
...