APS Logo

Differentiable Programming Tensor Networks

ORAL

Abstract

Differentiable programming is a fresh programming paradigm which composes parameterized algorithmic components and trains them using automatic differentiation (AD). We present theory and practice of programming tensor network algorithms in a fully differentiable way. By formulating the tensor network algorithm as a computation graph, one can compute higher order derivatives of the program accurately and efficiently using AD. We present essential techniques to differentiate through the tensor networks contractions, including stable AD for tensor decomposition and efficient backpropagation through fixed point iterations. As a demonstration, we compute the specific heat of the Ising model directly by taking the second order derivative of the free energy obtained in the tensor renormalization group calculation. Next, we perform gradient based variational optimization of infinite projected entangled pair states for quantum antiferromagnetic Heisenberg model and obtain start-of-the-art variational energy and magnetization with moderate efforts. Differentiable programming removes laborious human efforts in deriving and implementing analytical gradients for tensor network programs, which opens the door to more innovations in tensor network algorithms and applications.

Presenters

  • Hai-Jun Liao

    Institute of Physics, The Chinese Academy of Sciences

Authors

  • Hai-Jun Liao

    Institute of Physics, The Chinese Academy of Sciences

  • Jin-Guo Liu

    Institute of Physics, The Chinese Academy of Sciences

  • Lei Wang

    Institute of Physics, Institute of Physics, The Chinese Academy of Sciences, Chinese Academy of Sciences,Institute of Physics, Institute of Physics, Chinese Academy of Sciences

  • Tao Xiang

    Institute of Physics, The Chinese Academy of Sciences