SciPost Submission Page
Gauging tensor networks with belief propagation
by Joseph Tindall, Matt Fishman
This is not the latest submitted version.
This Submission thread is now published as
|Authors (as registered SciPost users):||Joseph Tindall|
|Preprint Link:||https://arxiv.org/abs/2306.17837v2 (pdf)|
|Date submitted:||2023-08-21 18:31|
|Submitted by:||Tindall, Joseph|
|Submitted to:||SciPost Physics|
Effectively compressing and optimizing tensor networks requires reliable methods for fixing the latent degrees of freedom of the tensors, known as the gauge. Here we introduce a new algorithm for gauging tensor networks using belief propagation, a method that was originally formulated for performing statistical inference on graphical models and has recently found applications in tensor network algorithms. We show that this method is closely related to known tensor network gauging methods. It has the practical advantage, however, that existing belief propagation implementations can be repurposed for tensor network gauging, and that belief propagation is a very simple algorithm based on just tensor contractions so it can be easier to implement, optimize, and generalize. We present numerical evidence and scaling arguments that this algorithm is faster than existing gauging algorithms, demonstrating its usage on structured, unstructured, and infinite tensor networks. Additionally, we apply this method to improve the accuracy of the widely used simple update gate evolution algorithm.
Submission & Refereeing History
You are currently on this page
Reports on this Submission
- Cite as: Anonymous, Report on arXiv:2306.17837v2, delivered 2023-10-11, doi: 10.21468/SciPost.Report.7931
1- very clearly written, with a lot of background information and comprehensive list of references; the text is also a good review work for gauge fixing in tensor networks;
2- new algorithm for gauge fixing a general tensor network, providing an approximate local environment for each tensor;
3- benchmarks showing good performance in different contexts.
No apparent weaknesses.
In this work, the authors introduce a new algorithm for finding an effective gauge of a general tensor networks. For the case of loop-free tensor networks (matrix product states and tree tensor networks), this algorithm reduces to the standard canonical form; for "loopy" tensor networks this algorithm converges to what the authors call the "Vidal gauge". The algorithm is explained in good detail and carefully benchmarked.
As the authors show, this gauge is often used implicitly in simple-update-style algorithms, which is often quite useful, but also often fails in capturing the environment accurately. The authors clearly discuss this drawback and make it clear that these types of algorithms should be used with this in mind.
Therefore, this is an interesting work that pushes the efficiency of simple-update-style algorithms, and is therefore expected to be commonly used in situations with only tree-like quantum correlations.
I recommend publication without any requested changes.
I don't request any changes.
- Cite as: Anonymous, Report on arXiv:2306.17837v2, delivered 2023-09-29, doi: 10.21468/SciPost.Report.7878
1- the intersection of tensor networks and tools such as BP is ripe for exploration. This paper both clarifies the links between BP and other algorithms such as Simple Update (SU), including the explicit gauge relation, and shows the practical advantage to using BP.
2- The paper is very clearly explained with diagrams and also structured, making it a good introductory and reference paper.
3- There is sufficient detail to reproduce everything, and the actual code is open source.
1- Theoretically it was already known that BP and SU are somehow equivalent and should converge to the same thing, and thus a critical reader might argue that the techniques introduced here don't fundamentally change what simulations are accessible c.f. SU.
2- The actual results shown are all at quite small scale and show fairly mild improvements as compared to SU or no re-gauging. Again a critical reader might find these underwhelming, though my personal opinion is that BP will be indeed be very useful for large simulations, for various reasons including beyond pure performance.
I think this is a very timely and clear paper, that makes another important link between BP and tensor networks, via the new gauging relation, and also shows the practical advantages of using. I think the paper is definitely worthy of publication, since, although one might argue there is not radically new theory or concepts, it provides a very clear, detailed and practical demonstration and how and why to use BP for tensor networks, a combination of methods that is likely to be very common in the future. I do think ideally the small points raised below should be addressed in some way.
One more minor comment for the authors:
1. the gauging procedure is the same as the projectors inserted in CTMRG and HOTRG, I don't know if this is a useful link for the authors to make or think about, or is maybe too trivial.
1- For the comparison with SU / eager BP and BP. As far as I understand for BP implemented here a 'parallel update' is always used (i.e. messages at step $t$ depend only on messages at step $t -1$). Whereas SU can be thought of as a sequential update. One is left wondering whether the difference in convergence is simply this difference. It might be nice to check with a 'sequential BP', where updated messages immediately replace the old ones. I think this is different from the eager BP shown - there is no gauging between steps required. While there are many convergence BP techniques and for one thing this might speed things up, more interestingly/importantly it might show whether BP and SU really converge at exactly the same rate (in terms of iterations) using the same update schedule. However, I would consider exploring this fully not strictly required.
2- The authors briefly discuss generalized BP and dub 'generalized BP gauging'. Though I am not an expert, I think what the authors refer to, including the referenced tensor network papers, is really still just BP but with larger 'regions'. Whereas a generalized BP algorithm (beyond the most simple case of BP itself) actually involves *overlapping* regions and a considerably more complicated relationship between the marginals and messages being passed around. As such, the name 'generalized BP gauging' might be a bit premature.