13. Infinite Matrix Product States#
This section discusses matrix product states (MPS) in the thermodynamic limit and their properties. Our discussion is mostly based on the excellent review [Vanderstraeten et al., 2019], which provides a thorough technical overview of tangent-space methods for uniform MPS. The formal exposition is supplemented with some very basic code examples on working with infinite MPS using MPSKit.jl at the end of this section. For more details on the numerical implementation of routines for uniform MPS we refer to the Julia version of the tutorials on uniform MPS, which is again based on [Vanderstraeten et al., 2019].
Contents
13.1. Matrix Product States in the Thermodynamic Limit#
13.1.1. Representation#
The finite MPS representation introduced in the previous previous section can be readily extended to the thermodynamic limit by constructing a quantum state of an infinite spin system as a product of an infinite chain of tensors. For infinite systems which are invariant under translations, it is natural to also impose transation-invariance on the corresponding MPS. This leads to a uniform MPS which has the same tensor \(A^{(i)} := A\) at every site, where \(A\) again has a physical dimension \(d\) and bond dimension \(D\). In diagramatic notation, a uniform MPS can be represented as
Note
In some cases, instead of assuming an MPS has the same tensor at each site it is more natural to use a state with a non-trivial repeating unit cell. A uniform MPS with a unit cell of size three would for example correspond to the state
While we will restrict our discussion to MPS with a single-site unit cell, most concepts and techniques apply just as well to the multi-site unit cell case.
One of the central objects when working with MPS in the thermodynamic limit is the transfer operator or transfer matrix, defined in our case as
The transfer matrix corresponds to an operator acting on the space of \(D\times D\) matrices, and can be interpreted as a 4-leg tensor \(\mathbb C^D \otimes \mathbb C^D \leftarrow \mathbb C^D \otimes \mathbb C^D\). The transfer matrix can be shown to be a completely positive map, such that its leading eigenvalue is a positive number. The eigenvalues of the transfer matrix characterize the normalization and correlation length of a uniform MPS, while its eigenvectors can be used to evaluate expectation values of local observables.
13.1.2. Normalization#
The norm of a uniform MPS corresponds to a contraction of the form
Clearly, this norm is nothing more than an infinite product of MPS transfer matrices defined above. Consider the spectral decomposition of the \(n\)th power \(\mathbb E^n\),
where \(l\) and \(r\) are the left and right fixed points which correspond to the largest magnitude eigenvalue \(\lambda_0\) of \(\mathbb E\),
and the \(\lambda_i\) represent the remaining eigenvalues of smaller mangitude, where writing the spectral decomposition we have implicitly assumed that the fixed points are properly normalized as
Taking the limit of this spectral decomposition, it follows that the infinite product of transfer matrices reduces to a projector onto the fixed points corresponding to the leading eigenvalue \(\lambda_0\),
To ensure a properly normalized state we should therefore rescale the leading eigenvalue \(\lambda_0\) to one by rescaling the MPS tensor as \(A \leftarrow A / \sqrt{\lambda_0}\).
With these properties in place, the norm of an MPS reduces to the overlap between the boundary vectors and the fixed points. Since there is no effect of the boundary vectors on the bulk properties of the MPS, we can always choose these such that MPS is properly normalized as \( \left \langle \psi(\bar{A})\middle | \psi(A) \right \rangle = 1\).
13.1.3. Expectation Values#
The fixed points of the transfer matrix can for example be used to compute expectation values of operators. Suppose we wish to evaluate expectation values of an extensive operator,
If we assume that each \(O_n\) acts on a single site and we are working with a properly normalized MPS, translation invariance dictates that the expectation value of \(O\) is given by the contraction
In the uniform gauge, we can use the fixed points of the transfer matrix to contract everything to the left and to the right of the operator, such that we are left with the contraction
13.1.4. Correlation Functions#
Correlation functions are computed similarly. Let us look at
where \(m\) and \(n\) are abritrary locations in the chain, and, because of translation invariance, the correlation function only depends on the difference \(m-n\). Again, we contract everything to the left and right of the operators by inserting the fixed points \(l\) and \(r\), so that
From this expression, we learn that it is the transfer matrix that determines the correlations in the ground state. Indeed, if we again use the spectral decomposition of the transfer matrix, recalling that now \(\lambda_0 = 1\), we can see that the correlation function reduces to
The first part is just the product of the expectation values of \(O^\alpha\) and \(O^\beta\), called the disconnected part of the correlation function, and the rest is an exponentially decaying part. This expression implies that connected correlation functions of an MPS always decay exponentially, which is one of the reasons why MPS generally have a harder time dealing with critical states. The correlation length \(\xi\) is determined by the second largest eigenvalue of the transfer matrix \(\lambda_1\) as
Note
The subleading eigenvalues of the transfer matrix typically also have a physical meaning, because they correspond to subleading correlations in the system. For example, by focussing on eigenvalues in a specific symmetry sector one can target the correlations associated to exitations corresponding to that particular symmetry. The subleading eigenvalues also play a crucial role in the powerful technique of finite entanglement scaling for infinite MPS [Rams et al., 2018]. Using this framework we can accurately capture critical phenomena using MPS, despite the ansatz inherently having exponentially decaying correlations.
13.2. Gauging Revisited#
13.2.1. Gauging in the Thermodynamic Limit#
While a given MPS tensor \(A\) corresponds to a unique state \(\left | \psi(A) \right \rangle\), the converse is not true, as different tensors may give rise to the same state. This is easily seen by noting that the gauge transform
leaves the physical state invariant. We may use this freedom in parametrization to impose canonical forms on the MPS tensor \(A\).
We start by considering the left-orthonormal form of an MPS, which is defined in terms of a tensor \(A_L\) that satisfies the condition
We can find the gauge transform \(L\) that brings \(A\) into this form
using an iterative procedure based on the QR docomposition, where starting from some initial guess \(L^0\) we repeatedly perform the QR-based update
This iterative procedure is bound to converge to a fixed point for which \(L^{(i+1)}=L^{(i)}=L\) and \(A_L\) is left orthonormal by construction:
Note that this left gauge choice still leaves room for unitary gauge transformations
which can be used to bring the right fixed point \(r\) into diagonal form. Similarly, we can find the gauge transform that brings \(A\) into right-orthonormal form
such that
and the left fixed point \(l\) is diagonal. A right-orthonormal tensor \(A_R\) and a matrix \(R\) such that \(A R = R A_R\) can be found using a similar iterative procedure.
Finally, we can define a mixed gauge for the uniform MPS by choosing one site, the ‘center site’, and bringing all tensors to the left of it in the left-orthonormal form and all the tensors to the right of it in the right-orthonormal form. Defining a new tensor \(A_C\) on the center site, we obtain the form
By contrast, the original representation using the same tensor at every site is commonly referred to as the uniform gauge. The mixed gauge has an intuitive interpretation. Defining \(C = LR\), this tensor then implements the gauge transform that maps the left-orthonormal tensor to the right-orthonormal one, thereby defining the center-site tensor \(A_C\):
This relation is called the mixed gauge condition and allows us to freely move the center tensor \(A_C\) through the MPS, linking the left- and right orthonormal tensors.
Finally we may bring \(C\) into diagonal form by performing a singular value decomposition \(C = USV^\dagger\) and absorbing \(U\) and \(V^\dagger\) into the definition of \(A_L\) and \(A_R\) using the residual unitary gauge freedom
13.2.2. Expectation Values Revisited#
In the mixed gauge, we can locate the center site where the operator is acting, and then contract everything to the left and right to the identity to arrive at the particularly simple expression for the expectation value
13.2.3. Entanglement Entropy#
The mixed canonical form with a diagonal \(C\) now allows to straightforwardly write down a Schmidt decomposition of the state across an arbitrary bond in the chain
where the states \(\left | \psi^i_L(A_L) \right \rangle\) and \(\left | \psi^i_R(A_R) \right \rangle\) are orthogonal states on half the lattice. The diagonal elements \(C_i\) are exactly the Schmidt coefficient of any bipartition of the MPS, and as such determine its bipartite entanglement entropy
13.2.4. Truncation#
The mixed canonical form also enables efficient truncatation of an MPS. The sum in the above Schmidt decomposition can be truncated, giving rise to a new MPS that has a reduced bond dimension for that bond. This truncation is optimal in the sense that the norm between the original and the truncated MPS is maximized. To arrive at a translation invariant truncated MPS, we can truncate the columns of the absorbed isometries \(U\) and \(V^\dagger\) correspondingly, thereby transforming every tensor \(A_L\) or \(A_R\). The truncated MPS in the mixed gauge is then given by
We note that the resulting state based on this local truncation is not guaranteed to correspond to the MPS with a lower bond dimension that is globally optimal. This would require a variational optimization of the cost function.
13.2.5. Code Example: MPSKit.InfiniteMPS
#
The Julia package MPSKit.jl provides many tools for working with infinite MPS. Without going into much detail, we can already check some aspects of our discussion above with this numerical implementation.
We can construct an
MPSKit.InfiniteMPS
by specifying the physical and virtual vector spaces of the MPS. We will use standard
complex vector spaces as specified by a
TensorKit.ComplexSpace
,
and choose a physical dimension \(d = 3\) and bond dimension \(D = 5\).
using MPSKit, TensorKit
d = 3 # physical dimension
D = 5 # bond dimension
mps = InfiniteMPS(ℂ^d, ℂ^D)
single site InfiniteMPS:
│ ⋮
│ CR[1]: TensorMap(ℂ^5 ← ℂ^5)
├── AL[1]: TensorMap((ℂ^5 ⊗ ℂ^3) ← ℂ^5)
│ ⋮
The infinite MPS is automatically stored in the mixed canonical form introduced above. For example, we can check that its normalization is indeed characterized by the center gauge tensors \(A_C\) and \(C\).
using LinearAlgebra
@show norm(mps)
@show norm(mps.AC[1])
@show norm(mps.CR[1]);
norm(mps) = 1.0000000000000002
norm(mps.AC[1]) = 1.0000000000000002
norm(mps.CR[1]) = 1.0
We can also explicitly verify the mixed gauge conditions on \(A_L\), \(A_R\), \(A_C\) and \(C\) by
evaluating the corresponding tensor network diagrams using the
TensorOperations.@tensor
macro.
using TensorOperations
@tensor AL_id[-1; -2] := mps.AL[1][1 2; -2] * conj(mps.AL[1][1 2; -1])
@tensor AR_id[-1; -2] := mps.AR[1][-1 1; 2] * conj(mps.AR[1][-2 1; 2])
@assert AL_id ≈ id(space(mps.AL[1], 3)') "AL not in left-orthonormal form!"
@assert AR_id ≈ id(space(mps.AR[1], 1)) "Ar not in right-orthonormal form!"
@tensor LHS[-1 -2; -3] := mps.AL[1][-1 -2; 1] * mps.CR[1][1; -3]
@tensor RHS[-1 -2; -3] := mps.CR[1][-1; 1] * mps.AR[1][1 -2; -3]
@assert LHS ≈ RHS && RHS ≈ mps.AC[1] "Center gauge MPS tensor not consistent!"
We can also easily evaluate the expectation value of local operators
O = TensorMap(randn, ℂ^d ← ℂ^d)
expectation_value(mps, O)
1-element Vector{ComplexF64}:
-1.2389355854733874 - 0.10352095070374498im
as well as compute the correlation length encoded in the MPS.
correlation_length(mps)
0.49884041326571316
MPSKit.jl exports a variety of infinite MPS algorithms, some of which will be discussed in the next section.