Abstract
The solution of linear non-autonomous ordinary differential equation systems (also known as the time-ordered exponential) is a computationally challenging problem arising in a variety of applications. In this work, we present and study a new framework for the computation of bilinear forms involving the time-ordered exponential. Such a framework is based on an extension of the non-Hermitian Lanczos algorithm to 4-mode tensors. Detailed results concerning its theoretical properties are presented. Moreover, computational results performed on real-world problems confirm the effectiveness of our approach.
Similar content being viewed by others
1 Introduction
In this paper, we present an extension of the non-Hermitian Lanczos algorithm (see, e.g., [25]) where the inputs are a 4-mode tensor \(\boldsymbol {\mathcal {A}} \in \mathbb {C}^{N \times N \times M \times M}\) and vectors \(\mathbf {w}, \mathbf {v} \in \mathbb {C}^{N}\) so that wHv≠ 0. We aim to use the introduced algorithm to approximate the bilinear form wHU(t)v, where \(\boldsymbol {\mathsf {U}}(t) \in \mathbb {C}^{N \times N}\) is the so-called time-ordered exponential, i.e., the solution of the ordinary differential equation
where \(I_{N}\in \mathbb {R}^{N\times N}\) is the identity matrix and \(\boldsymbol {\mathsf {A}}(t) \in \mathbb {C}^{N \times N}\) is a smooth matrix-valued function defined on the real interval I. Equation (1) can emerge in a variety of applications. For example, its solution is crucial in quantum physics where the matrix A(t) corresponds to the Hamiltonian operator. Situations where U(t) has no accessible expression are frequent in literature, see, e.g., [1, 7, 38, 49]. For instance, in Nuclear Magnetic Resonance (NMR) experiments, the associated bilinear form wHU(t)v represents the measurement of changes in an applied magnetic field caused by nuclear spins that are excited with electromagnetic waves, i.e., spectroscopy [29, 39]. Other applications are found in control theory, filter design, and model reduction problems [4, 5, 14, 37, 47]. In the mentioned applications, the matrix A(t) is often large-to-huge and sparse. The introduced algorithm is motivated and theoretically supported by a new expression for the bilinear form. This expression is given by combining the two symbolic methods known as Path-sum and ⋆-Lanczos algorithm [21,22,23,24]. Given the matrix-valued function A(t) and the vectors w,v, the two symbolic methods produce an expression for the bilinear form wHU(t)v composed of a finite and treatable number of integrals and scalar integral equations. To our knowledge, no other symbolic method can express the bilinear form with a treatable finite number of integral subproblems. Two commonly used alternative expressions are given by the Magnus series, i.e., an infinite series of nested integrals (e.g., [40]), and by the Floquet theory, where the solution of an infinite system of coupled linear differential equations is required (e.g., [7]).
The integrals and the integral equations generated by the ⋆-Lanczos and Path-sum methods do not always have an easily accessible solution. As a consequence, a numerical approach is needed. A possible strategy for the numerical approximation of the mentioned integrals and the integral equations is outlined in [23] and it is based on the discretization of the interval I into M − 1 equispaced subintervals. The algebraic objects resulting from the discretization strategy are the 4-mode tensor \(\boldsymbol {\mathcal {A}}\) (corresponding to A(t)) and the 3-mode tensors V,W (corresponding to v,w).
The outputs obtained by combining the ⋆-Lanczos algorithm with the mentioned discretization strategy are mathematically equivalent to the outputs of the tensor Lanczos algorithm presented here with, as inputs, \(\boldsymbol {\mathcal {A}}, \mathbf {v}, \mathbf {w}\). The main goal of this paper is to show that, in fact, the tensor Lanczos algorithm can converge to the outcome of the ⋆-Lanczos method within an accuracy of the same order as the discretization strategy. Moreover, the reported numerical experiments will show that the approximation of wHU(t)v obtained by combining the tensor Lanczos with the discretized Path-sum approach also converges to the solution within the order of the discretization. Naturally, many numerical methods for the solution of non-autonomous ODEs can be found in literature, see, for instance, [2, 6, 7, 10, 13, 15, 30, 34,35,36]. For large matrices, these numerical methods are known to be highly demanding both in terms of computational cost and storage. This motivates the research of novel approaches suitable for large-scale problems. In order to be competitive with the most advanced techniques, tensor Lanczos needs to be used in combination with more accurate discretization schemes. Development of suitable, faster converging discretization schemes is an ongoing research and out of the scope of this work. At the same time, it is important to note that the algorithm here proposed is part of a wider class of tensor extensions of Krylov subspace methods that recently appeared in the literature, see, e.g., [18, 26, 33, 46].
The Lanczos-type process we introduce can also be equivalently written as a block Lanczos method since the 4-mode tensor \(\boldsymbol {\mathcal {A}}\) can be seen as a block matrix; information about block Krylov subspace methods can be found, e.g., in [20]. Despite this fact, we prefer to interpret such a block structure in a tensorial fashion. Indeed, the tensorial approach has a direct translation in terms of a discretized ⋆-Lanczos algorithm. Moreover, as we will experimentally show, this interpretation is motivated by observing that several tensors from real-world examples related with (1) are characterized by a low parametric approximation known as Tensor Train decomposition (TT) [41, 42]. Such a low parametric approximation allows to efficiently manipulate and store the tensors. This paves the path for further improvements of our proposal, where the TT structure is fully exploited in the Lanczos-type procedure. Examples of tensor Krylov subspace methods combined with the TT decomposition can be found in [18, 48], further motivating our tensor-based point of view.
More in detail, this work is organized as follows. Preliminaries and definitions of tensor operations are given in Section 2. In Section 3 we discuss how to construct the non-Hermitian Lanczos procedure for tensors and we prove several crucial properties. In Section 4 we discuss the breakdown issue which typically arises when working with non-Hermitian Lanczos approaches. Numerical experiments are presented in Section 5 where we also give several examples exposing the low-rank TT structure of the considered tensors \(\boldsymbol {\mathcal {A}}\). Section 6 concludes the paper and A contains several proofs.
2 Preliminaries
In this work, we use a notation borrowed from Matlab®;. Fixing \(i_{1}\in \{1,\dots ,N_{1}\}\) and \(i_{2}\in \{1,\dots ,N_{2}\}\), if \(\boldsymbol {\mathcal {A}} \in \mathbb {C}^{N_{1} \times N_{2} \times M \times M }\), then \(\boldsymbol {\mathcal {A}}_{i_{1},i_{2},:,:}\) stands for the matrix
This notation similarly applies to 3-mode tensors, matrices, and vectors. Table 1 summarizes the notation used in the paper.
In the following, we define several tensorial operations, which can be seen as generalizations of the usual products involving matrices and vectors. We summarize them in Table 2.
In the following definitions we consider the tensors \(\boldsymbol {\mathcal {A}} \in \mathbb {C}^{N_{1}\times N_{2} \times M \times M },\boldsymbol {{\mathscr{B}}} \in \mathbb {C}^{N_{2} \times N_{3} \times M \times M }\), \(A \in \mathbb {C}^{N_{2} \times M \times M }\), \(B \in \mathbb {C}^{N_{1} \times M \times M }\), \(\boldsymbol {\alpha } \in \mathbb {C}^{M \times M}\). Moreover, the indices \(i_{1}\in \{1,\dots ,N_{1}\}\) \(i_{2}\in \{1,\dots ,N_{2}\}\) are fixed.
Definition 1 (∗-Tensor product)
The product \((\boldsymbol {\mathcal {A}}*\boldsymbol {{\mathscr{B}}})\in \mathbb {C}^{N_{1} \times N_{3} \times M \times M } \) is defined as
Definition 2 (Tensor-Hypervector product)
The product \((\boldsymbol {\mathcal {A}}*A) \in \mathbb {C}^{N_{1} \times M \times M }\) is defined as
We also need to define the action of a 3-mode tensor from the left. Every tensor with three modes that acts, will act, or is the outcome of a ∗-product from the left, will be denoted with a “D” (dual) as apex, and, in the remainder of this work, we will use \(B^{D}_{k,:,:}\) to denote (BD)k,:,:. We define, \( (B^{D}*\boldsymbol {\mathcal {A}})^{D}\in \mathbb {C}^{N_{2} \times M \times M }\) as
Note that the following 4-mode tensor is the identity for ∗-products introduced above
Definition 3 (Hypervector inner-product)
The product \((B^{D}*A) \in \mathbb {C}^{M\times M}\) is defined as
Definition 4 (Tensor-matrix product)
The products \((\boldsymbol {\mathcal {A}} \times \boldsymbol {\alpha }), (\boldsymbol {\alpha } \times \boldsymbol {\mathcal {A}} ) \in \mathbb {C}^{N_{1} \times N_{2} \times M \times M }\) are defined as
Definition 5 (Hypervector-matrix product)
The products \(({A} \times \boldsymbol {\alpha }), (\boldsymbol {\alpha } \times {A} ) \in \mathbb {C}^{N_{2} \times M \times M }\) are defined as
Definition 6 (Vector-to-Hypervector)
Given \(\mathbf {a} \in \mathbb {C}^{N}\) we define the product \( A=\mathbf {a}\otimes I_{M} \in \mathbb {C}^{N \times M \times M}\) as
Note that rearranging A as a block matrix, we get the usual Kronecker product. All the products are clearly distributive with respect to the usual addition. On the other hand, the associativity of some of the products is less obvious. Therefore, we state it in the following Lemma 1, postponing its proof to A.
Lemma 1
The following states that the tensor-tensor and tensor-hypervector ∗-products are associative.
-
Given \(\boldsymbol {\mathcal {A}} \in \mathbb {C}^{N_{1} \times N_{1} \times M \times M}\), \(A \in \mathbb {C}^{N_{1} \times M \times M}\) we have
$$ (\boldsymbol{\mathcal{A}}*\boldsymbol{\mathcal{A}})*{A}= \boldsymbol{\mathcal{A}}*(\boldsymbol{\mathcal{A}}*{A}){.} $$ -
Given \(B^{D} \in \mathbb {C}^{N_{1} \times M \times M}\), \({A} \in \mathbb {C}^{N_{2} \times M \times M }\), \(\boldsymbol {\mathcal {A}} \in \mathbb {C}^{N_{1}\times N_{2} \times M \times M }\), then
$$ (B^{D} * \boldsymbol{\mathcal{A}})^{D} *A=B^{D} * (\boldsymbol{\mathcal{A}} *A). $$ -
Given \(\boldsymbol {\mathcal {A}} \in \mathbb {C}^{N_{1}\times N_{2} \times M \times M }, \boldsymbol {{\mathscr{B}}} \in \mathbb {C}^{N_{2} \times N_{3} \times M \times M }, \boldsymbol {\mathcal {C}} \in \mathbb {C}^{N_{3}\times N_{1} \times M \times M }\), then
$$ (\boldsymbol{\mathcal{C}}*\boldsymbol{\mathcal{A}})*\boldsymbol{\mathcal{B}}=\boldsymbol{\mathcal{C}}*(\boldsymbol{\mathcal{A}}*\boldsymbol{\mathcal{B}}). $$
Having introduced the required products and their basic properties, we are ready to derive the tensor non-Hermitian Lanczos algorithm.
3 The Lanczos-type process
Using the operations given in Table 2, we propose a sensible generalization of Krylov subspaces where, instead of the usual matrix-vector product, the tensor-hypervector product is used to generate the subspaces. Section 3.1 describes these tensor Krylov-type subspaces in detail and defines biorthogonal bases for them. A Lanczos-type algorithm which generates these biorthogonal bases is proposed in Section 3.2. In Section 3.3 two important properties of the classical Lanczos algorithm are generalized, namely, the tensor representation of the three-term recurrence relations for the biorthogonal bases and the matching moment property. The computational cost and storage requirements of the algorithm are discussed in Section 3.4.
3.1 Krylov-type tensor subspaces
Consider the tensor \(\boldsymbol {\mathcal {A}}\in \mathbb {C}^{N \times N \times M \times M}\). We define the polynomials of degree ℓ of \(\boldsymbol {\mathcal {A}}\) as
where \(\boldsymbol {\mathcal {A}}^{k_{*}}\) stands for k ∗-multiplications of \(\boldsymbol {\mathcal {A}}\) by itself, and \(\boldsymbol {\alpha }_{k}^{H}\) is the conjugate transpose of α. Given the tensors \({A}\in \mathbb {C}^{N \times M \times M}\), \(B\in \mathbb {C}^{N \times M \times M}\) we can define the Krylov-type subspaces
Every element in \(\mathcal {K}_{n}(\boldsymbol {\mathcal {A}},A)\) is a tensor in \(\mathbb {C}^{N \times M \times M}\) and can be written as
From now on we will assume that A is of the form A = a ⊗ IM for some \(\mathbf {a} \in \mathbb {C}^{N}\). In this case, the matrices αk commute with A giving
An analogous result holds for \(B^{D}*p^{D}(\boldsymbol {\mathcal {A}})\) when \(B^D=\overline {\mathbf {b}} \otimes I_M\) for any \(\mathbf {b} \in \mathbb {C}^{N}\), \(\overline {\mathbf {b}}\) is the conjugated vector.
Driven by the analogy with the matrix case, our aim is to build two “biorthonormal bases” for the Krylov-type subspaces \(\mathcal {K}_{n}(\boldsymbol {\mathcal {A}},A)\) and \( \mathcal {K}_{n}^{D}(B^{D}, \boldsymbol {\mathcal {A}})\). The following Definition 7 allows to characterize spaces spanned by 3-mode tensors.
Definition 7
Given \(V_{1},\dots ,V_{n} \in \mathbb {C}^{N \times M \times M}\), \({W_{1}^{D}},\dots ,{W^{D}_{n}} \in \mathbb {C}^{N \times M \times M}\), we define the subspaces
We say that \(V_{1},\dots , V_{n}\) is a basis for the subspace \(\langle V_{1},\dots , V_{n} \rangle \) and \({W_{1}^{D}},\dots , {W_{n}^{D}}\) is a basis for the subspace \(\langle {W_{1}^{D}},\dots , {W_{n}^{D}} \rangle \).
Biorthonormal bases for Krylov-type subspaces are represented by the tensors \(\boldsymbol {\mathcal {V}}_{n} \in \mathbb {C}^{N\times n \times M \times M }\) and \(\boldsymbol {\mathcal {W}}_{n}\in \mathbb {C}^{n\times N \times M \times M }\) satisfying
with the hypervectors \(V_{k}:= (\boldsymbol {\mathcal {V}}_{n})_{:,k,:,:}\) and \({W^{D}_{k}}:= (\boldsymbol {\mathcal {W}}_{n})_{k,:,:,:}\), for \(k=1,\dots ,n\), forming, respectively, bases for \(\mathcal {K}_{n}(\boldsymbol {\mathcal {A}},A)\) and \( \mathcal {K}_{n}^{D}(B^{D}, \boldsymbol {\mathcal {A}})\), i.e.,
In the following section we derive such bases by constructing the tensor non-Hermitian Lanczos Algorithm.
3.2 The tensor Lanczos process
Given the inputs \(\boldsymbol {\mathcal {A}}\in \mathbb {C}^{N\times N \times M \times M}\) and \(\mathbf {v}, \mathbf {w} {\in \mathbb {C}^{N}}\), Algorithm 1 constructs, when no breakdown occurs, the bases \(\boldsymbol {\mathcal {W}}_{n}\) and \(\boldsymbol {\mathcal {V}}_{n}\), for \(\mathcal {K}_{n}(\boldsymbol {\mathcal {A}},A)\) and \( \mathcal {K}_{n}^{D}(B^{D}, \boldsymbol {\mathcal {A}})\), respectively, which satisfy the ∗-biorthogonality conditions (2).
Details on how the algorithm constructs these bases using three-term recurrences are described below.
-
By definition, the first hypervectors of the bases are \({W_{1}^{D}}, V_{1}\) satisfying \({W_{1}^{D}}*V_{1}=I_{M}\);
-
Consider the vector \(\widehat {W}^{D}_{2} \in \mathcal {K}^{D}_{2}(W^{D},\boldsymbol {\mathcal {A}})\) given by
$$ \widehat{W}^{D}_{2}:={W_{1}^{D}}*\boldsymbol{\mathcal{A}}-\boldsymbol {\alpha}_{1} \times {W_{1}^{D}}. $$Imposing that \(\widehat {W}_{2}^{D}\) satisfies the ∗-biorthogonal condition \(\widehat {W}_{2}^{D}*V_{1}=0\), we have \(\boldsymbol {\alpha }_{1}={W_{1}^{D}} * \boldsymbol {\mathcal {A}} *V_{1}\).
-
Analogously, define the vector \(\widehat {V}_{2} \in \mathcal {K}_{2}(\boldsymbol {\mathcal {A}},V)\) given by
$$ \widehat{V}_{2} := \boldsymbol{\mathcal{A}}*V_{1}-V_{1}\times \boldsymbol {\alpha}_{1}. $$Imposing the ∗-biorthogonality condition, we find the ∗-biorthogonal vectors
$$ V_{2}:=\widehat{V}_{2}\times \boldsymbol {\beta}_{1}^{-1} \text{ where } \boldsymbol {\beta}_{1}=\widehat{W}_{2}^{D}*\widehat{V}_{2}=\widehat{W}_{2}^{D}*V_{1} \text{ and } W_{2}=\widehat{W}_{2}. $$ -
Clearly \(\mathcal {K}_{2}({\boldsymbol {\mathcal {A}},V})=\langle V_{1},V_{2} \rangle \) and \(\mathcal {K}_{2}^{D}(W^{D},{\boldsymbol {\mathcal {A}}})=\langle {W^{D}_{1}},{W^{D}_{2}} \rangle \).
-
Now, assume the ∗-biorthonormal bases \(V_{1},\dots ,V_{{k}}\) and \({W^{D}_{1}},\dots ,W^{D}_{{k}}\) are available. Consider the hypervector
$$ \widehat{W}^{D}_{{k}+1}:={W}_{{k}}^{D}*\boldsymbol{\mathcal{A}}-{\sum}_{i=1}^{{k}}\boldsymbol{\boldsymbol{\eta}}_{i} \times {W}_{i}^{D}. $$The matrices ηi are determined by the conditions \(\widehat {W}_{{k}+1}^{D}*V_{i}=0\), for \(i~=~1,\dots ,~k\), which give
$$ \boldsymbol{\eta}_{i}=W_{{k}}^{D}*\boldsymbol{\mathcal{A}}*V_{i}, \quad \text{ for } i=1,\dots,{k}. $$In particular, since \(\boldsymbol {\mathcal {A}}*V_{i} \in \mathcal {K}_{i+1}(\boldsymbol {\mathcal {A}},V )\), we get ηi = 0 for \(i=1,\dots ,{k}-2\). An analogous argument is valid for \(\widehat {V}_{k+1}\). This leads to the following three-term recurrences
$$ \begin{array}{@{}rcl@{}} W_{k+1}^{D}={W_{k}^{D}}*\boldsymbol{\mathcal{A}}-\boldsymbol {\alpha}_{k}\times {W_{k}^{D}}-\boldsymbol {\beta}_{k}\times W_{k-1}^{D}, \end{array} $$(3a)$$ \begin{array}{@{}rcl@{}} V_{k+1}\times \boldsymbol {\beta}_{k+1}=\boldsymbol{\mathcal{A}} *V_{k}- V_{k}\times \boldsymbol {\alpha}_{k}-V_{k-1}, \end{array} $$(3b)with coefficients
$$ \boldsymbol {\alpha}_{k}= {W_{k}^{D}}*\boldsymbol{\mathcal{A}}*V_{k}, \boldsymbol {\beta}_{k+1}=W_{k+1}^{D}*\boldsymbol{\mathcal{A}}*V_{k}. $$(4) -
To prove that \(\langle V_{1}, \dots , V_{n} \rangle =\mathcal {K}_{n}(\boldsymbol {\mathcal {A}},V)\) and \(\langle {W_{1}^{D}}, \dots , {W_{n}^{D}} \rangle =\mathcal {K}_{n}(W^{D},\boldsymbol {\mathcal {A}})\), it is enough to use induction and the fact that \(V_{{k}} \in \mathcal {K}_{{k}}(\boldsymbol {\mathcal {A}},V)\) and \(W_{{k}}^{D} \in \mathcal {K}^{D}_{{k}}(W^{D},\boldsymbol {\mathcal {A}})\) for all \({k}=1,\dots ,n\).
Let us finally observe that, should βk+ 1 not be invertible, we would get a breakdown in the algorithm.
Different rescaling strategies are possible by setting an invertible coefficient γk+ 1 and noticing that
This last observation completes the construction of Algorithm 1.
3.3 Main properties of the tensor Lanczos algorithm
It is important to note that the coefficients in the three-term recurrences (3.2–3.2) can be represented by a sparse 4-mode tensor. To this aim, let us consider \(\boldsymbol {\mathcal {T}}_{n} \in \mathbb {C}^{n \times n \times M \times M}\) as the tensor defined as
where \(\boldsymbol {\alpha }_{i_{1}}, \boldsymbol {\beta }_{i_{1}}, \boldsymbol {\gamma }_{i_{1}}\) are the matrices in Algorithm 1. The tensor \(\boldsymbol {\mathcal {T}}_{n}\) is a generalization of the so-called (complex) Jacobi matrix associated with the non-Hermitian Lanczos algorithm; see, e.g., [45] and references therein. By using \(\boldsymbol {\mathcal {T}}_{n}\), Theorem 1 provides a compact representation of the three-term recurrences constructing the biorthogonal bases.
Theorem 1
The three-term recurrences (3.2–3.2) can be written in the compact form
where \( \widetilde {\boldsymbol {\mathcal {V}}}_{n} \in \mathbb {C}^{N \times n \times M \times M}\) is
and \(\widetilde {\boldsymbol {\mathcal {W}}}_{n} \in \mathbb {C}^{n \times N \times M \times M}\) is
Proof
By direct inspection. We have, for all \(i_{1} \in \{1,\dots ,N \}\), \(i_{2} \in \{1,\dots ,n-1\}\)
and
The equality between (7) and (8) follows using (3b) and proves (6a). The remaining part of the theorem can be proved analogously. □
If we ∗-multiply (6a) by \(\boldsymbol {\mathcal {W}_{n}}\) from the left we obtain the expression
The tensor \(\boldsymbol {\mathcal {T}}_{n}\) satisfies a generalization of the matching moment property which is stated in Theorem 2.
Theorem 2 (Matching Moment Property)
Let \(\boldsymbol {\mathcal {A}}, V, W\) and \(\boldsymbol {\mathcal {T}}_{n}\) be as described above, then
where E1 = e1 ⊗ IM and e1 is the first vector of the Euclidean base of \(\mathbb {C}^{N \times N}\).
The proof of Theorem 2 can be found in A.
3.4 Numerical properties
The tensor \(\boldsymbol {\mathcal {A}}\) is obtained by discretizing A(t) and stores in \(\boldsymbol {\mathcal {A}}_{k,l,:,:}\) the coefficients representing the (k,l)-th element of A(t). Different methods of discretization are possible. In this paper, following [23], we discretize the interval I obtaining the mesh
For this mesh the discretization of \(\boldsymbol {\mathsf {A}}(t) = \left [\boldsymbol {\mathsf {A}}_{k,\ell }(t) \right ]_{k,\ell }^{N}\) is the tensor
where the matrices \(\boldsymbol {\nu }_{k,\ell }\in \mathbb {C}^{M \times M}\) are lower triangular matrices defined as
This discretization scheme has an accuracy of order \(\mathcal {O}(h) = \mathcal {O}(1/M)\) and, indeed, in Section 5, we show that when this discretization scheme is used, the approximation of the bilinear form of interest also has an accuracy of \(\mathcal {O}(1/M)\).
The computational cost of Algorithm 1 depends on the chosen number of discretization points M and on the number of iterations n. In this algorithm the dominant cost is the multiplication of a 4-mode tensor with a 3-mode tensor, i.e., \(\boldsymbol {\mathcal {A}}\ast V_{k}\) and \({W_{k}^{D}}\ast \boldsymbol {\mathcal {A}}\). The worst case complexity of one such product is \(\mathcal {O}(M^{3} N^{2})\), for a total cost of \(\mathcal {O}(2 n M^{3} N^{2})\). However, since A(t) is sparse in all practical applications, the computational cost can be much lower. For example, if there are Nnz < N nonzeros in each column of A(t), then the cost reduces to \(\mathcal {O}(2 n M^{3} N N_{\text {nz}} )\). It is important to note, moreover, that the term M3 arises from the matrix-matrix multiplication between Vk and the blocks in \(\boldsymbol {\mathcal {A}}\). Since these blocks arise from a discretization strategy, it is likely that they will exhibit a particular structure that can be exploited for efficient computations. E.g., in the discretization used in this work, these blocks are lower triangular matrices for which the matrix-matrix multiplication has a cost of \(\frac {M^{3}}{2}\).
Finally, the storage cost of Algorithm 1 is three basisvectors Vi, three basisvectors \(W^{D}_{i}\) and 3n − 1 nonzero elements of \(\boldsymbol {\mathcal {T}_{n}}\), for a total of \(\mathcal {O}(6M^{2} N+ 3M^{2} n)\). Only three basisvectors must be kept in memory thanks to the underlying three-term recurrence relation.
Let us conclude this section observing that, as highlighted from the previous discussion, both, the computational cost and storage requirement depend strongly on the number M of discretization points used. For the discretization scheme described above, we expect that a large number of discretization points is required since its accuracy is \(\mathcal {O}(1/M)\). This justifies the search for more accurate discretization schemes, for example Legendre polynomial approximation. However, other discretization schemes will not be discussed here since they are subject of future research and since the discretization scheme introduced above suffices to illustrate the potential of Algorithm 1.
4 Breakdowns
If the matrix βk+ 1 is singular, then line 11 in Algorithm 1 cannot be performed and the algorithm breaks down. This breakdown issue is inherited from the (usual) non-Hermitian Lanczos algorithm; see, e.g., [19, 27, 28, 43, 52]. There are two different kinds of breakdowns. The first one, the so-called lucky breakdown, occurs when one of the Krylov-type subspaces \(\mathcal {K}_{k}(\boldsymbol {\mathcal {A}},A)\) or \(\mathcal {K}_{k}^{D}(B^{D},\boldsymbol {\mathcal {A}})\) becomes invariant under ⋆-multiplication with \(\mathcal {A}\) from the left or right, respectively. Suppose that \(\mathcal {K}_{k}(\boldsymbol {\mathcal {A}},A)\) is an invariant subspace, this will result, in exact arithmetic, in \(\widehat {V}_{k+1} = \boldsymbol {0}\) in Line 5 of Algorithm 1. In finite precision \(\widehat {V}_{k+1}\in \mathbb {C}^{N\times M \times M}\) will never be exactly zero. Therefore, the Frobenius norm
is used to define the following criterion to detect a lucky breakdown:
with 𝜖 << 1, a user-defined threshold close to machine precision. The same applies to the case \(\widehat {W}_{k+1}^{D} = \boldsymbol {0}\). The second kind of breakdown occurs when both \(\widehat {V}_{k+1} \neq \boldsymbol {0}\) and \(\widehat {W}_{k+1}^{D} \neq \boldsymbol {0}\), but \(\boldsymbol {\beta }_{k+1}\in \mathbb {C}^{M\times M}\) is still singular, then the algorithm breaks down. This case is known as a serious breakdown. In numerical computation, the condition number of βk+ 1 is monitored to decide if Line 11 can be computed sufficiently accurate. A user-defined threshold 𝜖s >> 1 specifies an upper bound on the allowed condition number of βk+ 1. That is, if the ratio of its largest and smallest singular value is larger than 𝜖s, i.e., \(\sigma _{\max \limits }(\boldsymbol {\beta }_{k+1})/\sigma _{\min \limits }(\boldsymbol {\beta }_{k+1}) > \epsilon _{s}\), then the algorithm breaks down. Note that the choice of γk+ 1 will influence the condition number of βk+ 1.
In the usual non-Hermitian Lanczos algorithm, a serious breakdown can be treated by using a so-called look-ahead strategy; see, e.g., [8, 9, 19, 27, 28, 43, 51]. Connection between serious breakdowns, (formal) orthogonal polynomials, and matching moment property can be found in [16, 44]. If needed, an analogous look-ahead strategy may be implemented for the tensor Lanczos algorithm. At the moment, an easier strategy to deal with serious breakdowns is to reformulate the problem so to change the input vectors v,w. For instance, when w = ei and v = ei, a serious breakdown is likely to happen due to the sparsity of \(\boldsymbol {\mathcal {A}}\). However, we can rewrite the approximation of the time-ordered exponential U(t) as
with \(\boldsymbol {e} = (1, \dots , 1)^{H}\). Then one can approximate \((\boldsymbol {e} + \boldsymbol {e}_{i})^{H} \boldsymbol {\mathsf {U}}(t) \boldsymbol {e}_{j}\) and \(\boldsymbol {e}^{H} \boldsymbol {\mathsf {U}}(t) \boldsymbol {e}_{j}\) separately, which are less likely going to have a breakdown thanks to the fact that e is a full vector; see, e.g., [25, Section 7.3] and [23].
5 Numerical examples
Let us consider the following smooth matrix-valued function defined on a real interval I = [a,b]:
As anticipated in the Introduction, the time-ordered exponential of A(t) is the unique matrix-valued function \(\boldsymbol {\mathsf {U}}(t) \in \mathbb {C}^{N\times N}\) defined on I that is the solution of the system of linear ordinary differential equations
see [17]. In this section, we aim to approximate the bilinear form wHU(t)v by using the tensor non-Hermitian Lanczos algorithm. If the matrix A is so that A(τ1)A(τ2) −A(τ2)A(τ1) = 0 for all τ1,τ2 ∈ I, then \(\boldsymbol {\mathsf {U}}(t)=\exp \left ({{\int \limits }_{s}^{t}} \boldsymbol {\mathsf {A}}(\tau ) \mathrm {d}\tau \right ).\) Unfortunately, U(t) – and the related bilinear forms – cannot be expressed by an analogous simple form in the general case. Indeed, even for small matrices, U(t) may be given by complicated special functions [32, 53].
A new approach for the approximation of a time-ordered exponential bilinear form was introduced in [22,23,24] and it is based on ⋆-Lanczos, which is a symbolic algorithm. This method is able to approximate the bilinear form
for the given vectors w,v, with wH,v≠ 0. The matrices \(\boldsymbol {\alpha }_{1},\dots ,\boldsymbol {\alpha }_{n}\), \(\boldsymbol {\beta }_{2},\dots ,\boldsymbol {\beta }_{n}\), and \(\boldsymbol {\gamma }_{2},\dots ,\boldsymbol {\gamma }_{n}\), which compose the 4-mode tensor \(\boldsymbol {\mathcal {T}}_{n}\) in (5), are obtained by running n iterations of Algorithm 1 with, as inputs, the 4-mode tensor \(\boldsymbol {\mathcal {A}}\) in (10) and the vectors v,w.
Sampling the true solution wHU(t)v on the discretization nodes τi gives the vector \(\hat {\mathbf {s}}\) defined as
Exploiting the results described in [23], the sampled solution vector \(\hat {\mathbf {s}}\) can be approximated by
where R∗ is the ∗-resolvent , i.e., the tensor
and
Overall, the accuracy of the approximation in (11) can not be better than \(\mathcal {O}(h)\). This is due to the fact that, as explained in [23], the discretization (10) is based on a rectangular quadrature rule. Finally, using the Path-sum method [21] we get the following explicit expression for the ∗-resolvent in terms of a continued fraction
with \(\widetilde {\boldsymbol {\alpha }}_{i} = I_{{M}} - \boldsymbol {\alpha }_{i}\). (12) is computed from the most inner term moving outward, where the inversion operation is performed using the backslash operator in Matlab®;. Note that the ∗-resolvent and all inverses appearing in (12) are expected to exist for h small enough, since their continuous counterparts exist under certain regularity conditions on A(t); see [22, 24].
The rest of the section is structured as follows. Section 5.1 describes the measures that will be used to quantify the errors of the final solution and of the computed biorthonormal bases for Krylov subspaces. In Section 5.2 two examples are discussed for which an analytical solution is available. This allows us to compare the approximation to an exact solution and to show that it converges with the expected rate of convergence. Small-scale examples from NMR spectroscopy are discussed in Section 5.3. Finally, in Section 5.4, we analyze the approximability of the previously considered tensors by the Tensor Train representation.
5.1 Error measures
In this section we define a series of error measures which quantify the quality of the generated biorthogonal bases and the accuracy of the approximation (11). These measures use the Frobenius norm, which, for a 4-modes tensor, is defined as
The main goal is to analyze the rate of convergence as the number of discretization points M is increased. To stress the dependence on M of computed quantities we use the superscript “(M)”.
A generalization of the usual error measures for Krylov subspace methods are used. As a measure for the biorthonormality of the bases \(\boldsymbol {\mathcal {V}}_{n} \in \mathbb {C}^{N\times n \times M \times M }\) and \(\boldsymbol {\mathcal {W}}_{n}\in \mathbb {C}^{n\times N \times M \times M }\) generated by n steps of the algorithm, we use
For a robust algorithm, it is paramount that the term \({\max \limits } (\Vert \boldsymbol {\mathcal {V}}_{n}^{(M)} \Vert _{F},\Vert \boldsymbol {\mathcal {W}}_{n}^{(M)} \Vert _{F})\) remains small as n increases. This can be obtained by employing an appropriate strategy to rescale the basisvectors in \(\boldsymbol {\mathcal {V}}_{n}^{(M)}\) and \(\boldsymbol {\mathcal {W}}_{n}^{(M)}\), i.e., by γk+ 1 in Algorithm 1. In this section we will choose γk+ 1 = IM for all k, i.e., no rescaling. An effective rescaling strategy can improve on the numerical results reported below, but developing such a strategy is subject to future research.
To measure the quality of the recurrences (1), we use
As a measure for the Matching Moment Property, see Theorem 2, we use
which should be close to zero for \(k=0,\dots , 2n-1\).
Finally, to quantify the quality of the solution, we consider as error measure for (11) the quantity
where, if no analytic expression is available, an approximation of s is obtained by using ode45 in Matlab®;. In the formula above, ∥⋅∥2 stands for the usual Euclidean norm. The rate at which errsol decreases as M increases is expected to be \(\mathcal {O}(h)= \mathcal {O}(1/M)\), i.e., the accuracy of the discretization used here.
5.2 Proof of concept
As a proof of concept, we test our proposal on two problems which originally appeared in [23]. In both experiments a discretization with M points is used and we run n iterations of Algorithm 1 with γk+ 1 ≡ IM. This produces the tensor \(\boldsymbol {\mathcal {T}}_{n}^{(M)}\) defined in (5) with coefficients \(\boldsymbol {\alpha _{1}}^{(M)},\dots , \boldsymbol {\alpha _{n}}^{(M)}\) and \(\boldsymbol {\beta _{2}}^{(M)},\dots , \boldsymbol {\beta _{n}}^{(M)}\), depending on M. For the two experiments considered here the result of the ⋆-Lanczos algorithm [23] is known. The coefficients resulting from the latter algorithm are bivariate functions \(\alpha _{1}(t,s),\dots ,\alpha _{n}(t,s)\) and \(\beta _{2}(t,s),\dots ,\beta _{n}(t,s)\), because ⋆-Lanczos is a symbolic algorithm. The tensor Lanczos algorithm is a discretization of the ⋆-Lanczos algorithm, which means that \(\boldsymbol {\alpha _{i}}^{(M)}\) and \(\boldsymbol {\beta _{i}}^{(M)}\) can be seen as discretizations of the functions αi(t,s) and βi(t,s), respectively. Consider the evaluation of these functions on the mesh τi:
then, we can define the errors \(\frac {\Vert \boldsymbol {\hat {\alpha }}_{i} - \boldsymbol {\alpha }_{i}^{(M)}\Vert _{2}}{\Vert \boldsymbol {\hat {\alpha }}_{i} \Vert _{2}}\), \(i=1,\dots ,n\), and \(\frac {\Vert \boldsymbol {\hat {\beta }}_{i} - \boldsymbol {\beta }_{i}^{(M)}\Vert _{2}}{\Vert \boldsymbol { \hat {\beta }}_{i} \Vert _{2}}\), \(i=2^,\dots ,n\). These errors will be used as a measure for the accuracy of the computed tensor \(\boldsymbol {\mathcal {T}}_{n}^{(M)}\). The number of iterations n is chosen equal to the problem size N, which allows us to compare all the available functions \(\alpha _{1}(t,s),\dots , \alpha _{N}(t,s)\) and \(\beta _{2}(t,s),\dots , \beta _{N}(t,s)\) with the elements in \(\boldsymbol {\mathcal {T}}_{N}^{(M)}\) and track the convergence rate with M of the latter.
5.2.1 Time-independent matrix
Consider a constant matrix and starting vectors
and the interval I = [0,1]. The inputs of Algorithm 1 are the starting hypervectors v ⊗ IM,w ⊗ IM and the tensor \(\boldsymbol {\mathcal {A}}\) whose components \(\boldsymbol {\mathcal {A}}_{i_{1},i_{2},:,:}\) are defined as
where \(\boldsymbol {0}\in \mathbb {R}^{M\times M}\) is the null matrix. The tensor \(\boldsymbol {\mathcal {A}}\) is obtained by sampling the matrix-valued function A(t) on the M point mesh (9) and following the definition in (10). The output for n = N = 3 iterations is \(\boldsymbol {\mathcal {T}}_{N}^{(M)}\). Table 3 reports the Krylov error measures, which behave as expected for the recurrence measures. The loss of biorthogonality observed for increasing values of M is presumably due to the fact that no rescaling is used in the algorithm.
On the other hand, this loss of biorthogonality does not compromise the moment matching capabilities of \(\boldsymbol {\mathcal {T}}_{N}^{(M)}\), as it becomes evident from Table 4 where we report errM(k) for k ≤ 5 = 2n − 1.
Moreover, as the values reported in Table 5 confirm, we can observe that the elements βi converge at the expected rate of \(\mathcal {O}(1/M)\). The elements αi are very accurate for M = 10 whereas for larger M, the accuracy of αi decreases: this decrease is, presumably, the result of error propagation in the numerical algorithm. This error is still smaller than the expected order of \(\mathcal {O}(1/M)\).
Moreover, in this particular case, the analytical solution to the ODE is known; see [23]:
with \(\left (\exp (At) \right )_{11} = -\frac {1}{2} \sinh (2t) + \frac {1}{2}\cosh (2t) + \frac {1}{2} \cosh (\sqrt {2}t)\). Hence, it is possible to compare this exact solution with (11). Note that for n = 3 the ∗-resolvent is given by the continued fraction:
Table 6 shows the error measure errsol for increasing M, which convergences at the expected rate \(\mathcal {O}(1/M)\).
5.2.2 Time-dependent matrix
Consider the time-dependent matrix
the starting vectors \(\mathbf {v} = \mathbf {w} = \begin {bmatrix} 1& 0& 0& 0 & 0 \end {bmatrix}^{\top }\), and the interval I = [10− 4,1]. As it becomes apparent from the results reported in Table 7, for this particular experiment, we obtain that the Krylov error measures and the recurrence measures are small whereas the biorthogonality measure is large.
The loss of orthogonality is an inherent feature of Lanczos-like algorithms, and it does not necessarily compromise algorithm’s capability to produce an approximation to the bilinear form. Indeed, Table 8 confirms that the matching moment property is not affected by the loss of ∗-biorthogonality.
On the other hand, the results presented in Table 9, where we report the error measures for the coefficients computed by Algorithm 1, show that the loss of ∗-biorthogonality of the computed bases has a limited impact for the convergence of the algorithm to the solution. Indeed, in this case, for all βi the expected convergence rate is observed whereas, for αi, only i = 1,2,3 show the expected decrease in the error measure.
Table 10 shows that the approximation to the quantity of interest converges at the rate \(\mathcal {O}(1/M)\). Hence, the loss of biorthogonality and the inaccurate coefficients of the tridiagonal tensor did not compromise this approximation.
5.3 NMR experiments
Nuclear magnetic resonance (NMR) spectroscopy studies the structure and dynamics of molecules by looking at nuclear spins [31, 39]. Computer simulations of NMR experiments are important because they can improve the design and analysis of laboratory experiments [50]. In this section, three small, realistic examples arising from NMR spectroscopy [3] are discussed. The ODE that governs the dynamics of nuclear spins during NMR spectroscopy is the Schrödinger equation
where H(t) is the so-called Hamiltonian, ϕ(t) the wave function, \(\tau _{\exp }\) the duration of the experiment and \(\imath = \sqrt {-1}\). The size of the Hamiltonian is related to the number of nuclear spins present in the system, for l nuclear spins H(t) is of the size 2l × 2l. Hence, H(t) grows exponentially with the number of spins, but it is a sparse matrix, making it an ideal candidate for a Lanczos-like algorithm.
The experiments discussed in this section use M = 500 discretization points, because of memory constraints. It is important to note that such memory constraints may be overcome using the Tensor Train approximation presented in Section 5.4. The number of iterations of the tensor Lanczos algorithm is chosen to obtain the maximal attainable accuracy, which is determined by the discretization scheme with M = 500. Since the discretization used here has an accuracy of \(\mathcal {O}(1/M)\), the smallest number of iterations n such that errsol is of order 10− 3 suffices. Choosing larger n will not decrease errsol further for a fixed M and will increase the computational cost.
5.3.1 Experiment 1: Weak coupling
Consider four nuclear spins with heteronuclear dipolar couplings. In this framework, the Hamiltonian for a magic angle spinning (MAS) experiment [29] is the diagonal matrix
with \(\alpha _{k},\beta _{k},\gamma _{k}\in \mathbb {R}\) and ν = 104. The diagonal matrix A(t) = −ı2πH(t) commutes with itself at all times and thus the solution U(t) can be computed
The starting vectors are chosen to excite and measure the lowest oscillatory components in U(t): \(\mathbf {w} =\mathbf {v} = \begin {bmatrix} 0 & 1 & 1 & 0 & 1 & 1& {\dots } & 0& 1& 1 \end {bmatrix}^{\top }\). A typical experiment would run for a time of the order 102 seconds. Since the problem is (highly) oscillatory and the current discretization requires many points to accurately compute a solution, we choose to restrict the experiment time to \(\tau _{\exp }=5\times 10^{-5}\). This is a valid approach since the total time interval of 10− 2 can be split into subintervals of length 5 × 10− 5 and the solutions on the subintervals can be combined to obtain the solution on the whole interval.
Algorithm 1 is run for n = 3 iterations and the corresponding Krylov error measures are shown in Table 11. A first observation concerns the fact that going from M = 5 to M = 50 a large decrease of the biorthogonality measure is observed. This is due to the fact that when discretizing with fewer discretization points, e.g., M = 5, the original matrix in the ODE − ı2πH(t) is translated into a simpler (and inaccurate) discretized input, for which the tensor Lanczos iteration converges fast. The discretization M = 50 represents the original input better, as is suggested by the stagnation of erro going from M = 50 to M = 500.
The error errsol is computed using the analytical solution wHU(t)v evaluated in the discretization points and decays at the expected rate \(\mathcal {O}(1/M)\). Figure 1 shows \(\hat {\mathbf {s}}\) and the approximation \(\mathbf {s}^{(M)}_{n}\) as a function of time; such approximation clearly converges for increasing M (x-axis reports time).
5.3.2 Experiment 2: Strong coupling
MAS with four nuclear spins with homonuclear dipolar couplings leads to the Hamiltonian
where \(\alpha _{k}\in \mathbb {R}\) is a scalar, and \(B,C\in \mathbb {R}^{16\times 16}\) are matrices with a sparsity structure as shown in Fig. 2.
A typical experiment time is 10− 2 seconds and ν = 104. The simulated experiment time is \(\tau _{\exp } = 5 \times 10^{-6}\), the size of the Krylov subspace is k = 4 and \(\mathbf {v}=\mathbf {w}=\begin {bmatrix} 0 & 1 & 1 & 0 & 1 & 1& {\dots } & 0& 1& 1 \end {bmatrix}^{\top }\). The corresponding error measures are shown in Table 12, which show a similar behavior as for Experiment 1. The measure errsol is computed by comparing \(\hat {\mathbf {s}}\) to the solution obtained by ode45.
5.3.3 Experiment 3: Uncoupled spins under a pulse wave
The Hamiltonian for four uncoupled spins under a pulse wave is
with \(\alpha _{k}\in \mathbb {R}\) and \(B\in \mathbb {R}^{16\times 16}\), \(C\in \mathbb {C}^{16 \times 16}\) have a structure as shown in Fig. 3.
A practical experiment time ranges from 10− 6 to 10− 3 seconds, here \(\tau _{\exp } = 10^{-3}\) is used. The starting vectors are \(\mathbf {v}= \mathbf {w}= \begin {bmatrix} 1 & {\dots } 1 \end {bmatrix}^{\top }\) and n = 4 iterations of the tensor Lanczos algorithm are run. The Krylov error measures shown in Table 13 behave similarly to the measures observed for Experiments 1 and 2. The measure errsol is obtained via ode45 and shows a convergence rate a bit slower than \(\mathcal {O}(1/M)\). The slower convergence rate can, in part, be explained by the fact that a comparison is made with the ode45 solution. Additional errors are incurred when comparing s with \(\hat {\mathbf {s}}\), because the former is only available in the points τi and the latter is available only in points which are determined by ode45.
5.4 Tensor Train approximations
As briefly mentioned in the Introduction, despite the fact that the block matrix and the tensor formulation of the problem (1) are mathematically equivalent, the tensor formulation introduced and analyzed in this work, allows the exploitation of particular low parametric representations. The aim of this section is indeed to show that for all the examples previously presented, the resulting tensors can be accurately and conveniently approximated using a low parametric representation called Tensor Train (TT) format [41, 42].
As a matter of fact, multilinear algebra, tensor analysis, and the theory of tensor approximations play increasingly important roles in nowadays computational mathematics and numerical analysis, thereby attracting tremendous interest in recent years [12]. In this panorama, Tensor Train (TT) approximations are a powerful technique for dealing with the curse of dimensionality, i.e., the particularly unpleasant feature where the number of unknowns and the computational complexity grow exponentially when the dimension of the problem increases.
Before presenting the computational results, we briefly survey the main features of the TT representation, addressing the interested reader to the surveys [11, 12]. We consider the Tensor Train (TT) format [41] for the tensors of interest in this work. Specifically, a 4-mode \(\boldsymbol {\mathcal {A}} \in \mathbb {C}^{N_{1} \times N_{2} \times M \times M }\) tensor is expressed in TT format when
where Gk(ik) is a matrix of dimension rk− 1 × rk and r0 = r4 = 1. The numbers rk are called TT-ranks, and Gk(ik) are the cores of the TT decomposition. If rk ≤ r, nk ≤ n, then storing the TT representation requires memorizing ≤ 4nr2 numbers. If r is small, then the memory requirement is much smaller than storing the full tensor, i.e., storing n4 numbers.
It is important to note that the TT representation allows to approximate various tensor operations efficiently, see, e.g., [41, Sec. 4]. In this paper, we do not propose a low parametric TT version of Algorithm 1. To be efficient, such a TT version would need a TT representation of the tensor products used in Algorithm 1. This paper aims to show that Algorithm 1 works; further enhancements are postponed to future investigations.
In Tables 14, 15, and 16 we present the TT ranks for all the tensors considered in Section 5.3. In particular, the tables present the details of the TT approximations obtained using the TT-toolbox [41] when the required accuracy for the approximation is set to 10− 5 and 10− 10. As becomes evident from the presented results, all the considered tensors are amenable of a low parametric representation provided by the TT format and, indeed, for all the presented results the Compression Factor (C.F.), which is defined as \(({\sum }_{k=1}^{4} r_{k-1} \times n_{k} \times r_{k})/nnz(\boldsymbol {\mathcal {A}})\), with \(nnz(\boldsymbol {\mathcal {A}})\) the number of nonzero elements of \(\boldsymbol {\mathcal {A}}\), lies in the interval (10− 3,0.5). It is important to observe that when increasing the accuracy from 10− 5 to 10− 10 the C.F. does not significantly change, suggesting the interesting fact that for the considered tensors, the TT format is closer to an exact representation rather than an approximation. Finally, it is important to note that the ranks of the TT approximations are robust across the choices of the parameter ν (cfr. the TT ranks in Table 14 and in Table 15) and to note also that, for some of the considered problems, the number of parameters needed for the approximation can be two orders of magnitude smaller than \(nnz(\boldsymbol {\mathcal {A}})\); see Tables 15 and 16.
6 Conclusions
In this work we introduced a non-Hermitian Lanczos algorithm for tensors and we provided the corresponding theoretical analysis. In particular, after introducing all the necessary theoretical background, we are able to interpret such Lanczos-type process in terms of tensor polynomials and to prove the related matching moment property. A series of numerical experiments performed on real-world problems confirm the effectiveness of our approach. Using a linearly converging approximation for the inputs, the algorithm produces a linearly converging approximation of the bilinear form wHU(t)v, where U(t) is the solution of the ODE (1). More accurate approximation schemes for the inputs are currently being developed by some of the paper’s authors, possibly leading to faster convergence. Moreover, in all the considered examples, the related tensors show a low parametric structure in terms of Tensor Train representation. This important feature paves the path for future efficiency improvements of our proposal where this representation is fully exploited in the Lanczos-type procedure.
References
Autler, S.H., Townes, C.H.: Stark effect in rapidly varying fields. Phys. Rev. 100, 703–722 (1955)
Bader, P., Iserles, A., Kropielnicka, K., Singh, P.: Efficient methods for linear Schrödinger equation in the semiclassical regime with time-dependent potential. Proc R Soc A Math Phys Eng Sci 472, 20150733 (2016)
Baligács, E., Bonhomme, C: https://github.com/BaligacsEni/TOMEexamples.git. Accessed 1st July 2022 (2022)
Benner, P., Cohen, A., Ohlberger, M., Willcox, K.: Model Reduction and Approximation: Theory and Algorithms. Computational Science and Engineering. SIAM, Philadelphia (2017)
Blanes, S.: High order structure preserving explicit methods for solving linear-quadratic optimal control problems. Numer. Algorithms 69(2), 271–290 (2015)
Blanes, S., Casas, F.: A concise introduction to geometric numerical integration. CRC Press, Bocan Raton (2017)
Blanes, S., Casas, F., Oteo, J., Ros, J.: The Magnus expansion and some of its applications. Phys. Rep. 470(5), 151–238 (2009)
Brezinski, C., Redivo-Zaglia, M., Sadok, H.: Avoiding breakdown and near-breakdown in Lanczos type algorithms. Numer. Algorithms 1(3), 261–284 (1991)
Brezinski, C., Redivo Zaglia, M., Sadok, H.: A breakdown-free Lanczos type algorithm for solving linear systems. Numer. Math. 63(1), 29–38 (1992)
Budd, C., Iserles, A., Nørsett, S.: On the solution of linear differential equations in Lie groups. Philos. Trans. R. Soc. London. Series A: Mathematical Phys. Eng. Sci. 357(1754), 983–1019 (1999)
Cichocki, A., Lee, N., Oseledets, I., Phan, A.-H., Zhao, Q., Mandic, D.: Low-rank tensor networks for dimensionality reduction and large-scale optimization problems: Perspectives and challenges part 1. arXiv:1609.00893(2016)
Cichocki, A., Mandic, D., Caiafa, C., Phan, A., Zhou, G., Zhao, Q., De Lathauwer, L.: Tensor decompositions for signal processing applications. IEEE Signal Processing Mag. (2013)
Cohen, D., Jahnke, T., Lorenz, K., Lubich, C.: Numerical integrators for highly oscillatory Hamiltonian systems: A review. In: Mielke, A. (ed.) Analysis, Modeling and Simulation of Multiscale Problems, pp 553–576. Springer, Berlin (2006)
Corless, M., Frazho, A.: Linear Systems and Control: An Operator Perspective. Pure and Applied Mathematics. Marcel Dekker, New York (2003)
Degani, I., Schiff, J.: RCMS: Right correction Magnus series approach for oscillatory ODEs. J. Comput. Appl. Math. 193(2), 413–436 (2006)
Draux, A.: Formal orthogonal polynomials revisited. Applic. Numer. Algorithms 11(1), 143–158 (1996)
Dyson, F.J.: Divergence of perturbation theory in quantum electrodynamics. Phys. Rev. 85(4), 631–632 (1952)
Feng, J., Yang, L.T., Zhang, R., Qiang, W., Chen, J.: Privacy preserving high-order Bi-Lanczos in cloud-fog computing for industrial applications. IEEE Transactions on Industrial Informatics, 1–1 (2020)
Freund, R.W., Gutknecht, M.H., Nachtigal, N.M.: An implementation of the look-ahead Lanczos algorithm for non-Hermitian matrices. SIAM J. Sci Comput. 14(1), 137–158 (1993)
Frommer, A., Lund, K., Szyld, D.B.: Block Krylov subspace methods for functions of matrices. Electron. Trans. Numer Anal. 47, 100–126 (2017)
Giscard, P.-L., Lui, K., Thwaite, S.J., Jaksch, D.: An exact formulation of the time-ordered exponential using path-sums. J. Math. Phys. 56(5), 053503 (2015)
Giscard, P.-L., Pozza, S.: Lanczos-like algorithm for the time-ordered exponential: The ∗-inverse problem. Appl. Math. 65(6), 807–827 (2020)
Giscard, P.-L., Pozza, S.: A Lanczos-like method for non-autonomous linear ordinary differential equations. arXiv:1909.03437 (2021)
Giscard, P.-L., Pozza, S.: Tridiagonalization of systems of coupled linear differential equations with variable coefficients by a Lanczos-like method. Linear Algebra Appl. 624, 153–173 (2021)
Golub, G.H., Meurant, G.: Matrices, Moments and Quadrature with Applications. Princeton Ser. Appl. Math. Princeton University Press, Princeton (2010)
Guide, M.E., Ichi, A.E., Jbilou, K., Sadaka, R.: On tensor GMRES and Golub-Kahan methods via the T-product for color image processing. Electron. J. Linear Algebra 37, 524–543 (2021)
Gutknecht, M.H.: A completed theory of the unsymmetric Lanczos process and related algorithms. I. SIAM J. Matrix Anal. Appl. 13(2), 594–639 (1992)
Gutknecht, M.H.: A completed theory of the unsymmetric Lanczos process and related algorithms. II. SIAM J. Matrix Anal. Appl. 15(1), 15–58 (1994)
Hafner, S., Spiess, H.-W.: Advanced solid-state NMR spectroscopy of strongly dipolar coupled spins under fast magic angle spinning. Concepts Magnetic Resonance 10(2), 99–128 (1998)
Hochbruck, M., Lubich, C.: Exponential integrators for quantum-classical molecular dynamics. BIT Numer. Math. 39(4), 620–645 (1999)
Hore, P.J.: NMR principles. In: Lindon, J.C. (ed.) Encyclopedia of Spectroscopy and Spectrometry, 2nd edn, pp 1833–1840. Academic Press, Oxford (1999)
Hortaçsu, M : Heun functions some of their applications in physics. Adv High Energy Phys 2018, 8621573 (2018)
Huang, B., Xie, Y., Ma, C.: Krylov subspace methods to solve a class of tensor equations via the Einstein product. Numer. Linear Algebra Appl. 26, 4 (2019)
Iserles, A.: On the global error of discretization methods for highly-oscillatory ordinary differential equations. BIT Numer. Math. 42(3), 561–599 (2002)
Iserles, A.: On the method of Neumann series for highly oscillatory equations. BIT Numer. Math. 44(3), 473–488 (2004)
Iserles, A., Munthe-Kaas, H.Z., Nørsett, S.P., Zanna, A.: Lie-group methods. Acta Numer. 9, 215–365 (2000)
Kwakernaak, H., Sivan, R.: Linear Optimal Control Systems, vol. 1. Wiley-interscience, New York (1972)
Lauder, M., Knight, P., Greenland, P.: Pulse-shape effects in intense-field laser excitation of atoms. Opt Acta 33(10), 1231–1252 (1986)
Levitt, M.H.: Spin Dynamics: Basics of Nuclear Magnetic Resonance, 2nd edn. Wiley, Chichester (2008)
Magnus, W.: On the exponential solution of differential equations for a linear operator. Comm. Pure Appl. Math. 7(4), 649–673 (1954)
Oseledets, I.: Tensor-train decomposition. SIAM J. Sci. Comput. 33(5), 2295–2317 (2011)
Oseledets, I., Tyrtyshnikov, E.: TT-cross approximation for multidimensional arrays. Linear Algebra Appl. 432(1), 70–88 (2010)
Parlett, B.N., Taylor, D.R., Liu, Z.A.: A look-ahead Lanczos algorithm for unsymmetric matrices. Math Comp. 44(169), 105–124 (1985)
Pozza, S., Pranić, M.: The Gauss quadrature for general linear functionals, Lanczos algorithm, and minimal partial realization. Numer. Algorithms 88, 647–678 (2021)
Pozza, S., Pranić, M. S., Strakoš , Z.: Gauss quadrature for quasi-definite linear functionals. IMA J. Numer Anal. 37(3), 1468–1495 (2017)
Reichel, L., Ugwu, U.O.: Tensor Arnoldi-Tikhonov and GMRES-type methods for ill-posed problems with a t-product structure. Journal of Scientific Computing, 90(1) dec (2021)
Reid, W.T.: Riccati matrix differential equations and non-oscillation criteria for associated linear differential systems. Pacific J. Math. 13(2), 665–685 (1963)
Ruymbeek, K., Meerbergen, K., Michiels, W.: Tensor-Krylov method for computing eigenvalues of parameter-dependent matrices. J. Comput. Appl Math. 408, 113869 (2022)
Shirley, J.H.: Solution of the Schrödinger equation with a Hamiltonian periodic in time. Phys. Rev. 138, B979–B987 (1965)
Smith, S.A., Palke, W.E., Gerig, J.T.: The Hamiltonians of NMR. Part I. Concepts in Magnetic Resonance 4(2), 107–144 (1992)
Taylor, D.R.: Analysis of the Look Ahead Lanczos Algorithm. PhD thesis. University of, California, Berkeley (1982)
Wilkinson, J.H.: The Algebraic Eigenvalue Problem. Monographs on Numerical Analysis. The Clarendon Press Oxford University Press, New York (1988)
Xie, Q., Hai, W.: Analytical results for a monochromatically driven two-level system. Phys. Rev. A 82, 032117 (2010)
Acknowledgements
The authors want to thank Enikö Baligács and Christian Bonhomme (Laboratoire de chimie de la matière condensée de Paris, Sorbonne University) for providing the data from real-world applications used in Section 5. This work was supported by Charles University Research programs No. PRIMUS/21/SCI/009 and UNCE/SCI/023, and by the Magica project ANR-20-CE29-0007 funded by the French National Research Agency. The author M.R.-Z. is a member of the GNCS-INdAM group.
Author information
Authors and Affiliations
Corresponding authors
Ethics declarations
Conflict of interest
The authors declare no competing interests.
Additional information
Data availability
Data sharing not applicable to this article as no datasets were generated or analyzed during the current study.
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendix
Appendix
1.1 A.1 The ∗-product associativeness
In the following, we prove the three statements of Lemma 1. First, we have
The second statement is proved by direct inspection as follows:
Finally, we have the following equality
This concludes the proof of Lemma 1.
1.2 A.2 Matching moment property
We prove Theorem 2 by giving a polynomial interpretation of the Lanczos process for tensors. The proof follows the same principles as the proof given in [23] for a different but analogous case. For simplicity, we consider the case in which the matrices γk are set equal to the matrix identity. The proof can be easily extended to the general case. Let us define the following set of polynomials
with \(\boldsymbol {\eta }_{{k}} \in \mathbb {C}^{M \times M}\). We say that the map \([\cdot , \cdot ]: \mathcal {P}_{*} \times \mathcal {P}_{*} \rightarrow \mathbb {C}^{M\times M}\) is a sesquilinear block form if and only if, given \(p,q,r,s \in \mathcal {P}_{\ast }\) and \(\boldsymbol {\alpha }, \boldsymbol {\beta } \in \mathbb {C}^{M \times M}\), it satisfies
In addition, from now on we assume that for every sesquilinear block form [⋅,⋅] it holds that
Then, [⋅,⋅] is determined by its moments defined as
We say that the sequences \(p_{0}, p_{1}, \dots \) and \(q_{0}, q_{1}, \dots \) from \(\mathcal {P}_{*}\) are sequences of biorthonormal polynomials with respect to [⋅,⋅] if and only if
with δij the Kronecker delta (hereafter the subindex k in pk and qk will stand for the degree of the polynomial). From now on, we also assume m0 = IM, getting p0 = q0 = IM. Let q1 be the polynomial from \(\mathcal {P}_{*}\)
the orthogonality conditions (14) imply
Analogously, we get
with
Repeating such a biorthogonalization process, we obtain the three-term recurrences for \(k=0,1,\dots \)
with p− 1 = q− 1 = 0 and
We remark that the recurrences are obtained using property (13). The previous derivation also constructively proves that the biorthonormal polynomials \(p_{0}, \dots , p_{n}\) and \(q_{0}, \dots , q_{n}\) exist if \(\boldsymbol {\beta }_{1}, \dots , \boldsymbol {\beta }_{n}\) are invertible matrices.
Let \(\boldsymbol {\mathcal {A}}, V, W\) be as in Theorem 2 and let us define the sesquilinear block form
Assume that there exist polynomials \(p_{0}, \dots , p_{n}\) and \(q_{0},\dots ,q_{n}\) from \(\mathcal {P}_{*}\) which are biorthonormal with respect to \([\cdot , \cdot ]_{\boldsymbol {\mathcal {A}}}\). Defining the vectors
and using the recurrences (A) we get the recurrences (3.2) of the non-Hermitian Lanczos for tensors. Moreover, the coefficients in (16) are the coefficients in (4).
Let \(\boldsymbol {\mathcal {T}}_{n}\) be as in Theorem 2. We can define the sesquilinear block form \([q, p]_{n}: \mathcal {P}_{*} \times \mathcal {P}_{*} \rightarrow \mathbb {C}^{M \times M}\) as
Note that here the vector e1 in the definition of E1 = e1 ⊗ IM has length n ≤ N. The following Lemmas will show that
concluding the proof of Theorem 2.
Lemma 2
Let \(p_{0},\dots ,p_{n} \in \mathcal {P}_{*}\) and \(q_{0},\dots ,q_{n} \in \mathcal {P}_{*}\) be biorthonormal polynomials with respect to \([\cdot ,\cdot ]_{\boldsymbol {\mathcal {A}}}\). Assume that \(\boldsymbol {\beta }_{1},\dots , \boldsymbol {\beta }_{n}\) in (A) are invertible matrices. Then the polynomials are also biorthonormal with respect to [⋅,⋅]n as defined above.
Proof
Let us define the tensors Ei := ei ⊗ IM for \(i=1,\dots , n\), with \(I_{n} = [\mathsf {e}_{1}, \dots , \mathsf {e}_{n}]\). We will first prove by induction that for \(i=0,\dots ,n-1\)
For i = 0, the relations (17) are trivial. Assume now that (17) hold for \(i=1,\dots ,k\), by (A) we get
Since βk+ 1 is invertible, direct computations prove that (17) holds i = k + 1.
As a consequence we have
which concludes the proof. □
Lemma 3
Let \(p_{0},\dots ,p_{n-1} \in \mathcal {P}_{*}\) and \(q_{0},\dots ,q_{n-1} \in \mathcal {P}_{*}\) be biorthonormal polynomials with respect to a sesquilinear block form \([\cdot ,\cdot ]_{\boldsymbol {\mathcal {A}}}\) and to a sesquilinear block form \([\cdot ,\cdot ]_{\boldsymbol {{\mathscr{B}}}}\). If \([1,1]_{\boldsymbol {\mathcal {A}}} = [1,1]_{\boldsymbol {{\mathscr{B}}}} = I_{M}\), then \([\lambda ^{{k_{*}}},1]_{\boldsymbol {\mathcal {A}}} = [\lambda ^{{k}_{*}},1]_{\boldsymbol {{\mathscr{B}}}}\) for \({k}=0,\dots ,2n-1\).
Proof
The proof is by induction. Let \(\boldsymbol {\mu }_{{k}} = [\lambda ^{{{k}_{*}}},1]_{\boldsymbol {\mathcal {A}}}\) and \(\widehat {\boldsymbol {\mu }}_{j} = [\lambda ^{{{k}_{*}}},1]_{\boldsymbol {{\mathscr{B}}}} \) for \({k}=0,1,\dots , 2n-1\). The coefficient formula (16) gives
Hence \(\boldsymbol {\mu }_{1} = \boldsymbol {\alpha }_{0} = \widehat {\boldsymbol {\mu }}_{1}\). Considering the induction assumptions \(\boldsymbol {\mu }_{{k}} = \widehat {\boldsymbol {\mu }}_{{k}}\) for \({k}=0,\dots ,2{j}-3\), we prove that \(\boldsymbol {\mu }_{2{j}-2} = \widehat {\boldsymbol {\mu }}_{2{j}-2}\) and \(\boldsymbol {\mu }_{2{j}-1} = \widehat {\boldsymbol {\mu }}_{2{j}-1}\), for \({j} = 2,\dots ,n\). By the formula in (16) we get
which we can rewrite as
where \(\boldsymbol {\eta }_{i}, \widehat {\boldsymbol {\eta }}_{{k}} \in \mathbb {C}^{M \times M}\) are the coefficients respectively of qj− 1 and pj− 2. By the induction assumption we obtain
The leading coefficients of the polynomials q2j− 2 and p2j− 2 are respectively ηj− 1 = 1 and \(\widehat {\boldsymbol {\eta }}_{{j}-2}=(\boldsymbol {\beta }_{{j}-2} \times {\cdots } \times \boldsymbol {\beta }_{1})^{-1}\). Hence \(\boldsymbol {\mu }_{2{j}-2} = \widehat {\boldsymbol {\mu }}_{2{j}-2}\). We conclude the proof repeating the same argument with the coefficient αj− 1 (16). □
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Cipolla, S., Pozza, S., Redivo-Zaglia, M. et al. A Lanczos-type procedure for tensors. Numer Algor 92, 377–406 (2023). https://doi.org/10.1007/s11075-022-01351-6
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11075-022-01351-6