Abstract
This paper is devoted to the numerical solution of constrained energy minimization problems arising in computational physics and chemistry such as the Gross–Pitaevskii and Kohn–Sham models. In particular, we introduce Riemannian Newton methods on the infinite-dimensional Stiefel and Grassmann manifolds. We study the geometry of these two manifolds, its impact on the Newton algorithms, and present expressions of the Riemannian Hessians in the infinite-dimensional setting, which are suitable for variational spatial discretizations. A series of numerical experiments illustrates the performance of the methods and demonstrates their supremacy compared to other well-established schemes such as the self-consistent field iteration and gradient descent schemes.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
The Kohn–Sham model [31, 37, 38] is a prototypical example of a constrained energy minimization problem stated on the infinite-dimensional Stiefel manifold. This means that the sought-after minimizer is a p-frame of \(L^2\)-orthonormal functions. Another well-known example is the Gross–Pitaevskii model for Bose–Einstein condensates of ultracold bosonic gases [39, 43]. Here, the special case \(p=1\) is of interest, where we seek a single (minimizing) function on the unit sphere in \(L^2\), representing a unit mass constraint. Since these two applications are relevant for different communities, numerical methods are mostly considered separately. One aim of this paper is to give a unified approach to solving energy minimization problems. More precisely, we introduce Riemannian Newton methods for minimizing energy functionals of Kohn–Sham type, which also includes the Gross–Pitaevskii model. In general, the here considered PDE problems require a special treatment in terms of sparsity and dimension-independent methods, which is not part of existing general purpose optimization packages.
The numerical solution of the Gross–Pitaevskii model has been studied extensively in recent years. The most common numerical techniques are iterative methods based on Riemannian (conjugate) gradient descent methods or discretized Riemannian gradient flows in various metrics [10, 11, 20, 22, 26, 29, 36, 44, 53]. A conceptually different approach is the J-method [5, 33] with its inimitable sensitivity with regard to spectral shifts, allowing remarkable speed-ups in a Rayleigh quotient iteration manner. Reformulating the minimization problem as an eigenvalue problem with eigenvector nonlinearity—also known as nonlinear eigenvector problem—the self-consistent field iteration (SCF) can be employed; see [14, 23, 45]. This method involves the solution of a linear eigenvalue problem in each step and is strongly connected to the Newton method [28, 34]. Considering the extended nonlinear system including the normalization constraint also allows a direct application of Newton or Newton-type methods [12, 13, 24]. For an extended review on numerical methods for the Gross–Pitaevskii model, we refer to [28].
Most of the above approaches (with appropriate adjustments) have been applied to the Kohn–Sham model as well. This includes the direct constrained minimization algorithm [3, 47, 52] and the energy-adaptive gradient descent method [8]—both based on Riemannian optimization—as well as gradient flow schemes [21, 32]. Moreover, the SCF algorithm with different types of mixing is very popular in the computational chemistry community; see, e.g., [9, 15, 18, 19, 40]. For a discretized and simplified Kohn–Sham model (without the external potential and the exchange-correlation energy), global convergence and local second-order convergence of an inexact Riemannian Newton method on the Grassmann manifold has been shown in [54]. An overview of existing software packages for density functional theory problems can be found in [35].
In this paper, the point of origin is an energy functional defined on the infinite-dimensional Stiefel manifold, which we introduce in Sect. 2. For a better understanding, we recall definitions and properties of the Stiefel manifold and corresponding retractions in Sect. 3. Moreover, we provide formulae for the Riemannian gradient and the Riemannian Hessian which are needed for the Newton iteration. Since the considered energy functional is invariant under orthogonal matrices, we also discuss the infinite-dimensional Grassmann manifold and examine a connection of its tangent space to a certain subspace of the tangent space of the Stiefel manifold. The resulting Newton algorithms are then subject of Sect. 4. In particular, we present an inexact Riemannian Newton method on the Grassmann manifold. In Sect. 5, we consider the two mentioned examples of the Gross–Pitaevskii and the Kohn–Sham model in more detail. For both applications, we derive the formulae including a spatial discretization and illustrate the supremacy of the inexact Newton approach compared to well-established methods such as the SCF iteration and gradient descent schemes.
1.1 Notation
The sets of \(p\times p\) real symmetric and skew-symmetric matrices are denoted by \({\mathcal {S}}_{\textrm{sym}}(p)\) and \({\mathcal {S}}_{\textrm{skew}}(p)\), respectively. For \(M\in {\mathbb {R}}^{p\times p}\), we write \({{\,\textrm{sym}\,}}M=\frac{1}{2}(M+M^T)\) for the symmetric part, and \({{\,\textrm{tr}\,}}M\) denotes the trace of M. Further, \(I_p\) and \(0_p\) denote the \(p\times p\) identity and zero matrices, respectively. The expression \({{\,\textrm{diag}\,}}(M)\) defines the column vector consisting of the diagonal elements of \(M\in {\mathbb {R}}^{n\times n}\) and \({{\,\textrm{Diag}\,}}(v)\) denotes the diagonal matrix with components of the vector \(v\in {\mathbb {R}}^n\) on the diagonal.
2 The Energy Functional and Nonlinear Eigenvector Problems
For a given spatial domain \(\Omega \subseteq {\mathbb {R}}^d\), \(d\le 3\), we consider the Hilbert spaces \(L^2(\Omega )\) and \({\tilde{V}}\!\subseteq \! H^1(\Omega )\). For \(p\ge 1\), we further define the Hilbert spaces \(V={\tilde{V}}^p\) and \(H=[L^2(\Omega )]^p\) of p-frames. Throughout this paper, we assume that V is dense in H and that \(V\subseteq H\subseteq V^*\) form a Gelfand triple, where \(V^*\) denotes the dual space of V.
For \(\varvec{v}= (v_1, \dots , v_p), \varvec{w}= (w_1, \dots , w_p) \in H\), we define the dot product
On the pivot space H, we further introduce an outer product
and an inner product
The inner product (2) induces the norm \(\Vert \varvec{v}\Vert _H=\sqrt{(\varvec{v},\varvec{v})_H}\) on H. The canonical identification \({\mathcal {I}}:V\rightarrow V^*\) is defined by
where \(\langle \,\cdot ,\cdot \,\rangle \) denotes the duality pairing on \(V^*\times V\). This identification operator can also be written as \({\mathcal {I}}= {\mathfrak {j}}^*\circ {\mathfrak {i}}_H\circ {\mathfrak {j}}\) with the trivial embedding \({\mathfrak {j}}:V\rightarrow H\) (the injective identity operator), the Riesz isomorphism \({\mathfrak {i}}_H:H\rightarrow H^*\), which reads \({\mathfrak {i}}_H(u) = (u,\,\cdot \,)_H\), and the adjoint operator \({\mathfrak {j}}^*:H^*\rightarrow V^*\) satisfying \({\mathfrak {j}}^*(f)=f\circ {\mathfrak {j}}\) for all \(f\in H^*\). Since all these operators act componentwisely, we have \({\mathcal {I}}(\varvec{v}\Lambda ) = {\mathcal {I}}(\varvec{v})\Lambda \) for all \(\varvec{v}\in V\) and \(\Lambda \in {\mathbb {R}}^{p\times p}\). Moreover, since V is a dense subspace of H, so is \({\mathfrak {j}}(V)\). Hence, \({\mathfrak {j}}^*\) is injective, and as the composition of injective operators, \({\mathcal {I}}\) is also injective. As a result, \({\mathcal {I}}\) has a left inverse \({\mathcal {J}}:V^*\rightarrow V\) such that \({\mathcal {J}}{\mathcal {I}}\varvec{v}=\varvec{v}\) for all \(\varvec{v}\in V\).
2.1 Energy and Applications
For a p-frame \({\varvec{\phi }}\in V\), we consider the energy functional
with an external potential \(\vartheta \), the density function \(\rho ({\varvec{\phi }}) = {\varvec{\phi }}\cdot {\varvec{\phi }}\), and a smooth nonlinearity \(\varGamma (\rho )\). Our aim is to minimize this energy functional on the infinite-dimensional Stiefel manifold of index p given by
In other words, we are interested in solving the constrained minimization problem
A state of lowest energy is called the ground state. Such states play an important role in quantum-mechanical models as they represent a most stable configuration of atoms and molecules. These models include two famous applications in computational physics and chemistry.
Example 1
(Gross–Pitaevskii model) For \(p=1\) and \(\varGamma (\rho ) = \frac{1}{2} \kappa \rho ^2\) with \(\kappa \in {\mathbb {R}}\), the energy functional takes the form
This is the well-known Gross–Pitaevskii energy used in the modeling of Bose–Einstein condensates of ultracold bosonic gases [39, 43]. Here, \(\vartheta \in L^\infty (\Omega )\) is the magnetic trapping potential, \(\phi \in H_0^1(\Omega )\) is the quantum state of the Bose–Einstein condensate, and \(\kappa \) characterizes the strength and the direction of particle interactions.
Example 2
(Kohn–Sham model) The (non-local) nonlinearity
yields the Kohn–Sham energy functional
where \({\varvec{\phi }}\) denotes a wave function with p components called single-particle orbitals and \(\rho ({\varvec{\phi }})\) is the electronic charge density. Moreover, \(\vartheta _\text {ion}\) is the ionic potential, and \(\epsilon _\text {xc}(\rho )\) is the exchange-correlation energy per particle in a homogeneous electron gas of density \(\rho \). This model is based on the so-called density functional theory [31], which allows a significant reduction of the degrees of freedom [17, 37, 38]. The last integral in (7) is a local density approximation to the exchange-correlation energy obtained by using semi-empirically knowledge of the model [42]. In the Kohn–Sham model, a ground state corresponds to the low-energy wave function of the considered molecule and the orthogonality condition \(\llbracket {{\varvec{\phi }}},{{\varvec{\phi }}}\rrbracket _H = I_p\) means that there is no interaction between the electrons in different orbitals.
At this point, it should be emphasized that, since the energy functional \({\mathcal {E}}\) in (3) is invariant under orthogonal transformations, i.e. \({\mathcal {E}}({\varvec{\phi }})={\mathcal {E}}({\varvec{\phi }}Q)\) for all orthogonal matrices \(Q\in {\mathbb {R}}^{p\times p}\), the optimal solution to the minimization problem (5) is not unique. To overcome this difficulty, we will transfer this problem to the infinite-dimensional Grassmann manifold defined in Sect. 3.2.
2.2 Connection to Nonlinear Eigenvector Problems
We observe that the directional derivative of \({\mathcal {E}}\) from (3) at \({\varvec{\phi }}\in V\) along \(\varvec{w}\in V\) has the form
where
with \(\gamma (\rho )=\frac{{\textrm{d}}}{{{\textrm{d}}}\rho }\varGamma (\rho )\). One can see that for fixed \({\varvec{\phi }}\in V\), \(a_{\varvec{\phi }}\) is a symmetric bilinear form on \(V\times V\). Further note that \(a_{\varvec{\phi }}\) exhibits a special structure, namely
with a symmetric bilinear form \({{\tilde{a}}}_{\varvec{\phi }}:{\tilde{V}}\times {\tilde{V}}\rightarrow {\mathbb {R}}\) given by
Within this paper, we assume that \({{\tilde{a}}}_{\varvec{\phi }}\) is bounded and coercive on \({\tilde{V}}\times {\tilde{V}}\). Obviously, the bilinear form \(a_{\varvec{\phi }}\) inherits these properties such that \(a_{{\varvec{\phi }}}\) is also bounded and coercive on \(V\times V\).
Introducing the Lagrangian \({\mathcal {L}}({\varvec{\phi }},\Lambda )={\mathcal {E}}({\varvec{\phi }})-\frac{1}{2}{{\,\textrm{tr}\,}}\big (\Lambda ^T(\llbracket {{\varvec{\phi }}},{{\varvec{\phi }}}\rrbracket _H - I_p)\big )\) with a Lagrange multiplier \(\Lambda \in {\mathcal {S}}_{\textrm{sym}}(p)\), the first-order necessary optimality conditions for the minimization problem (5) yield the nonlinear eigenvector problem (NLEVP)
with unknown \({\varvec{\phi }}_*\in V\), which is referred to as the eigenvector, and \(\Lambda _*\in {\mathcal {S}}_{\textrm{sym}}(p)\), whose eigenvalues are the lowest p eigenenergies of the system. Yet another formulation of the NLEVP () follows from the special structure of the bilinear form \(a_{\varvec{\phi }}\) given in (9): seek \({\varvec{\phi }}_*=(\phi _{*,1}, \dots , \phi _{*,p})\in \text {St}(p,V)\) and p eigenvalues \(\lambda _1,\dots ,\lambda _p\in {\mathbb {R}}\) such that
For fixed \({\varvec{\phi }}\in V\), we introduce the operator \({\mathcal {A}}_{\varvec{\phi }}:V\rightarrow V^*\), called the Hamiltonian, which is defined by
Then the NLEVP () can be written as
where \({{\textbf {0}}}^{*}\in V^*\) is the zero functional. Using the left inverse \({\mathcal {J}}\) of \({\mathcal {I}}\), we find that
Remark 1
Due to the symmetry of the bilinear form \({{\tilde{a}}}_{\varvec{\phi }}\), we conclude that
for \(i,j=1,\ldots ,p\). This means that \(\llbracket {{\varvec{\phi }}},{{\mathcal {J}}{\mathcal {A}}_{\varvec{\phi }}\,{\varvec{\phi }}}\rrbracket _H\) is symmetric for any \({\varvec{\phi }}\in V\).
3 The Infinite-Dimensional Stiefel and Grassmann Manifolds
In this section, we summarize definitions and properties of the infinite-dimensional Stiefel and Grassmann manifolds and their tangent spaces, which lay the foundation of the Riemannian optimization schemes in the upcoming section.
3.1 The Stiefel Manifold
We consider the infinite-dimensional Stiefel manifold \(\text {St}(p,V)\) defined in (4). It is an embedded submanifold of the Hilbert space V and has co-dimension \(p\,(p+1)/2\); see [8]. The tangent space of \(\text {St}(p,V)\) at \({\varvec{\phi }}\in \text {St}(p,V)\) is given by
The Riemannian structure of the Stiefel manifold \(\text {St}(p,V)\) strongly depends on an underlying metric. Within this paper, we equip \(\text {St}(p,V)\) with the metric given by
The normal space with respect to g is then defined as
It can also be represented as
Further, any \(\varvec{y}\in V\) can be decomposed as \(\varvec{y}= {\mathcal {P}}_{{\varvec{\phi }}}^{}(\varvec{y}) + {\mathcal {P}}_{{\varvec{\phi }}}^\perp (\varvec{y})\), where
are the orthogonal projections onto the tangent and normal spaces, respectively.
The Riemannian gradient of a smooth function \(\,{\mathcal {E}}:\text {St}(p,V)\rightarrow {\mathbb {R}}\) with respect to the metric g is the unique element \({{\,\textrm{grad}\,}}{\mathcal {E}}({\varvec{\phi }})\in T_{\varvec{\phi }}\,\text {St}(p,V)\) satisfying the condition
where \({\overline{{\mathcal {E}}}}\) denotes a smooth extension of \({\mathcal {E}}\) around \({\varvec{\phi }}\) in V and \({{\,\textrm{D}\,}}{\overline{{\mathcal {E}}}}({\varvec{\phi }})\) is the Fréchet derivative of \(\,{\overline{{\mathcal {E}}}}\) in V.
For the energy functional \({\mathcal {E}}\) in (3), the Riemannian gradient at \({\varvec{\phi }}\in \text {St}(p,V)\) with respect to the metric g can be determined by using the \(L^2\)-Sobolev gradient \(\nabla \,{\overline{{\mathcal {E}}}}({\varvec{\phi }})\in V\) which is defined as the Riesz representation of \({{\,\textrm{D}\,}}{\overline{{\mathcal {E}}}}({\varvec{\phi }})\) in the Hilbert space V with respect to the inner product \((\,\cdot ,\,\cdot \,)_H\). Then, for all \(\varvec{w}\in V\), we have
and, hence, \(\nabla \,{\overline{{\mathcal {E}}}}({\varvec{\phi }}) ={\mathcal {J}}{\mathcal {A}}_{\varvec{\phi }}\,{\varvec{\phi }}\). Furthermore, for all \({\varvec{\eta }}\in T_{{\varvec{\phi }}}\,\text {St}(p,V)\), we obtain
This implies that
The Riemannian Hessian of \(\,{\mathcal {E}}\) at \({\varvec{\phi }}\in \text {St}(p,V)\) with respect to the metric g, denoted by \({{\,\textrm{Hess}\,}}{\mathcal {E}}({\varvec{\phi }})\), is a linear mapping on the tangent space \(T_{\varvec{\phi }}\,\text {St}(p,V)\) into itself which is defined by
where \(\nabla _{{\varvec{\eta }}}\) denotes the covariant derivative along \({\varvec{\eta }}\) with respect to the connection \(\nabla \), cf. [1, Sect. 5.3] for the finite-dimensional case.
The following theorem provides two expressions for the Riemannian Hessian of \({\mathcal {E}}\) in terms of the directional derivative of \({{\,\textrm{grad}\,}}{\mathcal {E}}({\varvec{\phi }})\) and the \(L^2\)-Sobolev Hessian \(\nabla ^2\,{\overline{{\mathcal {E}}}}({\varvec{\phi }})\) of \(\,{\overline{{\mathcal {E}}}}\), which is a linear operator mapping \(\varvec{v}\in V\) onto the Riesz representation of \({{\,\textrm{D}\,}}^2{\overline{{\mathcal {E}}}}({\varvec{\phi }})[\varvec{v},\,\cdot \,]\) with respect to the inner product \((\,\cdot ,\cdot \,)_H\).
Theorem 1
Let \({\varvec{\phi }}\in \text {St}(p,V)\) and \({\varvec{\eta }}\in T_{{\varvec{\phi }}}\,\text {St}(p,V)\). Then the Riemannian Hessian of a smooth function \({\mathcal {E}}:\text {St}(p,V)\rightarrow {\mathbb {R}}\) admits the expressions
where \(\nabla \,{\overline{{\mathcal {E}}}}({\varvec{\phi }})\) and \(\nabla ^2\,{\overline{{\mathcal {E}}}}({\varvec{\phi }})\) denote, respectively, the \(L^2\)-Sobolev gradient and the \(L^2\)-Sobolev Hessian of a smooth extension \(\,{\overline{{\mathcal {E}}}}\) of \(\,{\mathcal {E}}\) around \({\varvec{\phi }}\) in V.
Proof
Since \(\text {St}(p,V)\) is an embedded submanifold of the Hilbert space V, the expression (19) can be shown similarly to the finite-dimensional case [1, Prop. 5.3.2].
In order to prove (20), we first compute the directional derivative
Let \(\varvec{c}(t)\subset \text {St}(p,V)\) be a smooth curve defined on a neighborhood of \(t=0\) such that \(\varvec{c}(0)={\varvec{\phi }}\) and \(\tfrac{\textrm{d}}{\textrm{d}t}\varvec{c}(0)={\varvec{\eta }}\). Then for all \(\varvec{y}\in V\), we have
Inserting (21) into (19) and taking into account that
we obtain (20). \(\square \)
In order to derive a formula for the Riemannian Hessian of the energy functional \({\mathcal {E}}\) in (3), we first compute the second-order derivative
where \(\beta (\rho )=\frac{{\textrm{d}}}{{{\textrm{d}}}\rho }\gamma (\rho )\) and the operator \({\mathcal {B}}_{\varvec{\phi }}:V\rightarrow V^*\) has the form
Hence, the \(L^2\)-Sobolev Hessian of \(\,{\overline{{\mathcal {E}}}}\) is given by \(\nabla ^2\,{\overline{{\mathcal {E}}}}({\varvec{\phi }})[\varvec{v}]={\mathcal {J}}\, {\mathcal {A}}_{\varvec{\phi }}\, \varvec{v}+ {\mathcal {J}}\,{\mathcal {B}}_{\varvec{\phi }}\,\varvec{v}\) for all \(\varvec{v}\in V\). By the definition of the orthogonal projection onto \(T_{\varvec{\phi }}\,\text {St}(p,V)\) in (17), we conclude that for \({\varvec{\eta }}\in T_{\varvec{\phi }}\,\text {St}(p,V)\), the Riemannian Hessian of \({\mathcal {E}}\) is given by
Within optimization methods, we need to transfer data from the tangent space to the manifold to keep the iterations on the search space. For this purpose, we can use retractions defined as follows. Let \(T\text {St}(p,V)\) be the tangent bundle to \(\text {St}(p,V)\). A smooth mapping \({\mathcal {R}}:T\text {St}(p,V)\rightarrow \text {St}(p,V)\) is called a retraction if for all \({\varvec{\phi }}\in \text {St}(p,V)\), the restriction of \({\mathcal {R}}\) to \(T_{\varvec{\phi }}\,\text {St}(p,V)\), denoted by \({\mathcal {R}}_{\varvec{\phi }}\), satisfies the following properties:
-
1)
\({\mathcal {R}}_{\varvec{\phi }}(\varvec{0}_{\varvec{\phi }})={\varvec{\phi }}\), where \(\varvec{0}_{\varvec{\phi }}\) denotes the origin of \(T_{\varvec{\phi }}\,\text {St}(p,V)\), and
-
2)
\(\tfrac{\textrm{d}}{\textrm{d}t} {\mathcal {R}}_{\varvec{\phi }}(t{\varvec{\eta }})\big |_{t=0}={\varvec{\eta }}\) for all \({\varvec{\eta }}\in T_{\varvec{\phi }}\,\text {St}(p,V)\).
Retractions provide first-order approximations to the exponential mapping on a Riemannian manifold and are often much easier to compute. A retraction \({\mathcal {R}}\) on \(\text {St}(p,V)\) is of second-order, if it satisfies \(\tfrac{\textrm{d}^2}{\textrm{d}t^2} {\mathcal {R}}_{\varvec{\phi }}(t{\varvec{\eta }})\big |_{t=0}\in T_{\varvec{\phi }}^{\perp }\,\text {St}(p,V)\) for all \(({\varvec{\phi }},{\varvec{\eta }})\in T\text {St}(p,V)\).
In [8], several retractions on the Stiefel manifold \(\text {St}(p,V)\) have been introduced. They can be considered as an extension of the corresponding concepts on the matrix Stiefel manifold (see, e.g., [2, 46]) to the infinite-dimensional case.
For \(\varvec{v}\in V\) with linearly independent components, we consider the qR decomposition \(\varvec{v}=\varvec{q}R\), where \(\varvec{q}\in \text {St}(p,V)\) and \(R\in {\mathbb {R}}^{p\times p}\) is upper triangular. Such a decomposition exists and is unique if we additionally require that R has positive diagonal elements. Then the qR decomposition based retraction is defined as \({\mathcal {R}}^{qR}({\varvec{\phi }},{\varvec{\eta }})={{\,\textrm{qf}\,}}({\varvec{\phi }}+{\varvec{\eta }})\), where \({{\,\textrm{qf}\,}}({\varvec{\phi }}+{\varvec{\eta }})\) denotes the factor from \(\text {St}(p,V)\) in the qR decomposition of \({\varvec{\phi }}+{\varvec{\eta }}\). Such a factor can be computed, e.g., by the modified Gram-Schmidt orthonormalization procedure on V presented in [8].
An alternative retraction can be defined by using the polar decomposition \(\varvec{v}=\varvec{u}S\), where \(\varvec{u}\in \text {St}(p,V)\) and \(S\in {\mathbb {R}}^{p\times p}\) is symmetric and positive definite. Assuming that the components of \(\varvec{v}\) are linearly independent, \(S=\llbracket {\varvec{v}},{\varvec{v}}\rrbracket _H^{1/2}\) and \(\varvec{u}=\varvec{v}\,\llbracket {\varvec{v}},{\varvec{v}}\rrbracket _H^{-1/2}\) are uniquely defined. This leads to the polar decomposition based retraction
which is of second-order. Indeed, computing the second-order derivative of \({\mathcal {R}}^\textrm{pol}_{\varvec{\phi }}(t{\varvec{\eta }})\) at \(t=0\) and exploiting (16), we obtain that
Note that second-order retractions are advantageous for second-order Riemannian optimization methods; see, e.g. [1, Sect. 6.3].
3.2 The Grassmann Manifold
Let \(\text {O}(p)\) be the orthogonal group of \({\mathbb {R}}^{p\times p}\). Following [47], we define the infinite-dimensional Grassmann manifold as the quotient
of the Stiefel manifold \(\text {St}(p,V)\) with respect to the equivalence relation
The Grassmann manifold \(\text {Gr}(p,V)\) can be interpreted as the set of the equivalence classes given by
for \({\varvec{\phi }}\in \text {St}(p,V)\). Similarly to the Grassmann matrix manifold [1, Prop. 3.4.6], one can show that \(\text {Gr}(p,V)\) admits a unique structure of quotient manifold. A canonical projection from the Stiefel manifold into the Grassmann manifold is defined by
and is a smooth submersion. This means that \({{\,\textrm{D}\,}}\! \pi ({\varvec{\phi }})\) is surjective, and, hence, the equivalence class is an embedded submanifold of \(\text {St}(p,V)\); see [1, Prop. 3.4.4.].
In the following, we examine a useful connection of the Stiefel manifold and the Grassmann manifold. More precisely, we show that there is a one-to-one relation between the tangent space of the Grassmann manifold and the so-called horizontal space, a subspace of the tangent space of the Stiefel manifold. The tangent space \(T_{\varvec{\phi }}\, \text {St}(p,V)\) at \({\varvec{\phi }}\in \text {St}(p,V)\) defined in (14) can be splitted with respect to the projection \(\pi \) and the metric g as \(T_{\varvec{\phi }}\, \text {St}(p,V)= {\mathcal {V}}_{{\varvec{\phi }}} \oplus {\mathcal {H}}_{\varvec{\phi }}\), where
is the vertical space at \({\varvec{\phi }}\) and
is the horizontal space at \({\varvec{\phi }}\); see [47, Lem. 2]. The orthogonal projection of a tangent vector \({\varvec{\eta }}\in T_{\varvec{\phi }}\text {St}(p,V)\) onto \({\mathcal {H}}_{{\varvec{\phi }}}\) is given by
One can see that, moving on a curve in the Stiefel manifold \(\text {St}(p,V)\) with direction in the vertical space \({\mathcal {V}}_{{\varvec{\phi }}}\), we stay in the equivalence class . The tangent space of the Grassmann manifold \(\text {Gr}(p,V)\) can then be identified with the horizontal space \({\mathcal {H}}_{{\varvec{\phi }}}\) in the sense that for any , there exists a unique \({\varvec{\psi }}_{{\varvec{\phi }}}^\textrm{h}\in {\mathcal {H}}_{{\varvec{\phi }}}\) such that \({{\,\textrm{D}\,}}\!\pi ({\varvec{\phi }})[{\varvec{\psi }}_{{\varvec{\phi }}}^\textrm{h}]={\varvec{\psi }}\). The unique element \({\varvec{\psi }}_{{\varvec{\phi }}}^\textrm{h}\) is called the horizontal lift of \({\varvec{\psi }}\) at \({\varvec{\phi }}\). This relation allows us to introduce a metric on the Grassmann manifold \(\text {Gr}(p,V)\), namely
where \({\varvec{\psi }}_{{\varvec{\phi }}}^\textrm{h}, {\varvec{\zeta }}_{{\varvec{\phi }}}^\textrm{h}\in {\mathcal {H}}_{{\varvec{\phi }}}\) are the horizontal lifts of \({\varvec{\psi }}\) and \({\varvec{\zeta }}\) at \({\varvec{\phi }}\), respectively. Due to \({\varvec{\psi }}_{{\varvec{\phi }}Q}^\textrm{h}={\varvec{\psi }}_{{\varvec{\phi }}}^\textrm{h}Q\) for all \(Q\in \text {O}(p)\), one can show that this metric does not depend on the choice of the representative \({\varvec{\phi }}\) of the equivalence class .
The connection of and \({\mathcal {H}}_{{\varvec{\phi }}}\) makes it possible to introduce optimization methods on the Grassmann manifold, while still working on the tangent space of the corresponding Stiefel manifold. Using the canonical projection \(\pi \), the minimization problem (5) on the Stiefel manifold \(\text {St}(p,V)\) can be written as the minimization problem
on the Grassmann manifold \(\text {Gr}(p,V)\), where the cost functional \({\mathcal {F}}\!:\! \text {Gr}(p,V)\!\rightarrow \!{\mathbb {R}}\) is induced by \({\mathcal {E}}\) as \({\mathcal {E}}({\varvec{\phi }})={{\mathcal {F}}}(\pi ({\varvec{\phi }}))\) and . Note that this definition is justified by the fact that \({\mathcal {E}}({\varvec{\phi }})={\mathcal {E}}({\varvec{\phi }}Q)\) for all \(Q\in \text {O}(p)\). The horizontal lift of the Riemannian gradient with respect to the metric \(g^\textrm{Gr}\) is given by
To obtain the horizontal lift of the Riemannian Hessian , we proceed as before but replace the projection \({\mathcal {P}}_{{\varvec{\phi }}}\) by the orthogonal projection \({\mathcal {P}}_{{\varvec{\phi }}}^\textrm{h}\) onto the horizontal space; see Eq. (25). This leads to
Retractions on the Grassmann manifold are inherited from that on the Stiefel manifold applied to the horizontal lift; see [1, Prop. 4.1.3]. For all and , we have
Note that these retractions are independent of the chosen point \({\varvec{\phi }}\), providing the same equivalence class on \(\text {Gr}(p,V)\).
Similar to the matrix case [1, 25], we can also derive an explicit expression for the Grassmann exponential \(\textrm{Exp}:T\text {Gr}(p,V)\rightarrow \text {Gr}(p,V)\), which maps to the end point of the unique geodesic starting at and going in the direction \({\varvec{\psi }}\). Let \({\varvec{\psi }}_{{\varvec{\phi }}}^\textrm{h}=\varvec{u}\Sigma W^T\) be a singular value decomposition of the horizontal lift \({\varvec{\psi }}_{{\varvec{\phi }}}^\textrm{h}\) of \({\varvec{\psi }}\), where \(\varvec{u}\in \text {St}(p,V)\), \(W\in \text {O}(p)\), and \(\Sigma \in {\mathbb {R}}^{p\times p}\) is diagonal with nonnegative diagonal elements. Then the Grassmann exponential is given by
Using \({\varvec{\psi }}_{{\varvec{\phi }}}^\textrm{h}\in {\mathcal {H}}_{\varvec{\phi }}\), one can verify that \({\varvec{\phi }}\,W\cos \Sigma +\varvec{u}\sin \Sigma \in \text {St}(p,V)\). Therefore, it can be considered as a representative of the resulting equivalence class.
Remark 2
For \(p=1\), the Stiefel manifold coincides with the Grassmann manifold and equals the unit sphere
Its tangent space is given by \(T_{\phi }\,{\mathbb {S}} = \{ \eta \in {\tilde{V}}: (\eta , \phi )_{L^2(\Omega )} = 0 \}\) and the orthogonal projection onto this space takes the form \({\mathcal {P}}_{\phi }(y)=y-(\phi ,y)_{L^2(\Omega )}\phi \) for \(y\in {\tilde{V}}\). Furthermore, for \((\phi ,\eta )\in T{\mathbb {S}}\), the second-order retraction and the exponential mapping on \({\mathbb {S}}\) are given by
respectively.
4 Riemannian Newton Methods
In this section, we present Riemannian Newton methods on the Stiefel manifold as well as on the Grassmann manifold and discuss the inexact version of the latter.
Within the Riemannian Newton method on the Stiefel manifold \(\text {St}(p,V)\), for given iterate \({\varvec{\phi }}_k\in \text {St}(p,V)\), we first compute the Newton search direction \({\varvec{\eta }}_k\in T_{{\varvec{\phi }}_k} \text {St}(p,V)\) by solving the Newton equation
The iterate is then updated by \({\varvec{\phi }}_{k+1}={\mathcal {R}}({\varvec{\phi }}_k,{\varvec{\eta }}_k)\) for any retraction on \(\text {St}(p,V)\). It should, however, be noted that due to the non-uniqueness of the minimizer of (5) caused by the invariance of \({\mathcal {E}}\) under orthogonal transformations, we cannot expect that \({{\,\textrm{Hess}\,}}{\mathcal {E}}({\varvec{\phi }}_k)\) is invertible on \(T_{{\varvec{\phi }}_k} \text {St}(p,V)\). This is, indeed, vindicated by the following theorem.
Theorem 2
Let \({\varvec{\phi }}\in \text {St}(p,V)\) and let \({\mathcal {V}}_{{\varvec{\phi }}}\) be the vertical space given in (24). Then the Riemannian Hessian of the energy functional \({\mathcal {E}}\) from (3) in \({\varvec{\phi }}\) is non-invertible on \({\mathcal {V}}_{{\varvec{\phi }}}\), i.e., \(\big ({\varvec{\xi }},{{\,\textrm{Hess}\,}}{\mathcal {E}}({\varvec{\phi }}) [{\varvec{\eta }}]\big )_H=0\) for all \({\varvec{\eta }},{\varvec{\xi }}\in {\mathcal {V}}_{{\varvec{\phi }}}\).
Proof
Let \({\varvec{\eta }},{\varvec{\xi }}\in {\mathcal {V}}_{{\varvec{\phi }}}\) be arbitrary. Then there exist the matrices \(\varTheta _{{\varvec{\eta }}},\varTheta _{{\varvec{\xi }}}\in {\mathcal {S}}_{\textrm{skew}}(p)\) such that \({\varvec{\eta }}={\varvec{\phi }}\,\varTheta _{{\varvec{\eta }}}\) and \({\varvec{\xi }}={\varvec{\phi }}\,\varTheta _{{\varvec{\xi }}}\). Using the definition of \({{\,\textrm{Hess}\,}}{\mathcal {E}}({\varvec{\phi }})\) in (23) and the symmetry of the matrix \(\llbracket {{\varvec{\phi }}},{{\mathcal {J}}{\mathcal {A}}_{\varvec{\phi }}\, {\varvec{\phi }}}\rrbracket _H\) shown in Remark 1, we have
and, hence,
We now show that \(\llbracket {{\varvec{\phi }}},{{\mathcal {J}}{\mathcal {B}}_{\varvec{\phi }}\,{\varvec{\phi }}}\rrbracket _H \,\varTheta _{{\varvec{\eta }}}= 0\). With the skew-symmetric matrix \(\varTheta _{{\varvec{\eta }}} = [\theta _{ij}]_{i,j=1}^p\), we first observe that
This implies \({\varvec{\phi }}\cdot ({\varvec{\phi }}\,\varTheta _{{\varvec{\eta }}}) = 0\) and, hence, \(\llbracket {{\varvec{\phi }}},{{\mathcal {J}}{\mathcal {B}}_{\varvec{\phi }}\,{\varvec{\phi }}}\rrbracket _H \,\varTheta _{{\varvec{\eta }}}= 0\). As a result, we conclude that \(\big ({\varvec{\xi }},{{\,\textrm{Hess}\,}}{\mathcal {E}}({\varvec{\phi }}) [{\varvec{\eta }}]\big )_H=0\) for all \({\varvec{\eta }},{\varvec{\xi }}\in {\mathcal {V}}_{{\varvec{\phi }}}\). \(\square \)
It follows from Theorem 2 that if Eq. (29) is solvable, its solution is not unique. To overcome this difficulty, we pass on to the Grassmann manifold \(\text {Gr}(p,V)\). Given , the Newton direction is computed by solving the Newton equation
By applying the horizontal lift expressions (27) and (28), this equation leads to
with unknown \(({\varvec{\psi }}_k)_{{\varvec{\phi }}_k}^\textrm{h}\in {\mathcal {H}}_{{\varvec{\phi }}_k}\) being the horizontal lift of \({\varvec{\psi }}_k\) at \({\varvec{\phi }}_k\). Note that this equation is well-defined, since \(-{{\,\textrm{grad}\,}}{\mathcal {E}}({\varvec{\phi }}_k)\) is an element of the horizontal space \({\mathcal {H}}_{{\varvec{\phi }}_k}\).
For solving the Newton equation (30), we can employ any matrix-free iterative linear solver which does not require the storage of the coefficient matrix explicitly but accesses it by computing the matrix–vector product or—as in our case—by evaluating the linear operator given in (28). The resulting Riemannian Newton method on the Grassmann manifold is presented in Algorithm 1.
The Newton equation (30) can also be formulated as a saddle point problem. To this end, we introduce the bilinear form
Then, the equivalent problem to (30) reads: find \(({\varvec{\psi }}_k)_{{\varvec{\phi }}_k}^\textrm{h}\in V\) and a Lagrange multiplier \(M_k\in {\mathbb {R}}^{p\times p}\) such that
The constraint (31b) implies that \(({\varvec{\psi }}_k)_{{\varvec{\phi }}_k}^\textrm{h}\in {\mathcal {H}}_{{\varvec{\phi }}_k}\). Further note that any function \(\varvec{w}\in {\mathcal {H}}_{{\varvec{\phi }}_k}\) satisfies \(\llbracket {{\varvec{\phi }}_k},{\varvec{w}}\rrbracket _H= 0_p\). Hence, for all \(\varvec{w}\in {\mathcal {H}}_{{\varvec{\phi }}_k}\), Eq. (31a) reads
which is equivalent to the Newton equation (30).
One important property guaranteeing an isolated local minimum of the energy is that the Hessian is positive at a stationary point. For a global minimizer of (5), denoted by \({\varvec{\phi }}_*\), we consider the following linear eigenvalue problem: seek \(\phi \in {\tilde{V}}\) and \(\lambda \in {\mathbb {R}}\) such that
Then, due to (11), we know that the components of \({\varvec{\phi }}_* = (\phi _{*,1}, \dots , \phi _{*,p})\) satisfy Eq. (32) together with the smallest p eigenvalues denoted by \(0 < \lambda _1 \le \dots \le \lambda _p\). In the following, we will assume that these eigenfunctions can be extended to a basis of \({\tilde{V}}\).
Assumption 3
(Basis and spectral gap) The eigenfunctions \(\phi _{*,1}, \phi _{*,2}, \ldots \in {\tilde{V}}\) of the eigenvalue problem (32) form an \(L^2\)-orthonormal basis of \({\tilde{V}}\). The corresponding eigenvalues \(\lambda _1 \le \lambda _2 \le \dots \) are ordered by size with a spectral gap \(\lambda _p < \lambda _{p+1}\).
Theorem 4
(Positive Hessian) Let \({\varvec{\phi }}_*\) be a global minimal solution of (5) and let the corresponding eigenvalue problem (32) satisfy Assumption 3. Furthermore, assume that the operator \({\mathcal {B}}_{{\varvec{\phi }}_*}\) fulfills \(({\mathcal {J}}{\mathcal {B}}_{{\varvec{\phi }}_*} {\varvec{\psi }}_{{\varvec{\phi }}_*}^\textrm{h}, {\varvec{\psi }}_{{\varvec{\phi }}_*}^\textrm{h})_H\!\ge \! 0\) for all \({\varvec{\psi }}_{{\varvec{\phi }}_*}^\textrm{h}\in {\mathcal {H}}_{{\varvec{\phi }}_*}\). Then the Riemannian Hessian of \({\mathcal {F}}\) at is positive, i.e.,
for all nonzero .
Proof
We extend the proof of [54, Th. 5.1], which considers the finite-dimensional case for the simplified Kohn–Sham problem, to the infinite-dimensional case in a more general setting. We know from (28) that the horizontal lift of the Riemannian Hessian of \({\mathcal {F}}\) at \({\varvec{\phi }}_*\) takes the form
For the first term, we make the following considerations. Due to the orthogonality, each component of \({\varvec{\psi }}_{{\varvec{\phi }}_*}^\textrm{h}\) satisfies \((\phi _{*,l}, ({\varvec{\psi }}_{{\varvec{\phi }}_*}^\textrm{h})_j)_{L^2(\Omega )} = 0\) for \(l,j= 1,\dots , p\). Hence, the jth component of \({\varvec{\psi }}_{{\varvec{\phi }}_*}^\textrm{h}\) takes the form \(\psi _j = ({\varvec{\psi }}_{{\varvec{\phi }}_*}^\textrm{h})_j = \sum _{l>p} \alpha _{lj} \phi _{*,l}\) for some coefficients \(\alpha _{lj}\in {\mathbb {R}}\) and \(\phi _{*,l}\) denoting the basis from Assumption 3. As a consequence, we get
for all \(i,j=1,\ldots , p\) and, hence,
For the second term, we get with the assumption on \({\mathcal {B}}_{{\varvec{\phi }}_*}\) that
Finally, by using (13), the third term takes the form
Let the columns of \(U\in \text {O}(p)\) form a basis of eigenvectors corresponding to the eigenvalues \(\lambda _1,\ldots ,\lambda _p\) of \(\Lambda _*\) and let \({\varvec{\psi }}_{{\varvec{\phi }}_*}^\textrm{h}U=({\tilde{\psi }}_1,\ldots ,\tilde{\psi _p})\). Due to the assumed spectral gap, this yields all together
which completes the proof. \(\square \)
Remark 3
(Connection to the Lagrange–Newton method) The optimal solution of the constrained minimization problem (5) can also be determined by the Lagrange–Newton method. Based on the first-order optimality conditions () with a symmetric Lagrange multiplier, we aim to solve the nonlinear system of equations
Computing the Jacobian of f, the Lagrange-Newton iteration is given as follows: for given \({\varvec{\phi }}_k\in V\) and \(\Lambda _k\in {\mathbb {R}}^{p\times p}\), solve the equations
for \({\varvec{\eta }}_k\in V\), \(\Xi _k\in {\mathbb {R}}^{p\times p}\) and update \({\varvec{\phi }}_{k+1}={\varvec{\phi }}_k+{\varvec{\eta }}_k\), \(\Lambda _{k+1}=\Lambda _k+\Xi _k\). Note that \({\varvec{\phi }}_{k+1}\) does not necessarily belong to \(\text {St}(p,V)\). Assuming \({\varvec{\phi }}_k\in \text {St}(p,V)\) and \(\Lambda _k\in {\mathcal {S}}_{\textrm{sym}}(p)\), however, Eqs. (33b) and (33c) imply that \({\varvec{\eta }}_k\in T_{{\varvec{\phi }}_k}\text {St}(p,V)\) and \(\Xi _k\in {\mathcal {S}}_{\textrm{sym}}(p)\), respectively. Resolving Eq. (33a) for symmetric \(\Xi _k\), we find that
Inserting this matrix into (33a) yields the Newton equation (29). This shows that the Lagrange–Newton method with the modified update
is equivalent to the Riemannian Newton method on the Stiefel manifold.
5 Examples and Numerical Experiments
This section is devoted to the numerical investigation of the Riemannian Newton methods. To this end, we consider the Gross–Pitaevskii eigenvalue problem from Example 1 and the Kohn–Sham model from Example 2.
5.1 Gross–Pitaevskii Eigenvalue Problem
The minimization of the Gross–Pitaevskii energy functional \({\mathcal {E}}_{\textrm{GP}}\) in (6) leads to the following nonlinear eigenvector problem: find \(\phi \in {\tilde{V}} = H^1_0(\Omega )\) with \(\Vert \phi \Vert _{L^2(\Omega )} = 1\) and \(\lambda \in {\mathbb {R}}\) such that
for some space-dependent external potential \(\vartheta \ge 0\) and an interaction constant \(\kappa >0\). The latter means that the particle interactions are repulsive, i.e., we consider the so-called defocussing regime. In this case, we get the operators
for \(v,w\in {\tilde{V}}\). One can see that the bilinear form defined through (35a) corresponds to the Laplacian with the \(L^2\)-shift \(2\vartheta +\kappa \phi ^2\). Assuming this shift to be constant and \(\Omega =(0,1)^d\) as the spatial domain, Assumption 3 is satisfied; see [50, Ch. 12]. Moreover, the nonlinear operator from (35b) fulfills
such that Theorem 4 is applicable.
For the spatial discretization of the Gross–Pitaevskii problem (34), we use a biquadratic finite element method on a Cartesian mesh of width h; see [16] for the corresponding error analysis. The resulting discrete eigenvalue problem reads
with \(\varphi \in {\mathbb {R}}^n\), where n denotes the number of degrees of freedom. Here, A is the stiffness matrix, M is the mass matrix, and \(M_{\vartheta }\) and \(M_{\varphi ^2}\) are the weighted mass matrices, respectively, where \(\varphi ^2\) should be understood as the elementwise product. Then the discrete version of \(\llbracket {\phi },{{\mathcal {J}}\!{\mathcal {A}}_\phi \,\phi }\rrbracket _H=\big (\phi ,{\mathcal {J}}\!{\mathcal {A}}_\phi \,\phi \big )_{L^2(\Omega )}\) equals
and the Newton equation takes the form
with unknown \(\psi \in \{\xi \in {\mathbb {R}}^n\;:\; \xi ^TM\varphi = 0\}=\text {im}(I-\varphi \varphi ^T\! M)\).
We demonstrate the performance of the resulting Riemannian Newton method in comparison with the SCF iteration combined with the optimal damping algorithm (ODA) proposed in [23] and the energy-adaptive Riemannian gradient descent method (RGD) of [29] with a non-monotone step size control as outlined in [8]. The numerical experiments are performed on a sufficiently large bounded domain \(\Omega = (-L,L)^2\), \(L=8\), for two types of trapping potentials. In Sect. 5.1.1, we consider a simple harmonic trap, whereas in Sect. 5.1.2, we add an additional disorder potential. The interaction parameter \(\kappa \) as well as the spatial resolution h will be specified separately for each case.
5.1.1 Ground State in a Harmonic Trap
For the first numerical experiment, we consider the harmonic trapping potential
and the interaction parameters \(\kappa =10,100,1000\). The resulting ground states computed on a Cartesian mesh of width \(h/(2\,L) = 2^{-10}\) are depicted in Fig. 1.
To generate a joint and sufficiently accurate initial value for the three solvers of the discretized nonlinear eigenvector problem, we run the energy-adaptive RGD method starting from the biquadratic finite element interpolation of the constant 1 (respecting the homogeneous Dirichlet boundary condition). We stopped this iteration once the residual fell below the \(10^{-2}\) tolerance and used the approximated ground state as the initial state to compare the asymptotic behavior of the three different solvers. The corresponding convergence histories are presented in Fig. 2 showing the evolution of the residuals during the iteration processes. It can be observed that the Riemannian Newton method (with sparse direct solution of the Newton equation using the Sherman–Morrison formula [48]) reaches the tolerance of \(10^{-8}\) in only three steps. While the performances of the SCF iteration and the energy-adaptive RGD method abate with increasing \(\kappa \), the Riemannian Newton scheme appears to be extremely robust. We would like to emphasize that, although one Newton step is slightly more expensive than one step of any other competing method, the overall costs are much smaller, especially for increasing \(\kappa \).
The convergence behavior of the Riemannian Newton scheme is also robust to the underlying mesh size h and, hence, independent of the dimension of the discretization space. This is demonstrated in Fig. 3 with a fixed choice of \(\kappa =1000\). We consider a sequence of meshes with \(h/(2L) = 2^{-1},\ldots ,2^{-10}\) and use the same procedure as above to generate initial guesses with residuals of order \(10^{-2}\). The left graph shows the number of (outer) iterations of the Riemannian Newton method to fall below the tolerance of \(10^{-10}\) for each of these mesh sizes. An increase in the number of iterations with smaller mesh size is not observed. In our experience, this mesh independence of the Riemannian Newton optimization scheme is representative for many other choices of potentials and interaction parameters. Note, however, that this does not mean that the costs of a Newton step are independent of h. As for every Laplace-type problem, methods such as multigrid need to be implemented in order to obtain a mesh-independence also for the inner iteration. This holds for all competing methods in the same way. For completeness, Fig. 3 also shows the corresponding errors in the minimum energy approximation as a function of the mesh size, demonstrating the optimal fourth-order convergence rate of the biquadratic finite element implementation [30].
5.1.2 Localized Ground State in a Disorder Potential
The second experiment considers the computationally more difficult case where the external potential is the sum of the harmonic potential \(\vartheta _\text {harm}\) defined in (36) and a potential \(\vartheta _\text {rand}\) reflecting a high degree of disorder. The disorder part \(\vartheta _\text {rand}\) is chosen as a piecewise constant function on the Cartesian mesh of width \(2L\varepsilon \), \(\varepsilon = 2^{-6}\), taking values 0 or \(\varepsilon ^{-2}\) as depicted in Fig. 4. For a potential in such a scaling regime, the low-energy eigenstates essentially localize in terms of an exponential decay of their moduli relative to the small parameter \(\varepsilon \). For the linear case, i.e., for \(\kappa =0\), this has been analyzed in [4]. For growing \(\kappa \), the ground state consists of a growing number of localized peaks; see Fig. 4. Further details on the phenomenon of localization in the Gross–Pitaevskii equation and the onset of delocalization can be found in [6, 7]. As in the previous experiments, we use biquadratic finite elements on a Cartesian mesh of width \(h/(2L) = 2^{-10}\). To illustrate the localization behavior that occurs with the current parameter scaling for \(\kappa \lesssim 1\), we consider the interaction parameters \(\kappa =0.1,1,10\). The ground states for \(\kappa =1,10\) are shown in Fig. 4. The ground state for \(\kappa =0.1\) is hardly distinguishable from the one for \(\kappa =1\) and, therefore, it is not shown in a separate figure.
Figure 5 displays the convergence history of the residuals for \(\kappa =0.1,1,10\). We employed the same strategy as in Sect. 5.1.1 to generate suitable initial guesses with the residuals of order \(10^{-2}\) used for all methods. The results clearly indicate that the ground state computations with the disorder potential are already challenging for smaller values of \(\kappa \). Particularly, the energy-adaptive RGD method needs much larger iteration counts, which according to [27], may be related to smaller spectral gaps between the first and second eigenvalue. The Riemannian Newton method, on the other hand, still performs well and reaches the prescribed tolerance \(10^{-8}\) for the residual in only a few steps in all three examples. For comparison, the SCF iteration converges very fast in the almost linear case but suffers from larger values of \(\kappa \) as the energy-adaptive RGD method. Here, again, the higher costs per Newton step are compensated by far by the very small number of needed iteration steps.
5.2 Kohn–Sham Model
For the Kohn–Sham energy functional \({\mathcal {E}}_{\textrm{KS}}\) introduced in (7), we have
with \(\mu _{\textrm{xc}}(\rho ) = \frac{{\textrm{d}}}{{{\textrm{d}}}\rho } \big (\rho \,\epsilon _{\textrm{xc}}(\rho )\big )\). Moreover, the operator \({\mathcal {B}}_{\varvec{\phi }}\) has the form
where \(\zeta _{\textrm{xc}}(\rho )=\frac{\textrm{d}}{\textrm{d}\rho } \mu _\textrm{xc}(\rho )\). The exchange-correlation function \(\epsilon _\text {xc}(\rho )\) can additively be decomposed as \(\epsilon _\text {xc}(\rho ) =\epsilon _\text {x}(\rho ) + \epsilon _\text {c}(\rho )\), where the exchange component \(\epsilon _\text {x}(\rho )\) has the particular analytical expression \(\epsilon _\text {x}(\rho )=-\frac{3}{4}\big (\frac{3}{\pi }\rho \big )^{1/3}\) and the correlation component \(\epsilon _\text {c}(\rho )\) is usually unknown, but can be fitted by using quantum Monte-Carlo data [41]. For the numerical experiments, we use the MATLAB toolbox KSSOLV [35, 51], in which the correlation component is implemented as
where \(r(\rho )=\big (\frac{4\pi }{3}\rho \big )^{-1/3}\) is the Wigner-Seitz radius, and \(a_j,b_j\in {\mathbb {R}}\) are fitted constants; see [42, App. C].
For the spatial discretization, we employ the planewave discretization method as implemented in KSSOLV. With n denoting the number of degrees of freedom, the matrix \(\Phi \in {\mathbb {C}}^{n\times p}\) contains the coefficients of the approximation of the wave function \({\varvec{\phi }}\). Then the discretized Kohn–Sham energy functional is given by
where \(\Phi ^*\) denotes the complex conjugate transpose of \(\Phi \), \(L\in {\mathbb {C}}^{n\times n}\) is the discrete Laplace matrix, \(L^+\in {\mathbb {C}}^{n\times n}\) is its pseudoinverse, \(D_\textrm{ion}\in {\mathbb {R}}^{n\times n}\) is the discretized ionic potential, and \(\rho _h(\Phi ) = {{\,\textrm{diag}\,}}(\Phi \Phi ^*) \in {\mathbb {R}}^n\) is the discretized electronic charge density. Note that the matrix L is Hermitian and \(D_{\textrm{ion}}\) is diagonal. In this setting, the minimization problem
on the (compact) Stiefel manifold \( \text {St}(p,n)= \big \{ \Phi \in {\mathbb {C}}^{n\times p}\;:\; \Phi ^* \Phi = I_p \big \} \) leads to the finite-dimensional nonlinear eigenvector problem
where the discrete Kohn–Sham Hamiltonian is given by
Further, the Riemannian gradient of \(E(\Phi )\) becomes
In the Riemannian Newton method on the Stiefel manifold \(\text {St}(p,n)\), we need to solve the equation
for \(\Psi \) belonging to the tangent space \(T_\Phi \, \text {St}(p,n)\). Therein,
is the orthogonal projector onto \(T_\Phi \,\text {St}(p,n)\) and
is the discretization of \({\mathcal {B}}_{\varvec{\phi }}\). On the Grassmann manifold \(\text {Gr}(p,n)=\text {St}(p,n)/\text {U}(p)\) with the unitary group \(\text {U}(p)\), the Newton equation takes the form
for \(\Psi _\Phi ^\textrm{h}\in {\mathcal {H}}_{\Phi } = \{\Psi \in T_\Phi \, \text {St}(p,n)\;:\; \Phi ^*\Psi =0\}\).
In our experiments, we compare the calculation of the ground state by using the SCF iteration with the Anderson charge mixing scheme [51], the energy-adaptive RDG with non-monotone step size control [8], and the Riemannian Newton methods on the Stiefel manifold (RNS) and on the Grassmann manifold (RNG). In all these methods, we choose the same initial guess for the wave function by performing one SCF step with a randomly generated starting point and stop the iterations once the Frobenius norm of the residual \(R(\Phi _k) = A(\Phi _k)\Phi _k-\Phi _k\Lambda _k\) with \(\Lambda _k=\Phi _k^*A(\Phi _k)\Phi _k\) is smaller than the tolerance \(10^{-8}\). Note that due to (37), \(\Vert R(\Phi _k)\Vert _F=\Vert {{\,\textrm{grad}\,}}E(\Phi _k)\Vert _F\), i.e., the norms of the residuals provide the information on the size of the Riemannian gradients. In both Newton methods and the energy-adaptive RGD method, we use the qR decomposition based retractions. The reference minimal energy \(E_{\min }\) is computed by the RNG method with the tolerance \(10^{-10}\).
All algorithms are performed in an inexact manner, i.e., the occurring linear systems are only solved up to a certain tolerance. In RNG, for instance, we follow Algorithm 1 using the MATLAB built-in function minres as a linear system solver with the adaptive tolerance \(\min (1/k, 10^{-3} \Vert {{\,\textrm{grad}\,}}E(\Phi _{k-1})\Vert _F)\) and the maximal number of inner iterations \(\ell _{\max }=15\). The remaining parameters are chosen as \(\eta =10^{-8}\), \(\delta =0.5\), and \(\sigma =10^{-4}\). In RNS, we proceed similarly, with the only difference that instead of (39) we solve the Newton equation of the form (38). For solving the linear eigenvalue problems in SCF, we employ the KSSOLV built-in LOBPCG algorithm for the pentacene model in Sect. 5.2.1 and the MATLAB built-in function eigs for the graphene model in Sect. 5.2.2. Switching to another eigenvalue solver is necessary due to the ill-conditioning in LOBPCG for the latter example. In both cases, the tolerance for the inner iterations is set to be \(\min (10^{-3}, 10^{-3} \Vert {{\,\textrm{grad}\,}}E(\Phi _{k-1})\Vert _F)\). For the linear system solvers, we use the kinetic energy preconditioner, which provides, especially for the energy-adaptive RGD method, better numerical results than the KSSOLV built-in Teter–Payne–Allan preconditioner [49].
Finally, we would like to mention that similarly to the Gross–Pitaevskii example in Sect. 5.1, the Riemannian Newton schemes for the Kohn–Sham model are again robust in terms of the dimension of the discretization space used in KSSOLV. This means that the number of Newton iterations needed to fall below a certain tolerance is not effected by finer discretizations (as long as the number of inner iterations is sufficiently large).
5.2.1 Pentacene Molecule
In the first numerical experiment, we calculate the ground state for the pentacene molecule \(\textrm{C}_{22}\textrm{H}_{14}\) with \(p=51\) electron orbitals. A spatial planewave discretization on a \(80\times \ 55\times 160\) sampling grid gives the discrete model of dimension \(n=44791\). In Fig. 6, we present the convergence history of the residuals and the energy reduction during the iterations. One can see that the RNS and RNG methods have very similar behavior and converge within 8 and 10 iterations, respectively. In comparison, the SCF method and the energy-adaptive RGD method require 18 and 54 iterations to converge, respectively. In terms of computing time, all methods perform quite similarly in this experiment. This can also be seen in Table 1, which shows the values of the energy functional, the reached residuals, the number of (outer) iterations, the total number of Hamiltonian evaluations, and the CPU time.
5.2.2 Graphene Lattice
As the second model, we consider a graphene lattice consisting of carbon atoms arranged in 9 hexagons with \(p=67\) electron orbitals. We use a \(32\times \ 55\times 160\) sampling grid for the wave function and get a discrete model of dimension \(n=12279\). Figure 7 presents the evolution of the residuals and errors in the energy. We observe again that both Newton methods converge very fast compared to the energy-adaptive RGD method which needs 64 iterations to achieve the tolerance \(10^{-8}\) for the residual. In contrary, the SCF iteration has difficulties to converge. Also other mixing strategies implemented in KSSOLV do not improve the convergence property of SCF for the graphene model. This behaviour may be explained by a missing spectral gap between the excited and non-excited states. A detailed comparison, including the overall CPU time is part of Table 1. In this experiment (with the particular implementation and used hardware), one can say that the computational complexity of the methods follows the rule of thumb
Overall, this example clearly shows the supremacy of the Newton approach for more challenging examples.
6 Conclusion
In this paper, we have derived Riemannian Newton methods on the infinite-dimensional Stiefel and Grassmann manifolds for Kohn–Sham type energy minimization problems. Starting from an energy functional, we present a unified approach for applications in computational physics (e.g., the Gross–Pitaevskii eigenvalue problem) and computational chemistry (e.g., the Kohn–Sham model). The remarkable gain in computational efficiency of the Riemannian Newton methods compared to the so far more popular methods such as SCF and gradient descent methods is demonstrated by a series of numerical experiments.
Data Availability
Data is available from the corresponding author on reasonable request.
References
Absil, P.A., Mahony, R., Sepulchre, R.: Optimization Algorithms on Matrix Manifolds. Princeton University Press, Princeton, NJ (2008)
Absil, P.A., Malick, J.: Projection-like retractions on matrix manifolds. SIAM J. Optim. 22(1), 135–158 (2012). https://doi.org/10.1137/100802529
Alouges, F., Audouze, C.: Preconditioned gradient flows for nonlinear eigenvalue problems and application to the Hartree-Fock functional. Numer. Methods Partial Differ. Equ. 25(2), 380–400 (2009). https://doi.org/10.1002/num.20347
Altmann, R., Henning, P., Peterseim, D.: Quantitative Anderson localization of Schrödinger eigenstates under disorder potentials. Math. Models Methods Appl. Sci. 30(5), 917–955 (2020). https://doi.org/10.1142/S0218202520500190
Altmann, R., Henning, P., Peterseim, D.: The \(J\)-method for the Gross–Pitaevskii eigenvalue problem. Numer. Math. 148, 575–610 (2021). https://doi.org/10.1007/s00211-021-01216-5
Altmann, R., Henning, P., Peterseim, D.: Localization and delocalization of ground states of Bose–Einstein condensates under disorder. SIAM J. Appl. Math. 82(1), 330–358 (2022). https://doi.org/10.1137/20M1342434
Altmann, R., Peterseim, D.: Localized computation of eigenstates of random Schrödinger operators. SIAM J. Sci. Comput. 41, B1211–B1227 (2019). https://doi.org/10.1137/19M1252594
Altmann, R., Peterseim, D., Stykel, T.: Energy-adaptive Riemannian optimization on the Stiefel manifold. ESAIM Math. Model. Numer. Anal. 56(5), 1629–1653 (2022). https://doi.org/10.1051/m2an/2022036
Bai, Z., Li, R.C., Lu, D.: Sharp estimation of convergence rate for self-consistent field iteration to solve eigenvector-dependent nonlinear eigenvalue problems. SIAM J. Matrix Anal. Appl. 43(1), 301–327 (2022). https://doi.org/10.1137/20M136606X
Bao, W., Chern, I.L., Lim, F.Y.: Efficient and spectrally accurate numerical methods for computing ground and first excited states in Bose–Einstein condensates. J. Comput. Phys. 219(2), 836–854 (2006). https://doi.org/10.1016/j.jcp.2006.04.019
Bao, W., Du, Q.: Computing the ground state solution of Bose–Einstein condensates by a normalized gradient flow. SIAM J. Sci. Comput. 25(5), 1674–1697 (2004). https://doi.org/10.1137/S1064827503422956
Bao, W., Tang, W.: Ground-state solution of Bose–Einstein condensate by directly minimizing the energy functional. J. Comput. Phys. 187(1), 230–254 (2003). https://doi.org/10.1016/S0021-9991(03)00097-4
Caliari, M., Ostermann, A., Rainer, S., Thalhammer, M.: A minimisation approach for computing the ground state of Gross–Pitaevskii systems. J. Comput. Phys. 228(2), 349–360 (2009). https://doi.org/10.1016/j.jcp.2008.09.018
Cancès, E.: SCF algorithms for HF electronic calculations. In: Mathematical Models and Methods for Ab Initio Quantum Chemistry, Lecture Notes in Chemistry, vol. 174, pp. 17–43. Springer, Berlin, Heidelberg (2000). https://doi.org/10.1007/978-3-642-57237-1_2
Cancès, E.: Self-consistent field algorithms for Kohn–Sham models with fractional occupation numbers. J. Chem. Phys. 114(24), 10616–10622 (2001). https://doi.org/10.1063/1.1373430
Cancès, E., Chakir, R., Maday, Y.: Numerical analysis of nonlinear eigenvalue problems. J. Sci. Comput. 45, 90–117 (2010). https://doi.org/10.1007/s10915-010-9358-1
Cancès, E., Chakir, R., Maday, Y.: Numerical analysis of the planewave discretization of some orbital-free and Kohn-Sham models. ESAIM Math. Model. Numer. Anal. 46(2), 341–388 (2012). https://doi.org/10.1051/m2an/2011038
Cancès, E., Kemlin, G., Levitt, A.: Convergence analysis of direct minimization and self-consistent iterations. SIAM J. Matrix Anal. Appl. 42(1), 243–274 (2021). https://doi.org/10.1137/20M1332864
Cancès, E., Le Bris, C.: On the convergence of SCF algorithms for the Hartree–Fock equations. ESAIM Math. Model. Numer. Anal. 34(4), 749–774 (2000). https://doi.org/10.1051/m2an:2000102
Chen, Z., Lu, J., Lu, Y., Zhang, X.: On the convergence of Sobolev gradient flow for the Gross–Pitaevskii eigenvalue problem. SIAM J. Numer. Anal. 62(2), 667–691 (2024). https://doi.org/10.1137/23M1552553
Dai, X., Wang, Q., Zhou, A.: Gradient flow based Kohn–Sham density functional theory model. Multiscale Model. Simul. 18(4), 1621–1663 (2020). https://doi.org/10.1137/19M1276170
Danaila, I., Protas, B.: Computation of ground states of the Gross–Pitaevskii functional via Riemannian optimization. SIAM J. Sci. Comput. 39(6), B1102–B1129 (2017). https://doi.org/10.1137/17M1121974
Dion, C.M., Cancès, E.: Ground state of the time-independent Gross–Pitaevskii equation. Comput. Phys. Comm. 177(10), 787–798 (2007). https://doi.org/10.1016/j.cpc.2007.04.007
Du, C.E., Liu, C.S.: Newton-Noda iteration for computing the ground states of nonlinear Schrödinger equations. SIAM J. Sci. Comput. 44(4), A2370–A2385 (2022). https://doi.org/10.1137/21M1435793
Edelman, A., Arias, T.A., Smith, S.T.: The geometry of algorithms with orthogonality constraints. SIAM J. Matrix Anal. Appl. 20(2), 303–353 (1998). https://doi.org/10.1137/S0895479895290954
García-Ripoll, J.J., Pérez-García, V.M.: Optimizing Schrödinger functionals using Sobolev gradients: applications to quantum mechanics and nonlinear optics. SIAM J. Sci. Comput. 23(4), 1316–1334 (2001). https://doi.org/10.1137/S1064827500377721
Henning, P.: The dependency of spectral gaps on the convergence of the inverse iteration for a nonlinear eigenvector problem. Math. Models Methods Appl. Sci. 33(07), 1517–1544 (2023). https://doi.org/10.1142/S0218202523500343
Henning, P., Jarlebring, E.: The Gross–Pitaevskii equation and eigenvector nonlinearities: numerical methods and algorithms. Preprint (2022)
Henning, P., Peterseim, D.: Sobolev gradient flow for the Gross–itaevskii eigenvalue problem: global convergence and computational efficiency. SIAM J. Numer. Anal. 58(3), 1744–1772 (2020). https://doi.org/10.1137/18M1230463
Henning, P., Yadav, M.: On discrete ground states of rotating Bose–Einstein condensates. Math. Comput. (2024). https://doi.org/10.1090/mcom/3962
Hohenberg, P., Kohn, W.: Inhomogeneous electron gas. Phys. Rev. 136, B864–B871 (1964). https://doi.org/10.1103/PhysRev.136.B864
Hu, G., Wang, T., Zhou, J.: A linearized structure-preserving numerical scheme for a gradient flow model of the Kohn–Sham density functional theory. East Asian J. Appl. Math. 13(2), 299–319 (2023). https://doi.org/10.4208/eajam.2022-134.081022
Jarlebring, E., Kvaal, S., Michiels, W.: An inverse iteration method for eigenvalue problems with eigenvector nonlinearities. SIAM J. Sci. Comput. 36(4), A1978–A2001 (2014). https://doi.org/10.1137/S1064827500366124
Jarlebring, E., Upadhyaya, P.: Implicit algorithms for eigenvector nonlinearities. Numer. Algorithms 90, 301–321 (2022). https://doi.org/10.1007/s11075-021-01189-4
Jiao, S., Zhang, Z., Wu, K., Wan, L., Ma, H., Li, J., Chen, S., Qin, X., Liu, J., Ding, Z., Yang, J., Li, Y., Hu, W., Lin, L., Yang, C.: KSSOLV 2.0: An efficient MATLAB toolbox for solving the Kohn–Sham equations with plane-wave basis set. Comput. Phys. Comm. 279, 108424 (2022). https://doi.org/10.1016/j.cpc.2022.108424
Kazemi, P., Eckart, M.: Minimizing the Gross–Pitaevskii energy functional with the Sobolev gradient—Analytical and numerical results. Int. J. Comput. Methods 7(3), 453–475 (2010). https://doi.org/10.1142/S0219876210002301
Kohn, W., Sham, L.J.: Self-consistent equations including exchange and correlation effects. Phys. Rev. 140, A1133–A1138 (1965). https://doi.org/10.1103/PhysRev.140.A1133
Le Bris, C.: Computational chemistry from the perspective of numerical analysis. Acta Numer. 14, 363–444 (2005). https://doi.org/10.1017/S096249290400025X
Lieb, E.H., Seiringer, R., Yngvason, J.: A rigorous derivation of the Gross–Pitaevskii energy functional for a two-dimensional Bose gas. Commun. Math. Phys. 224(1), 17–31 (2001). https://doi.org/10.1007/s002200100533
Liu, X., Wen, Z., Wang, X., Ulbrich, M., Yuan, Y.: On the analysis of the discretized Kohn–Sham density functional theory. SIAM J. Numer. Anal. 53(4), 1758–1785 (2015). https://doi.org/10.1137/140957962
Perdew, J., Wang, Y.: Accurate and simple analytic representation of the electron-gas correlation energy. Phys. Rev. B 45(23), 13244–13249 (1992). https://doi.org/10.1103/PhysRevB.45.13244
Perdew, J.P., Zunger, A.: Self-interaction correction to density-functional approximations for many-electron systems. Phys. Rev. B 23, 5048–5079 (1981). https://doi.org/10.1103/PhysRevB.23.5048
Pitaevskii, L.P., Stringari, S.: Bose–Einstein Condensation. Oxford University Press, Oxford (2003)
Raza, N., Sial, S., Siddiqi, S.S., Lookman, T.: Energy minimization related to the nonlinear Schrödinger equation. J. Comput. Phys. 228(7), 2572–2577 (2009). https://doi.org/10.1016/j.jcp.2008.12.016
Roothaan, C.C.J.: New developments in molecular orbital theory. Rev. Mod. Phys. 23, 69–89 (1951). https://doi.org/10.1103/RevModPhys.23.69
Sato, H., Aihara, K.: Cholesky QR-based retraction on the generalized Stiefel manifold. Comput. Optim. Appl. 72(2), 293–308 (2019). https://doi.org/10.1007/s10589-018-0046-7
Schneider, R., Rohwedder, T., Neelov, A., Blauert, J.: Direct minimization for calculating invariant subspaces in density functional computations of the electronic structure. J. Comput. Math. 27(2–3), 360–387 (2009)
Sherman, J., Morrison, W.J.: Adjustment of an inverse matrix corresponding to a change in one element of a given matrix. Ann. Math. Statist. 21(1), 124–127 (1950). https://doi.org/10.1214/aoms/1177729893
Teter, M.P., Payne, M.C., Allan, D.C.: Solution of Schrödinger’s equation for large systems. Phys. Rev. B 40, 12255–12263 (1989). https://doi.org/10.1103/PhysRevB.40.12255
Wloka, J.: Partial Differential Equations. Cambridge University Press, Cambridge (1987)
Yang, C., Meza, J.C., Lee, B., Wang, L.W.: KSSOLV—a MATLAB toolbox for solving the Kohn–Sham equations. ACM Trans. Math. Softw. 36(2), 1–35 (2009). https://doi.org/10.1145/1499096.1499099
Yang, C., Meza, J.C., Wang, L.W.: A constrained optimization algorithm for total energy minimization in electronic structure calculation. J. Comput. Phys. 217(2), 709–721 (2006). https://doi.org/10.1016/j.jcp.2006.01.030
Zhang, Z.: Exponential convergence of Sobolev gradient descent for a class of nonlinear eigenproblems. Commun. Math. Sci. 20, 377–403 (2022). https://doi.org/10.4310/CMS.2022.v20.n2.a4
Zhao, Z., Bai, Z.J., Jin, X.Q.: A Riemannian Newton algorithm for nonlinear eigenvalue problems. SIAM J. Matrix Anal. Appl. 36(2), 752–774 (2015). https://doi.org/10.1137/140967994
Funding
Open Access funding enabled and organized by Projekt DEAL. The work of D. Peterseim is part of a project that has received funding from the European Research Council (ERC) under the European Union’s Horizon 2020 research and innovation programme (Grant agreement No. 865751 – RandomMultiScales).
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The authors declare that they have no conflict of interest.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Altmann, R., Peterseim, D. & Stykel, T. Riemannian Newton Methods for Energy Minimization Problems of Kohn–Sham Type. J Sci Comput 101, 6 (2024). https://doi.org/10.1007/s10915-024-02612-3
Received:
Revised:
Accepted:
Published:
DOI: https://doi.org/10.1007/s10915-024-02612-3
Keywords
- Riemannian optimization
- Stiefel manifold
- Grassmann manifold
- Newton method
- Kohn–Sham model
- Gross–Pitaevskii eigenvalue problem