Abstract
In this paper we present a general procedure for designing higher strong order methods for linear Itô stochastic differential equations on matrix Lie groups and illustrate this strategy with two novel schemes that have a strong convergence order of 1.5. Based on the Runge–Kutta–Munthe–Kaas (RKMK) method for ordinary differential equations on Lie groups, we present a stochastic version of this scheme and derive a condition such that the stochastic RKMK has the same strong convergence order as the underlying stochastic Runge–Kutta method. Further, we show how our higher order schemes can be applied in a mechanical engineering as well as in a financial mathematics setting.
Similar content being viewed by others
1 Introduction
In the last decades, more and more interrelations in mechanics and finance have been modeled by stochastic differential equations (SDEs) on Lie groups. A trend can be observed that shows that kinematic models, which were previously expressed by ordinary differential equations (ODEs), are now extended by terms that include stochastic processes in order to include possible stochastic perturbations. Examples can be found in the modeling of rigid bodies such as satellites, vehicles and robots [5, 6, 16, 31]. Furthermore, SDEs on Lie groups are also considered in the estimation of object motion from a sequence of projections [28] and in the representation of the precessional motion of magnetization in a solid [1].
In financial mathematics, the consideration of stochastic processes is essential, and the solution of SDEs has been performed for many years, but usually not on Lie groups. However, the use of Lie groups to solve existing or to create new financial models could be of central importance for dealing with geometric constraints. We are confronted with geometric constraints, e.g. in the form of a positivity constraint on interest rates [14, 25, 30] or a symmetry and positivity constraint on covariance and correlation matrices [22], which are important in, for example, risk management and portfolio optimization.
Despite these diversified applications, the available literature on analysis and numerical methods for SDEs on Lie groups is limited, in contrast to the available literature on ODEs on Lie groups (for example [4, 7, 9, 11, 23, 24]). Furthermore, the available literature on Lie group SDEs mainly concerns Stratonovich SDEs [1, 3, 8, 16, 21, 30]. Readers interested in Itô SDEs on Lie groups will only find the geometric Euler–Maruyama scheme with strong order \(\gamma =1\) appearing in [18, 19, 26] and more recently the existence and convergence proof of the stochastic Magnus expansion in [12]. However, the consideration of Itô SDEs is crucial for its application in finance.
Our contribution to this field of research is a general procedure on how to set up structure-preserving schemes of higher strong order for Itô SDEs on matrix Lie groups. Based on the Magnus expansion we apply Itô–Taylor schemes or stochastic Runge–Kutta (SRK) schemes to solve a corresponding SDE in the Lie algebra. Using a SRK method can be interpreted as a stochastic version of Runge–Kutta–Munthe–Kaas (RKMK) methods and if our considered Itô SDEs were transformed to Stratonovich SDEs, this approach would be equivalent to the stochastic Munthe–Kaas schemes presented in [16]. Nevertheless, a proof of convergence for this method is still missing. Under these circumstances, we derive a condition such that the stochastic RKMK scheme inherits the strong convergence order \(\gamma \) of the SRK method applied in the Lie algebra.
The remainder of the paper is organized as follows. We start with an introduction to matrix Lie groups, their corresponding Lie algebras and the linear Itô matrix SDE which we consider in this geometric setting in Sect. 2. In Sect. 3 we take a closer look on how SDEs on Lie groups can be solved numerically and present our higher strong order methods. Then we provide some numerical and application examples in Sect. 4. A conclusion of our results is given in Sect. 5.
2 SDEs on matrix Lie groups
A Lie group is a differentiable manifold, which is also a group G with a differentiable product that maps \(G\times G\rightarrow G\). Matrix Lie groups are Lie groups, which are also subgroups of GL(n) for \(n\in {\mathbb {N}}\). The tangent space at the identity of a matrix Lie group G is called Lie algebra \({\mathfrak {g}}\). The Lie algebra is closed under forming Lie brackets \([\cdot ,\cdot ]\) (also called commutators) of its elements. For further details on Lie groups and Lie algebras we refer the interested reader to [10].
The matrix exponential, \(\exp (\varOmega )=\sum _{k=0}^{\infty }\frac{1}{k!}\varOmega ^k\), serves as a map from the Lie algebra to the corresponding Lie group, \(\exp :{\mathfrak {g}}\rightarrow G\), and is a local diffeomorphism near \(\varOmega =0_{n\times n}\). Its directional derivative in the direction of an arbitrary matrix \(H\in {\mathfrak {g}}\) is given by
where
(see [9, p. 83]). By \(\mathrm{ad}_{\varOmega }:{\mathfrak {g}}\rightarrow {\mathfrak {g}}\), \(\mathrm{ad}_{\varOmega }(H)=[\varOmega ,H]= \varOmega H - H\varOmega \) we express the adjoint operator which is used iteratively
The inverse of \(d\exp _{-\varOmega }\) is given in the following Lemma [9, p. 84].
Lemma 2.1
(Baker, 1905) If the eigenvalues of the linear operator \(\mathrm{ad}_{\varOmega }\) are different from \(2\ell \pi i\) with \(\ell \in \{\pm 1,\pm 2,\ldots \}\), then \(d\exp _{-\varOmega }\) is invertible. Furthermore, we have for \(\Vert \varOmega \Vert <\pi \) that
where \(B_k\) are the Bernoulli numbers, defined by \(\sum _{k=0}^{\infty }(B_k/k!)x^k=x/(e^x-1)\).
We recall that the first three Bernoulli numbers are given by \(B_0=1\), \(B_1=-\frac{1}{2}\), \(B_2=\frac{1}{6}\) and that \(B_{2m+1}=0\) holds for \(m\in {\mathbb {N}}\).
Let \((\varTheta ,{\mathscr {F}},{\mathbb {P}})\) be a complete probability space and let \(W_t=(W_t^1,W_t^2,\ldots ,W_t^d)\) be a d-dimensional standard Brownian motion w.r.t. a filtration \({\mathscr {F}}_t\) for \(t\ge 0\) which satisfies the usual conditions. On a matrix Lie group G we consider the linear matrix-valued Itô SDE
where \(K_t, V_{1,t},\ldots , V_{d,t} \in {\mathbb {R}}^{n \times n}\) are given coefficient matrices and \(I_{n\times n}\) is the n-dimensional identity matrix. In general, there exists no closed form solution to (2.3). However, a solution can be defined via a Magnus expansion \(Q_t=Q_0\exp (\varOmega _t)\) (see [15, 19, 24]), where \(\varOmega _t{{\in {\mathfrak {g}}}}\) obeys the following matrix SDE
The drift and diffusion coefficients, \(A,\varGamma _i:{\mathfrak {g}}\rightarrow {\mathfrak {g}}\), are given by
with \(C_i:{\mathfrak {g}}\rightarrow {\mathfrak {g}}\),
which can be specified by
for \(i=1,\ldots ,d\). We refer to [19] for the proof.
2.1 The Cayley map as local parameterization
With all these series related to the matrix exponential, the question arises whether there is another mapping \(\psi :{\mathfrak {g}} \rightarrow G\), which is also a local diffeomorphism near 0 but not based on the evaluation of an infinite number of summands. In case of a quadratic Lie group the answer is yes, there is a mapping, namely the Cayley transformation
A quadratic Lie group G is a set of matrices Q that fulfill the equation \(Q^{\top }PQ = P\) for a given constant matrix P. For the derivative of \(\mathrm{cay}(\varOmega )\) we have
The analogue expression to (2.1) is
with the inverse given by
see [9, p. 128].
Using the Cayley map instead of the matrix exponential as a local parameterization, the coefficients of (2.4) are
for \(i=1,\ldots ,d\) (see Appendix for proof).
2.2 Example: SDEs on SO(n)
As an example of a matrix Lie group, we take a closer look at the special orthogonal group
which is a quadratic Lie group, such that the Cayley map is also applicable as a local parameterization. The corresponding Lie algebra consists of skew-symmetric matrices:
Since we are interested in structure preservation we need conditions that tell us when the solution of an SDE on SO(n) is kept on the manifold.
Theorem 2.1
For the solution \(Q_t\) of (2.3), then \(Q_t \in \) SO(n) if and only if the coefficient matrices satisfy \(V_{i,t} \in \mathfrak {so}(n)\) for \(i=1,\ldots ,d\) and \(K_t+K_t^{\top }=\sum _{i=1}^d V_{i,t}^2\).
For the proof of this theorem we refer to [19].
3 Numerical methods for SDEs on Lie groups
Applying standard numerical methods for SDEs directly to the linear matrix-valued Itô SDE (2.3) will result in a drift off, i.e. the numerical approximations do not stay on the manifold. Consequently, one needs to consider special numerical methods that preserve the geometric properties of the Lie group G.
As the Lie algebra \({\mathfrak {g}}\) represents a linear space with Euclidean-like geometry, it appears reasonable to compute the numerical approximations of the matrix SDE (2.4) and to project the solution back onto the Lie group G.
A simple scheme based on the Runge–Kutta–Munthe–Kaas schemes for ODEs [23] that puts the described approach into practice can be found in [18] and is presented in the following algorithm.
Algorithm 3.1
Divide the time interval \([0,T]\) uniformly into J subintervals \([t_j, t_{j+1} ]\), \(j=0,1,\ldots ,J-1\) and define the time step \(\varDelta = t_{j+1} - t_j\). Let \(Q_t=Q_0\psi (\varOmega _t)\) with \(\psi :{\mathfrak {g}}\rightarrow G\) be a local parameterization of the Lie group G. Starting with \(t_0=0\), \(Q_0=I_{n\times n}\) and \(\varOmega _0=0_{n\times n}\) the following steps are repeated over successive intervals \([t_j, t_{j+1}]\) until \(t_{j+1}=T\).
-
1.
Initialization step: Let \(Q_j\) be the approximation of \(Q_t\) at time \(t=t_j\). Similarly, let \(K_j\) and \(V_{i,j}\) be the coefficient matrices \(K_t\) and \(V_{i,t}\) evaluated at \(t=t_j\) for \(i=1,\ldots ,d\).
-
2.
Numerical method step: Let \(\varOmega _\varDelta \) be the exact solution of (2.4) after one time step, i.e. at \(t=t_1=\varDelta \). Compute an approximation \(\varOmega _1\approx \varOmega _{\varDelta }\) by applying a stochastic Itô–Taylor or stochastic Runge–Kutta method to the matrix SDE (2.4).
-
3.
Projection step: Set \(Q_{j+1}=Q_j\, \psi (\varOmega _1)\).
Stochastic Runge–Kutta schemes consist of linear operations which would lead to approximations drifting off the Lie group G if they were applied directly to (2.3). In Algorithm 3.1 these linear operations are all carried out in the vector space \({\mathfrak {g}}\) such that the approximation \(\varOmega _1\) of (2.4) is also an element of the Lie algebra \({\mathfrak {g}}\). This approximation is then mapped to the Lie group G by \(\varOmega _1\mapsto Q_{j+1}=Q_j\psi (\varOmega _1)\), i.e. an approximation of (2.3) is obtained. Thus, all schemes obtained by following Algorithm 3.1 produce approximations in G and the structure of the Lie group is preserved.
The order of convergence of these Lie group structure-preserving schemes clearly depends on the numerical method used in the second step of the algorithm since \(\varOmega _1\mapsto Q_{j+1}=Q_j\psi (\varOmega _1)\) is a smooth mapping, i.e. the order on G is as high as the order of the scheme used for the corresponding equation in \({\mathfrak {g}}\). In order to analyze the accuracy of our geometric numerical methods we recall that an approximating process \(X_t^{\varDelta }\) is said to converge in a strong sense with order \(\gamma > 0\) to the Itô process \(X_t\) if there exists a finite constant K and a \(\varDelta '>0\) such that
for any time discretization with maximum step size \(\varDelta \in (0,\varDelta ')\) [13].
3.1 Geometric schemes of strong order 1
Using the Euler–Maruyama scheme in the numerical method step of Algorithm 3.1 results in
where \(\varDelta W \sim {\mathscr {N}}(0,\varDelta )\). Note that \(C_i(\varOmega _0)=0_{n\times n}\), \(i=1,\ldots ,d\) for both mappings \(\psi =\exp \) (see (2.7)) and \(\psi =\mathrm{cay}\) (see (2.10)) which is why we neglect this coefficient from here on.
Since this scheme (3.2) preserves the geometry of the Lie group G it was called the geometric Euler–Maruyama scheme [19]. It can be specified according to the mapping.
For \(\psi =\exp \), we get
where inserting \(\varOmega _0=0_{n \times n}\) is equivalent to truncating the infinite series (2.2) after the first summand, right before any dependence on \(\varOmega \) appears.
Using \(\psi =\mathrm{cay}\) instead, we obtain
In both cases we see that the diffusion term is only dependent on time and not on the solution itself. This is called additive noise [13] and it is the reason why these schemes have strong order \(\gamma =1\) instead of \(\gamma =0.5\) as expected for the traditional Euler–Maruyama method. A general proof of the geometric Euler–Maruyama method converging with strong order \(\gamma =1\) can be found in [26].
3.2 Geometric schemes of higher order
A higher strong order than \(\gamma = 1\) can be achieved by applying the strong Itô–Taylor approximation of order \(\gamma =1.5\) (see [13, p. 351]) in the second step of Algorithm 3.1. By doing so, we obtain for \(d=1\)
As seen in the previous section \(\varOmega _1\) depends on \(t_j\) since
Representing the double integral \(\int _{\tau _{\ell }}^{\tau _{\ell +1}} \int _{\tau _{\ell }}^{s_2} dW_{s_1}ds_2\), the random variable \(\varDelta Z\) is normally distributed with mean \({\mathbb {E}}[\varDelta Z]=0\), variance \({\mathbb {E}}\bigl [(\varDelta Z)^2\bigr ]=\frac{1}{3}\varDelta ^3\) and covariance \({\mathbb {E}}[\varDelta Z \varDelta W]=\frac{1}{2}\varDelta ^2\). We consider the matrix derivatives as directional derivatives, e.g.
which we then evaluate at \(\varOmega _0\). The computation of the needed matrix derivatives for \(\psi =\exp \) and \(\psi =\mathrm{cay}\) is provided in the Appendix.
In the case where \(d>1\) one could apply the corresponding Itô–Taylor scheme for \(d>1\) (see [13, p. 353]) to (2.4) followed by a projection step.
A strong order of \(\gamma =1.5\) can also be achieved by applying a stochastic Runge–Kutta method of that order to the SDE (2.4). By using the stochastic Runge–Kutta scheme of order \(\gamma =1.5\) of Rößler [27], we can avoid computing the derivatives in (3.4) and we obtain for \(d=1\)
with the stage values
The exploitation of stochastic Runge–Kutta methods gives us the benefit of a derivative-free scheme. However, using the mapping \(\psi =\exp \) raises the question of how large the truncation index q must be chosen in the truncated approximation for (2.2),
in order to maintain a strong order of \(\gamma =1.5\). More generally, a condition is needed that connects the truncation index q with the aimed strong convergence order \(\gamma \).
Inspired by [9, Theorem IV.8.5.] for Runge–Kutta–Munthe–Kaas methods to solve deterministic matrix ODEs we formulate the following theorem.
Theorem 3.2
Consider Algorithm 3.1 with \(\psi =\exp \). Let the applied stochastic Runge–Kutta method in the second step of Algorithm 3.1 be of strong order \(\gamma \). If the truncation index q in (3.6) satisfies \(q\ge 2\gamma -2\), then the method of Algorithm 3.1 is of strong order \(\gamma \).
Proof
According to the definition of strong convergence (3.1) we have to show that
where \(\varOmega _{\varDelta }\) is the exact solution of (2.4) with \(\psi =\exp \) at \(t=\varDelta \), \(\varOmega _1\) is the numerical approximation obtained in the second step of Algorithm 3.1 and K is a finite constant.
Let \(\varOmega _{\varDelta }^q\) be the exact solution of the truncated version of (2.4) with \(\psi =\exp \) at \(t=\varDelta \), namely
Our proof is divided into six steps.
Step 1: Numerical error
We consider the Frobenius norm of matrices in the Lie algebra \({\mathfrak {g}}\) and estimate the error in the \(L^1\)-norm by the \(L^2\)-norm. Then, we use the Minkowski inequality by introducing \(\varOmega _{\varDelta }^q\),
We are left with the modelling error, which corresponds to the first summand, and the numerical error, the second summand. The numerical error can be estimated by
as we assume that we are applying a SRK method of strong order \(\gamma \).
In other words, it remains to be shown that
holds for the modelling error.
Step 2: Itô isometry
Inserting the integral equation of (2.4) and its truncated version, we get
where we also used the Minkowski inequality, the Itô isometry and the properties of a matrix norm. Now, the summands in the last line differ only in the input matrix of the adjoint operator.
Step 3: Adjoint operator
We estimate the Frobenius norm of the adjoint operator of \(V_{i,s}\) for a fixed \(s\in [0,\varDelta ]\) and keep in mind that analogous estimates hold for the adjoint operator of \(K_s-\frac{1}{2}\sum _{i=1}^d V_{i,s}^2\). Since the Frobenius norm is submultiplicative, we have
As a direct consequence, it holds
which can also be shown via induction. Inserting this result in the expected value considered in the last line of the previous step, we get
Step 4: Estimate for the remainder
It is known that the Bernoulli numbers are implicitly defined by \(\sum _{k=0}^{\infty }(B_k/k!)x^k=x/(e^x-1)\). Inserting the absolute values of the Bernoulli numbers instead, then
(see [11, p. 48]). Let \(f:I\rightarrow {\mathbb {R}}\), \(x \mapsto \frac{x}{2}\left( 1+\cot \left( \frac{x}{2}\right) \right) +2\) with \(I=\{x \in {\mathbb {R}}: \frac{x}{2\pi }\not \in {\mathbb {Z}}\}\). Applying Taylor’s theorem to the function f at the point 0 gives
where we consider the Lagrange form of the remainder for some real number \(\xi \) between 0 and x.
Setting \(x=2\Vert \varOmega _s\Vert \) and recalling that the expression (2.2) only converges for \(\Vert \varOmega \Vert <\pi \), we now consider \(\left. f\right| _{\tilde{I}}:\tilde{I}\rightarrow {\mathbb {R}}\), \(x \mapsto \frac{x}{2}\left( 1+\cot \left( \frac{x}{2}\right) \right) +2\) with \(\tilde{I}=\{x \in {\mathbb {R}}: |x|<2\pi \}\). The restriction of f to \(\tilde{I}\) is bounded, in particular there exists an upper bound \(M_q\) such that \(|\left. f\right| _{\tilde{I}}^{(q+1)}(\xi )|\le M_q\) for all \(\xi \) between 0 and x. Moreover, the following estimate for the remainder holds
Using this estimate in the expected value of the last line of the previous step results in
Step 5: Itô–Taylor expansion
The goal of this step is to find an estimate for \({\mathbb {E}}\left[ \Vert \varOmega _s\Vert ^{2q+2}\right] \). For this purpose, we examine the following Itô–Taylor expansion
where \(C_1\) is a finite constant, for details see [13, Proposition 5.9.1]. This leads to the estimate
Step 6: Overall estimate
Gathering the results of the previous steps and inserting a Taylor expansion for \(V_{i,s}\) where \(C_2<\infty \) gives
Thus,
Analogously, one can show that
which concludes the proof.\(\square \)
Remark 3.1
-
1.
Assuming that multiple Wiener integrals can be simulated efficiently, a general procedure for designing methods of strong order \(\gamma \) for linear Itô SDEs on matrix Lie groups can be achieved by applying a SRK scheme of order \(\gamma \) in the second step of Algorithm 3.1 while evaluating the sum (3.6) involved in the coefficients of equation (2.4) only up to index \(2\gamma -2\). We refer the reader to [17] and the references therein for more details on the approximation of iterated stochastic integrals.
-
2.
Theorem 3.2 is in accordance with our results of Sect. 3.1, where the geometric Euler–Maruyama scheme (3.3) can be interpreted as a stochastic RKMK method with \(\gamma =1\) and \(q=0\).
-
3.
Due to the definition of the Cayley map as a finite product of matrices, there is no modelling error and therefore, no such theorem is needed if \(\psi =\mathrm{cay}\) is chosen as the local parameterization in Algorithm 3.1.
4 Numerical examples
In the following we provide numerical examples that illustrate the effectiveness of the proposed geometric methods, firstly, by simulating the strong convergence order of the proposed schemes and secondly, by showing the Lie group structure preservation of our methods.
For checking the convergence order, we set \(G=\) SO(3) and \({\mathfrak {g}}=\mathfrak {so}(3)\). In order to ensure the conditions of Theorem 2.1 we have used the set up of matrices \(K_t\) and \(V_t\) proposed by Muniz et al. [22] for \(d=1\). Specifically, we chose the time-dependent functions
to compute a skew-symmetric matrix \(V_t\) as a linear combination,
where \(G_i\), \(i=1,2,3\) are the following generators of the Lie algebra \(\mathfrak {so}(3)\),
Note that the functions \(f_i\), \(i=1,2,3\) can be chosen arbitrarily. We then set the matrix \(K_t\) as the lower triangular matrix of \(V_t^2\) where the diagonal entries of \(K_t\) are 0.5 times the diagonal entries of \(V_t^2\) such that \(K_t + K_t^{\top } = V_t^2\).
We simulated \(M=1000\) different paths of two independent realizations of a standard normally distributed random variable, \(U_1, U_2 \sim {\mathscr {N}}(0,1)\). Then, the random variables used in the numerical method step in Algorithm 3.1 were simulated as \(\widehat{\varDelta W} = U_1\sqrt{\varDelta }\) and \(\widehat{\varDelta Z} = \frac{1}{2}\varDelta (\widehat{\varDelta W} + U_2\sqrt{\frac{\varDelta }{3}})\). The absolute error as defined in (3.1) was estimated by using the Frobenius norm at \(t_j=T\), i.e. by
where the approximations \(Q_{T,i}^{\varDelta }\) were obtained by using Algorithm 3.1 with step sizes \(\varDelta = 2^{-14}, 2^{-13}, 2^{-12}, 2^{-11}, 2^{-10}, 2^{-9}\) and for the reference solution \(Q_{T,i}^{\text {ref}}\) we used the same method with \(\psi =\mathrm{cay}\) and step size \(\varDelta =2^{-16}\), respectively.
A log–log plot of the estimation of the absolute error against the step sizes can be viewed in Fig. 1. It indicates the strong order of convergence claimed in the sections above for the geometric Euler–Maruyama scheme (3.2), the geometric version of the Itô–Taylor scheme (3.4) and the geometric stochastic Runge–Kutta scheme (3.5).
Examples from financial mathematics and multibody system dynamics verify that the structure-preserving methods derived above can be applied in practice.
In the first example we apply our methods of strong order \(\gamma =1.5\) to an SDE on SO(2) in the context of stochastic correlation modelling. The second example shows how our methods can be used in the modeling of rigid bodies, such as satellites. Although, we have restricted our research for this paper to considering only linear SDEs on Lie groups, the second example shows that our methods can also be applied to nonlinear SDEs on e.g. SO(3).
4.1 A stochastic correlation model
Let us assume that a risk manager retrieves from the middle office’s reporting system an initial value of the correlation between two assets and a density function of the considered correlation. Moreover, we assume that the risk manager was given the task to generate correlation matrices that not only approximate the given density function but also respect the stochastic behaviour of correlations.
This problem can be solved by the stochastic correlation model presented in [22]. The main ideas of the approach are outlined below.
For this example we consider historical prices of the S&P 500 index and the Euro/US-Dollar exchange rate and compute moving correlations with a window size of 30 days to obtain correlations from January 03, 2005 to January 06, 2006 (see Fig. 2).
The corresponding initial correlation matrix calculated from this data and imputed to the risk manager is
Furthermore, we estimate a density function from the historical data using kernel smoothing functions, which is also plotted in Fig. 3. For more details on the density estimation see [2].
As a first step, we focus on covariance matrices \(P_t\), \(t\ge 0\). The authors of [29] utilised the principal axis theorem and defined the covariance flow
where \(P_0\) is the initial covariance matrix computed based on \(R_0^{\mathrm{hist}}\) and \(Q_t\) is an orthogonal matrix which without loss of generality can be assumed to have determinant +1, i.e. \(Q_t \in \mathrm {SO}(2)\). Following the approach in [22] the matrix \(Q_t\) is now assumed to be driven by the SDE (2.3) which can be solved by using Algorithm 3.1. With the resulting matrices approximations of \(P_t\) can be computed with (4.1), which can then be transformed to corresponding correlation matrices
with \(\varSigma _t=\bigl (\mathrm{diag}(P_t)\bigr )^{\frac{1}{2}}\).
At last, a density function is estimated from this correlation flow and the free parameters involved are calibrated such that the density function matches the density function from the historical data, see [22] for details.
We executed this procedure using the geometric Itô–Taylor scheme (3.4) with \(\psi =\mathrm{cay}\) (gIT) and the geometric Rößler scheme (3.5) with \(\psi =\exp \) and truncation index \(q=1\) (gSRK) in the second step of Algorithm 3.1, respectively. The results are plotted in Fig. 4, which shows that both density functions approximate the density function of the historical data quite well.
4.2 The stochastic rigid body problem
Consider a free rigid body, whose centre of mass is at the origin. Let the vector \(y=(y_1,y_2,y_3)^{\top }\) represent the angular momentum in the body frame and \(I_1\), \(I_2\) and \(I_3\) be the principal moments of inertia [20]. Then the motion of this free rigid body is described by the Euler equations
We suppose that the rigid body is perturbed by a Wiener process \(W_t\) and compute a matrix K(y) such that the dynamics are kept on the manifold, i.e. we compute K(y) from the condition \(K(y) + K^{\top }(y) = V^2(y)\). Consequently, we regard the Itô SDE
where the solution evolves on the unit sphere if the initial value \(y_0\) satisfies \(|y_0|=1\). Note that stochastic versions of the rigid body problem have already been considered in [16] and [30] but as Stratonovich SDEs.
Since the solution of (4.2) can also be written as \(y=Qy_0\) where \(Q\in \mathrm {SO}(3)\), we focus on the nonlinear matrix SDE
The coefficients of the corresponding SDE in the Lie algebra (2.4) are
Now, SDE (4.2) can be solved by applying Algorithm 3.1 to the SDE (4.3). Note that we deal with right multiplication of the solution Q on the right hand side of (4.3) instead of left multiplication as in (2.3). As a consequence, the sign of the index of the operator \(d\psi ^{-1}\) is changed in (4.4) and the solution of the Projection step in Algorithm 3.1 should be \(Q_{j+1}=\psi (\varOmega _{j+1})Q_j\). We refer to [23] for more details on this matter.
In Fig. 5 we simulated 200 steps of the trajectory of (4.2) with a step size of \(\varDelta =0.03\) by using Algorithm 3.1 with the initial values \(y_0=(\sin (1.1),0,\cos (1.1))^{\top }\) and the moments of inertia \(I_1=2\), \(I_2=1\) and \(I_3=2/3\). For the numerical method step of Algorithm 3.1 we used the Euler–Maruyama scheme with \(\psi =\mathrm{cay}\). Emphasizing the structure-preserving character of Algorithm 3.1 we also plotted a sample path of the traditional Euler–Maruyama scheme applied directly to (4.2), whose trajectory clearly fails to stay on the manifold. This phenomenon can also be viewed in Fig. 6 where we visualize the distance of the approximate solutions from the manifold.
5 Conclusion
We have presented stochastic Lie group methods for linear Itô SDEs on matrix Lie groups that have a higher strong convergence order than the known geometric Euler–Maruyama scheme. Based on RKMK methods for ODEs on Lie groups, we have proven a condition on the truncation index of the inverse of \(d\exp _{-\varOmega }(H)\) such that the stochastic RKMK method inherits the convergence order of the underlying SRK. This allows us to construct even higher order schemes than the here presented methods of strong order \(\gamma =1.5\) if one were able to approximate multiple Wiener integrals efficiently. Additionally, we have shown examples for the application of our methods in mechanical engineering and in financial mathematics.
Our methods require further investigations for the application to nonlinear Itô SDEs on matrix Lie groups, which we consider as future work. Moreover, we have restricted our research for this paper to the strong convergence order. In future research, an investigation on the weak convergence order of stochastic Lie group methods will also be conducted.
Change history
27 January 2022
A Correction to this paper has been published: https://doi.org/10.1007/s10543-022-00911-5
References
Ableidinger, M., Buckwar, E.: Weak stochastic Runge–Kutta Munthe–Kaas methods for finite spin ensembles. Appl. Numer. Math. 118, 50–63 (2017)
Bowman, A.W., Azzalini, A.: Applied Smoothing Techniques for Data Analysis. Oxford University Press, New York (1997)
Burrage, K., Burrage, P.M.: High strong order methods for non-commutative stochastic ordinary differential equation systems and the Magnus formula. Phys. D 133, 34–48 (1999)
Celledoni, E., Marthinsen, H., Owren, B.: An introduction to Lie group integrators—basics, new developments and applications. J. Comput. Phys. 257(Part B), 1040–1061 (2014)
Chirikjian, G.S.: Stochastic Models, Information Theory, and Lie Groups, vol. 1: Classical Results and Geometric Methods. Springer, Boston (2009)
Chirikjian, G.S.: Stochastic Models, Information Theory, and Lie Groups, vol. 2: Analytic Methods and Modern Applications. Springer, Boston (2011)
Crouch, P.E., Grossman, R.: Numerical integration of ordinary differential equations on manifolds. J. Nonlinear Sci. 3, 1–33 (1993)
Elworthy, K.D.: Stochastic Differential Equations on Manifolds, vol. 70. Cambridge University Press, Cambridge (1982)
Hairer, E., Lubich, C., Wanner, G.: Geometric Numerical Integration. Springer Series in Computational Mathematics, vol. 31, 2nd edn. Springer, Berlin (2006)
Hall, B.C.: Lie Groups, Lie Algebras, and Representations. Graduate Texts in Mathematics, vol. 222, 2nd edn. Springer, Heidelberg (2015)
Iserles, A., Munthe-Kaas, H.Z., Nørsett, S.P., Zanna, A.: Lie group methods. Acta Numer. 9, 215–365 (2005)
Kamm, K., Pagliarani S., Pascucci, A.: On the stochastic Magnus expansion and its application to SPDEs. arXiv preprint arXiv:2001.01098 (2020)
Kloeden, P.E., Platen, E.: Numerical Solution of Stochastic Differential Equations. Springer, Berlin (1992)
Lim, N., Privault, N.: Analytic bond pricing for short rate dynamics evolving on matrix Lie groups. Quant. Finance 16(1), 119–129 (2016)
Magnus, W.: On the exponential solution of differential equations for a linear operator. Commun. Pure Appl. Math. 7, 649–673 (1954)
Malham, S.J.A., Wiese, A.: Stochastic Lie group integrators. SIAM J. Sci. Comput. 30(2), 597–617 (2008)
Malham, S.J.A., Wiese, A.: Efficient almost-exact Lévy area sampling. Stat. Probab. Lett. 88, 50–55 (2014)
Marjanovic, G., Piggott, M.J., Solo, V.: A simple approach to numerical methods for stochastic differential equations in Lie groups. In: Proceedings of the 54th IEEE Conference on Decision and Control, IEEE, Osaka, Japan, pp. 7143–7150 (2015)
Marjanovic, G., Solo, V.: Numerical methods for stochastic differential equations in matrix Lie groups made simple. IEEE Trans. Autom. Control 63(12), 4035–4050 (2018)
Marsden, J.E., Ratiu, T.S.: Introduction to Mechanics and Symmetry, 2nd edn. Springer, New York (1999)
Misawa, T.: A Lie algebraic approach to numerical integration of stochastic differential equations. SIAM J. Sci. Comput. 23(3), 866–890 (2001)
Muniz, M., Ehrhardt, M., Günther, M.: Approximating correlation matrices using stochastic Lie group methods. Mathematics 9(1), 94 (2021)
Munthe-Kaas, H.: Runge–Kutta methods on Lie groups. BIT Numer. Math. 38(1), 92–111 (1998)
Munthe-Kaas, H.: High order Runge–Kutta methods on manifolds. Appl. Numer. Math. 29, 115–127 (1999)
Park, F.C., Chun, C.M., Han, C.W., Webber, N.: Interest rate models on Lie groups. Quant. Finance 11(4), 559–572 (2010)
Piggott, M.J., Solo, V.: Geometric Euler–Maruyama schemes for stochastic differential equations in SO(n) and SE(n). SIAM J. Numer. Anal. 54(4), 2490–2516 (2016)
Rößler, A.: Explicit order 1.5 schemes for the strong approximation of Itô stochastic differential equations. PAMM 5(1), 817–818 (2005)
Soatto, S., Perona, P., Frezza R., Picci, G.: Motion estimation via dynamic vision. In: Proceedings of the 33rd IEEE Conference on Decision and Control, vol.4, pp. 3253–3258. IEEE, Piscataway, NJ (1994)
Teng, L., Wu, X., Günther, M., Ehrhardt, M.: A new methodology to create valid time-dependent correlation matrices via isospectral flows. ESAIM Math. Model. Numer. Anal. 54(2), 361–371 (2020)
Wang, Z., Ma, Q., Yao, Z., Ding, X.: The magnus expansion for stochastic differential equations. J. Nonlinear Sci. 30, 419–447 (2020)
Zuyev, A., Vasylieva, I.: Partial stabilization of stochastic systems with application to rotating rigid bodies. IFAC-PapersOnLine 52(16), 162–167 (2019)
Acknowledgements
The authors would like to thank Martin Friesen (Dublin City University) for the in-depth discussions that improved the content of this paper. The work of the authors was partially supported by the bilateral German-Slovakian Project MATTHIAS—Modelling and Approximation Tools and Techniques for Hamilton-Jacobi-Bellman equations in finance and Innovative Approach to their Solution, financed by DAAD and the Slovakian Ministry of Education. Further the authors acknowledge partial support from the bilateral German-Portuguese Project FRACTAL – FRActional models and CompuTationAL Finance financed by DAAD and the CRUP - Conselho de Reitores das Universidades Portuguesas.
Open Access
This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.
Funding
Open Access funding enabled and organized by Projekt DEAL.
Author information
Authors and Affiliations
Corresponding author
Additional information
Communicated by Antonella Zanna Munthe-Kaas.
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
The original online version of this article was revised: unfortunately, the original article was published without ORCID numbers. This has since been corrected.
Appendices
Proofs
Theorem A.1
The solution of (2.3) can be written as \(Q_t=Q_0 \,\mathrm{cay}(\varOmega _t)\), where \(\varOmega _t\) obeys the SDE (2.4) with coefficients given by
with
Proof
Let \(f:{\mathfrak {g}}\rightarrow G\), \(f(\varOmega _t)=Q_0\mathrm{cay}(\varOmega _t)\) such that \(Q_t=f(\varOmega _t)\). Due to the Itô rules, i.e. \(dt\cdot dt=dt\cdot dW_t=dW_t\cdot dt=0\) and \(dW_t\cdot dW_t=dt\), \(dQ_t=d(f(\varOmega _t))\) is given exactly by the first two terms of the Taylor expansion
where we have used the fact that
We use (2.8) to specify the first part of the drift coefficient
Analogously, we have \(\bigl (\frac{d}{d\varOmega _t}f(\varOmega _t)\bigr )\,\varGamma _i(\varOmega _t)= Q_t d\mathrm{cay}_{-\varOmega _t}\bigl (\varGamma _i(\varOmega _t)\bigr )\). For the second derivative we obtain
where \(C_i(\varOmega _t)=\Bigl (\frac{d}{d\varOmega _t}d\mathrm{cay}_{-\varOmega _t}\bigl (\varGamma _i(\varOmega _t)\bigr )\Bigr )\varGamma _i(\varOmega _t)\). Comparing these results with SDE (2.3) we get
and thus
\(\square \)
Lemma A.1
The coefficient (A.1) is given by
Proof
Inserting for both H and \(\tilde{H}\) the diffusion coefficient \(\varGamma _i(\varOmega _t)=d\mathrm{cay}_{-\varOmega _t}^{-1}(V_{i,t}) = \frac{1}{2}(I+\varOmega _t)V_{i,t}(I-\varOmega _t)\) we get
\(\square \)
Matrix derivatives
In this section we provide the matrix derivatives that we used in the geometric version of the Itô–Taylor scheme of strong order \(\gamma =1.5\) (see (3.4)).
1.1 Derivatives for \(\psi =\mathrm{cay}\)
Computing the derivative of (2.9) in the direction of an arbitrary matrix \(\tilde{H}\), we get
Subsequently, the second directional derivative is
Inserting \(H=V\) and \(\tilde{H}=\varGamma (\varOmega )=d\mathrm{cay}_{-\varOmega }^{-1}(V)\), we obtain
Similar expressions are obtained for \(\left( \frac{d}{d\varOmega }A(\varOmega )\right) \varGamma (\varOmega )\), \(\left( \frac{d}{d\varOmega }A(\varOmega )\right) A(\varOmega )\) and \(\left( \frac{d}{d\varOmega }\varGamma (\varOmega )\right) A(\varOmega )\) by inserting \(H=K-\frac{1}{2}V^2\) and \(\tilde{H}=\varGamma (\varOmega )\), \(H=K-\frac{1}{2}V^2\) and \(\tilde{H}=A(\varOmega )\) and \(H=V\) and \(\tilde{H}=A(\varOmega )\) in (B.1), respectively.
Proceed accordingly to compute the second derivatives \(\left( \frac{d^2}{d\varOmega ^2}A(\varOmega )\right) \varGamma ^2(\varOmega )\) and \(\left( \frac{d^2}{d\varOmega ^2}\varGamma (\varOmega )\right) \varGamma ^2(\varOmega )\).
1.2 Derivatives for \(\psi =\exp \)
In the following we present derivatives of (2.1) up to \(k=4\), i.e. of
Computing the directional derivative we get
Whereas the second directional derivative is given by
Note that evaluating the derivatives at \(\varOmega _0=0_{n\times n}\) causes many summands to become zero, which makes computing higher summands (\(k>4\)) unnecessary.
The needed derivatives for the geometric Itô–Taylor scheme (3.4) can be computed from the formulas above by inserting correspondingly into H and \(\tilde{H}\) (see B.1 for instructions).
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Muniz, M., Ehrhardt, M., Günther, M. et al. Higher strong order methods for linear Itô SDEs on matrix Lie groups. Bit Numer Math 62, 1095–1119 (2022). https://doi.org/10.1007/s10543-021-00905-9
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10543-021-00905-9