Abstract
According to a version of Donsker’s theorem, geodesic random walks on Riemannian manifolds converge to the respective Brownian motion. From a computational perspective, however, evaluating geodesics can be quite costly. We therefore introduce approximate geodesic random walks based on the concept of retractions. We show that these approximate walks converge in distribution to the correct Brownian motion as long as the geodesic equation is approximated up to second order. As a result, we obtain an efficient algorithm for sampling Brownian motion on compact Riemannian manifolds.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
Probabilistic models in continuous time with geometric constraints lead to Brownian motion and stochastic differential equations (SDEs) on Riemannian manifolds (M, g). The theory of Brownian motion and SDEs on manifolds has been extensively studied (see, e.g., [16, 20])—and many results from the Euclidean setting can be generalized to manifoldvalued SDEs. One example is Donsker’s theorem. Here, one considers a geodesic random walk for which the update step at the current location \(x_i\) is obtained by uniformly sampling a unit tangent vector \({\bar{v}} \sim {\text {Unif}}( S_{x_i})\) with \(S_x:= \{u\in T_{x}M: \Vert u\Vert _g = 1\}\) and following a geodesic path by setting \(x_{i+1}:= {\text {Exp}}_{x_i} (\varepsilon v)\) for some small enough parameter \(\varepsilon \), where \(v:=\sqrt{m} {\bar{v}}\), m denotes the dimension of the manifold, and \({\text {Exp}}_{x}:T_xM \rightarrow M\) is the exponential map. We will discuss the reason for scaling by \(\sqrt{m}\) below. Jørgensen proved that on complete Riemannian manifolds that satisfy mild assumptions (which in particular hold for compact manifolds), the scaling limit of such a geodesic random walk is the Brownian motion on (M, g), see [21]. This result has recently been generalized to the setting of Finsler manifolds, see [26].
In principle, it is thus possible to simulate Brownian motion on Riemannian manifolds by computing geodesic random walks—analogously to how one would use Donsker’s theorem in the Euclidean case. Unfortunately, this procedure is not computationally efficient on manifolds since the solution of a nonlinear secondorder geodesic ODE has to be computed in every step. With an eye on applications, we are interested in efficient and simple, yet convergent, methods in order to approximate Brownian motion on Riemannian manifolds.
We introduce a class of algorithms that yield approximate geodesic random walks by making use of retractions that approximate the exponential map and have been introduced in the context of numerical optimization on manifolds, see [1,2,3, 8]. We show that retractions which approximate the geodesic ODE up to second order converge to Brownian motion in the Skorokhod topology, see Sect. 3. We consider two prevalent scenarios, where random walks based on such secondorder retractions can be computed efficiently: the case of parameterized manifolds and the case of implicitly defined manifolds, see Sect. 2. Our approach generalizes the setting of approximate geodesic walks on manifolds with positive sectional curvatures presented in [27]. Moreover, our method works for arbitrary dimensions, and its cost only depends on the evaluation of the respective retraction.
A retraction in our sense can be understood as both, an approximation of the geodesic ODE, or, equivalently, as a 2jet in the sense of [5], which allows for treating coordinatefree SDEs. Armstrong and King developed a 2jet scheme for simulating SDEs on manifolds in [6], which uses an unbounded stepsize in each iteration. Different from our retractionbased approach, which is based on bounded stepsizes, the case of unbounded stepsizes hampers efficiency (for the case of local charts) or might even prohibit an implementation (for the case of implicitly defined submanifolds when using a stepandproject approach). Moreover, our random walk method is based on a Donskertype theorem, while their approach improves Euler–Maruyama approximations—resulting in a different type of convergence in terms of the respective topologies.
Outside the setting of 2jets, one typically considers SDEs on embedded manifolds \(M \subset {\mathbb {R}}^n\) and first solves them in ambient Euclidean space, followed by projecting back to M using Lagrange multipliers, see, e.g., [7, 25]. Projectionbased approaches, however, do not immediately extend to the case of parameterized manifolds. Within our unified framework based on retractions, we cover both, the case of embedded manifolds and the case of parameterized manifolds, see Sects. 2.1 and 2.2. In particular, our approach leads to an efficient and convergent numerical treatment of Brownian motion on Riemannian manifolds (including drift—see Remark 2.4).
A different scenario where geodesic random walks are commonly used is the problem of sampling the volume measure of polytopes (defined as the convex hull resulting from certain linear constraints). Originally, Euclidean random walk methods had been considered in this context, see [30]. Although the Euclidean perspective is natural in this situation, small stepsizes are required in order to reach points near the boundary, which hampers efficiency in higher dimensions. To circumvent this issue, Lee and Vempala [24] introduced a Riemannian metric based on the Hessian of a logbarrier—to the effect of rescaling space as the walk approaches the boundary. In their implementation, they used collocation schemes for solving the requisite ODEs in order to simulate geodesic random walks. We leave for future research the question whether such approaches can be reformulated in terms of retractions.
Furthermore, the problem of sampling on manifolds is commonly addressed by Markov chain Monte Carlo (MCMC) methods, see, e.g., [10,11,12, 31]. Different from our focus, MCMC methods are typically concerned with obtaining samples from a given probability distribution on M. Our algorithm, however, yields approximate sample paths of a Brownian motion on a manifold. We are not aware of any results concerned with proving that an MCMC method samples the correct dynamics of a given SDE. Although not the main focus of our exposition, we show that our method also correctly recovers the stationary measure of a geodesic random walk in the limit of stepsize tending to zero, see Theorem 3.2.
2 RetractionBased Random Walks
Throughout this exposition, we consider mdimensional compact and orientable Riemannian manifolds (M, g) without boundary. We use retractions to approximate the exponential map on such manifolds.
Definition 2.1
Let \({\text {Ret}}: TM\xrightarrow {} M\) be a smooth map, and denote the restriction to the tangent space \(T_xM\) by \({\text {Ret}}_x\) for any \(x\in M\). \({\text {Ret}}\) is a retraction if the following two conditions are satisfied for all \(x\in M\) and all \(v\in T_xM\):

1.
\({\text {Ret}}_x(0) = x\), where 0 is the zero element in \(T_xM\) and

2.
\(\frac{\textrm{d}}{\textrm{d}\tau }{\text {Ret}}_x (\tau v)\big _{\tau = 0} = v\) (where we identify \(T_0T_xM \simeq T_xM\)).
A retraction is a secondorder retraction if it additionally satisfies that for all \(x \in M\) and for all \(v\in T_xM\) one has that
where \(\frac{\textrm{D}}{\textrm{d}\tau }(\frac{\textrm{d}}{\textrm{d}\tau } \gamma (\tau ))\) denotes covariant differentiation of the tangent vector field \({{\dot{\gamma }}} (\tau ) = \frac{\textrm{d}}{\textrm{d}\tau } \gamma (\tau )\) along the curve \(\gamma \) (following standard notation, see, e.g., [15]). If M is a submanifold of a Euclidean space, Eq. (1) is equivalent to
where \({\mathcal {N}}_xM\) denotes the normal bundle of M. Consequently, for the case of submanifolds, one has
for all \(x\in M\) and \(v\in T_xM\) as \(\tau \rightarrow 0\).
Clearly, the exponential map is itself a retraction. The main benefit of retractions is, however, that they can serve as computationally efficient approximations of the exponential map. Using retractions yields Algorithm 1 for simulating random walks on a Riemannian manifold.
Notice that the uniform sampling in Step 3 of the above algorithm is performed with respect to the Riemannian metric restricted to the unit sphere in the attendant tangent space. We provide details for this step in Sects. 2.1 and 2.2. Notice furthermore that \(\varepsilon \) takes the role of a spatial stepsize, while the total physical time simulated by Algorithm 1 is given by \(\varepsilon ^2N\); see also the time rescaling in Theorem 2.2. Step 4 of the algorithm is necessary since uniform sampling on the unit sphere in a space of dimension m results in a covariance matrix \(\frac{1}{m}{\text {Id}}\), while convergence to Brownian motion requires the identity as covariance matrix. The normalizing factor \(\sqrt{m}\) in Step 4 precisely ensures the latter. Indeed, without the renormalization in Step 4 the random walk constructed in Algorithm 1 converges to a timechanged Brownian motion.
The next theorem shows convergence of the retractionbased random walk resulting from Algorithm 1 and provides an explicit expression for the generator of the resulting limit process:
Theorem 2.2
Consider the sequence of random variables \((X_i^\varepsilon )_{i\in {\mathbb {N}}}\) constructed in Algorithm 1 (with \(N=\infty \)). The continuoustime process \(X^\varepsilon := \big (X_{\lfloor \varepsilon ^{2}t\rfloor }^\varepsilon \big )_{t\ge 0}\) converges in distribution to the stochastic process with generator
in the Skorokhod topology (see Sect. 3 for details). Here, \(S_x\) denotes the unit sphere in \(T_x M\), \(\omega _m\) is the volume of this sphere, and \(\Delta _g\) is the Laplace–Beltrami operator on (M, g).
We defer the proof to Sect. 3. Notice that Theorem 2.2 immediately implies a necessary and sufficient condition for a retractionbased random walk to converge to Brownian motion on a Riemannian manifold (M, g):
Corollary 2.3
Using the same notation as in Theorem 2.2, the continuoustime process \(X^\varepsilon \) converges in distribution to the Brownian motion on M if and only if the Laplacian (also known as the tension) of \({\text {Ret}}_x: T_xM\xrightarrow {} M\) vanishes at \(0 \in T_{x}M\) for every \(x\in M\). In particular, this holds for secondorder retractions.
Proof
By definition, the Laplacian of \({\text {Ret}}_x: T_xM\xrightarrow {} M\) is defined as
where \(\nabla d {\text {Ret}}_x\) is the Hessian of the \(C^\infty \)map \({\text {Ret}}_x\). Clearly, the Laplacian of \({\text {Ret}}_x\) vanishes at \(0 \in T_{x}M\) if and only if
Applying Theorem 2.2 proves the first claim. If \({\text {Ret}}_x\) is a secondorder retraction, the second claim immediately follows since \(\frac{\textrm{D}}{\textrm{d}\tau } \frac{\textrm{d}}{\textrm{d}\tau } {\text {Ret}}_x (\tau v)\big _{\tau = 0}=0\) for all \(v\in T_xM\).
\(\square \)
Thus, random walks generated by secondorder retractions converge to the Brownian motion on (M, g).
Remark 2.4
Drift can seamlessly be incorporated into our approach. Indeed, when modeling drift by a vector field X, then Step 4 of Algorithm 1 needs to be altered according to \(v:=\sqrt{m}{\bar{v}} + \varepsilon X_{x_{i1}^\varepsilon }\). The stochastic process resulting from this altered version of Algorithm 1 converges to the process with generator \(Lf+df(X)\), where Lf is the generator given by (2). The proof of this fact follows the proof of Theorem 2.2 with obvious modifications.
In the next two subsections, we consider two concrete examples of secondorder retractions that are computationally efficient.
2.1 Retractions Based on Local Parameterizations
Our first example of computationally efficient retractions is based on local parameterizations of compact mdimensional Riemannian manifolds (M, g). Consider an atlas consisting of finitely many charts \(\{(U_i, \phi _i)\}_{i\in I}\), where every \(U_i \subset {\mathbb {R}}^m\) is open, \(\phi _i:U_i \rightarrow M\) is a diffeomorphism onto its image in M, and where the relatively open sets \(\{\phi _i(U_i)\}\) cover M. In concrete applications, such parameterizations often arise naturally, e.g., for isometrically embedded manifolds \(M\subset {\mathbb {R}}^{n}\).
For a given \(x\in M\), \(v\in T_xM\), and \((U_i,\phi _i)\) with \(x\in \phi _i(U_i)\), let \({\tilde{x}} = \phi _i^{1} (x)\) and \({\tilde{v}} = (d\phi _i^{1})_x v\). We assume that for some fixed \(\varepsilon > 0\) and any \(x\in M\), there exists a chart \((U_k, \phi _k)\) with \({{\tilde{x}}}\in U_{k}\) such that
for all \({{\tilde{v}}} =(d\phi _k^{1})_x (\varepsilon \sqrt{m}\,{\bar{v}})\) with \({\bar{v}}\in S_x\), where \((\Gamma _{ij})\) denotes the Christoffel symbols of the respective chart. This assumption is readily satisfied for any compact M, provided that \(\varepsilon \) is chosen small enough. In order to find such a sufficiently small \(\varepsilon \) in Algorithm 1, we propose to restart the algorithm with stepsize \(\frac{\varepsilon }{2}\) if the condition is violated. Then, we define the retraction
Notice that Christoffel symbols can be computed numerically. For lowdimensional manifolds, symbolic differentiation accelerates their computation.
Lemma 2.5
The retraction defined by Eq. (3) is of second order.
Proof
Let \((\gamma (\tau ))_{\tau \ge 0}\) denote the geodesic satisfying \(\gamma (0) = x\) and \({\dot{\gamma }}(0) = v\). In the parameter domain, \(\gamma \) satisfies the ordinary differential equation
The Taylor expansion of \(\gamma \) reads
Since \({\text {Exp}}_x (\tau v) = \gamma (\tau )\), this shows that
as \(\tau \rightarrow 0\), which proves that \({\text {pRet}}\) is indeed a secondorder retraction. \(\square \)
Another aspect of parameterizationbased approaches concerns the computation of tangent vectors \({\tilde{v}}\) in the parameter domain that are uniformly distributed with respect to the Riemannian metric g. An efficient procedure for this task is presented in Algorithm 2.
Notice that Eq. (5) is well defined since \(w_k\ne 0\) a.s. for all \(1\le k \le m\) and that z is a sample of a uniformly distributed random variable on \({\mathbb {S}}^{m1}\). It is straightforward to verify the following claim:
Lemma 2.6
The vector \({\tilde{v}}\) constructed in Algorithm 2 is a unit tangent vector that is uniformly sampled with respect to the Riemannian metric \((g_{ij})\).
In summary, using \({\tilde{v}}\) as computed in Algorithm 2, setting \({{\bar{v}}} = {\tilde{v}}\) in Step 3 of Algorithm 1, and using the retraction \({\text {pRet}}\) from (3), provides an efficient method for computing random walks for parameterized manifolds. See Table 1 for a comparison between geodesic random walks and retractionbased random walks, showing that the latter are significantly faster than the former. Notice that the attendant algorithm might require to update the respective local chart during the simulated random walk. Results of a simulated random walk on a 2torus (using a single chart with periodic boundary conditions) are presented in Fig. 1.
2.2 ProjectionBased Retractions
Our second example of computationally efficient retractions is based on a stepandproject method for compact manifolds that are given as zerolevel sets of smooth functions. More specifically, we consider the setting where
is a smooth function, \(k<n\), and the mdimensional manifold in question, with \(m=(nk)\), is given by
We assume that 0 is a regular value of f, i.e., that the differential df has full rank along M. In this case, a retraction can be defined via
where \(x\in M\), \(v\in T_xM\), and \( \pi _M(y) = \mathop {\mathrm {arg\,min}}\limits _{p\in M} \Vert py\Vert \) denotes the (closest point) projection to M. Notice that \( \pi _M(y)\) is well defined as long as the distance d(y, M) is less then the socalled reach of M. The reach of M is the distance of M to its medial axis, which in turn is defined as the set of those points in \({\mathbb {R}}^n\) that do not have a unique closest point on M. Smoothly embedded compact manifolds always have positive reach.
The following result is important in our setup. For a proof, see, e.g., Theorem 4.10 in [3].
Lemma 2.7
\(\pi \)\({\text {Ret}}\) is a secondorder retraction.
It remains to specify the computation of a random unit direction \(v\in T_xM\) as well as the implementation of the projection step. As for the former, a uniformly sampled unit tangent direction at some \(x\in M\) can be computed by first sampling a randomly distributed unit vector \(u \in {\mathbb {S}}^{n1}\) in ambient space, and denoting its projection to \(\textrm{im} (df_{x})\) by \({\tilde{u}}\). Then,
yields the requisite uniformly distributed unit tangent vector. Finally, for computing the closest point projection of some given point \(y\in {\mathbb {R}}^n\) to M, consider the Lagrangian
where \(\lambda \in {\mathbb {R}}^{k}\) denotes the (vectorvalued) Lagrange multiplier. Then, Newton’s method offers an efficient implementation for solving the Euler–Lagrange equations resulting from \({\mathcal {L}}_y\). In our implementation, we let Newton’s method run until the threshold \(f(z)/\Vert df_z\Vert < \varepsilon ^3\) is satisfied. Notice that efficiency results from the fact that the point \(y=x+\tau v\) is close to M for \(x\in M\) and \(\tau \) small enough.
In summary, using \({\bar{v}}\) as computed in (7) for Step 3 of Algorithm 1, together with the retraction \(\pi \)\({\text {Ret}}\) from (6), provides an efficient method for computing random walks for implicitly given submanifolds. Results of a simulated random walk on the 2torus are presented in Fig. 2.
Remark 2.8
Our projectionbased retraction is an instance of retractions induced by socalled normal foliations. The latter are defined as foliations in a local neighborhood of \(M= f^{1}(0)\) whose leaves have dimension k and intersect M orthogonally. As shown in [32], retractions induced by normal foliations are always of second order and therefore provide alternative approaches for sampling Brownian motion on implicit manifolds. Examples of retractions induced by normal foliations include the gradient flow along \(\nabla f\) and Newton retractions based on the update step \(\delta z = (d f)^{\dagger }(z) f(z)\), where \((d f)^{\dagger }\) is the Moore–Penrose pseudoinverse of the Jacobian df.
2.3 Stationary Measure
For secondorder retractions, we show below that the stationary measure of the process in Algorithm 1 converges to the stationary measure of the Brownian motion, which is the normalized Riemannian volume form, see Theorem 3.2. Nonetheless, the empirical approximation \(\frac{1}{N}\sum _{i=1}^N \delta _{x_i^\varepsilon }\) of the Riemannian volume form using random walks—e.g., in order to obtain results as in Figs. 1 and 2—requires the stepsize \(\varepsilon \) and the number of steps N in Algorithm 1 to be chosen appropriately. An indicator at which time the empirical approximation is close to the stationary measure is the socalled \(\delta \)cover time, i.e., the time needed to hit any \(\delta \)ball. Hence, as a rough heuristic choice of \(\varepsilon \) and N, one may utilize [13, 14]. For a Brownian motion on a 2dimensional manifold M, the \(\delta \)cover time \(C_\delta \) of M satisfies [14, Theorem 1.3]
where A denotes the Riemannian area of M. Based on the observation that a secondorder retraction converges to Brownian motion in the limit of vanishing stepsize \(\varepsilon \), and using that the physical time of the walk resulting from Algorithm 1 is given by \(\varepsilon ^2 N\), one may set \(\varepsilon ^2 N= C_\delta \). If \(\delta \) is small enough, one may then choose N according to
For dimension \(m>2\), the corresponding result in [13], Theorem 1.1 yields the heuristic choice
where
V(M) denotes the volume of M, and \(\omega _m\) the volume of the unit sphere in \({\mathbb {R}}^m\).
Such heuristics, however, do not quantify the error between the empirical measure obtained by simulating a random walk and the Riemannian volume measure. In the setting of Euclidean random walks \((Y_n)_{n\in {\mathbb {N}}}\) with invariant distribution \(\nu \), it has been shown in [23] that under some conditions on Y the expectation of the 1Wasserstein distance between the empirical and the invariant measure asymptotically decreases according to the law
as \(N\rightarrow \infty \) with constants \(\alpha ,\beta ,\gamma > 0\) depending on the dimension m and \(\vartheta \in (0,1)\) depending on the process Y. Our simulations, see Fig. 3, followed by a simple fit to the model given by the righthand side of Eq. (8) suggest that a similar qualitative behavior might be present for retractionbased random walks on Riemannian manifolds. We leave a deeper analysis for future research.
3 Convergence
This section is devoted to the proof of Theorem 2.2, which yields convergence in the Skorokhod topology of the process constructed in Algorithm 1 as the stepsize \(\varepsilon \rightarrow 0\). The Skorokhod space \({\mathcal {D}}_M[0,\infty )\) is the space of all càdlàg, i.e., rightcontinuous functions with lefthand limits from \([0,\infty )\) to M, equipped with a metric turning \({\mathcal {D}}_M[0,\infty )\) into a complete and separable metric space, and hence a Polish space (see, e.g., [17], Chapter 3.5). The topology induced by this metric is the Skorokhod topology, which generalizes the topology of local uniform convergence for continuous functions \([0,\infty )\rightarrow M\) in a natural way. For local uniform convergence, we have uniform convergence on any compact time interval \(K\subset [0,\infty )\). In the Skorokhod topology, functions with discontinuities converge if the times and magnitudes of the discontinuities converge (in addition to local uniform convergence of the continuous parts of the functions). Thus, \({\mathcal {D}}_M[0,\infty )\) is a natural and—due to its properties—also a convenient space for convergence of discontinuous stochastic processes.
With reference to the notation used in Theorem 2.2, notice that the process \(X^\varepsilon = \big (X_{\lfloor \varepsilon ^{2}t\rfloor }^\varepsilon \big )_{t\ge 0}\) has deterministic jump times and is therefore not Markov as a process in continuous time. In order to apply the theory of Markov processes, let \(\left( \eta _t\right) _{t\ge 0}\) be a Poisson process with rate \(\varepsilon ^{2}\) and define the pseudocontinuous process \(Z^\varepsilon = (Z_t^\varepsilon )_{t\ge 0}\) by \(Z_t^\varepsilon := X_{\eta _t}^\varepsilon \) for any \(t\ge 0\), which is Markov since the Poisson process is Markov. Then, the convergence of \(Z^\varepsilon \) implies convergence of \(X^\varepsilon \) to the same limit by the law of large numbers (see [22], Theorem 19.28). First we restate Theorem 2.2, using an equivalent formulation based on the process \(Z^\varepsilon \):
Theorem 3.1
Consider the sequence of random variables \((X_i^\varepsilon )_{i\in {\mathbb {N}}}\) obtained by the construction from Algorithm 1. The process \(Z^\varepsilon = \left( X^\varepsilon _{\eta _t}\right) _{t\ge 0}\) converges in distribution in the Skorokhod topology to the Ldiffusion Z, i.e., the stochastic process with generator L defined in (2). This means that for any continuous and bounded function \(f:{\mathcal {D}}_M[0,\infty ) \rightarrow {\mathbb {R}}\) we have
Certainly one cannot, in general, deduce convergence of stationary measures from timelocal convergence statements about paths. However, standard arguments from the theory of Feller processes indeed allow to do so. This yields the following statement regarding stationary measures.
Theorem 3.2
Let \(\mu _\varepsilon \) be the stationary measure of the retractionbased random walk with stepsize \(\varepsilon \). Then, the weak limit of \(\left( \mu _\varepsilon \right) _{\varepsilon >0}\) as \(\varepsilon \rightarrow 0\) is the stationary measure of the Ldiffusion, with generator L defined in (2). In the case of secondorder retractions, the limit is the Riemannian volume measure.
Remark 3.3
Notice that due to the approximation of the exponential map, the process \(Z^\varepsilon \) is not in general reversible and the generator of \(Z^\varepsilon \) is not in general selfadjoint. This is relevant since the stationary measure corresponds to the kernel of the adjoint of the generator.
Our proof of Theorems 3.1 and 3.2 (presented below) hinges on convergence of generators that describe the infinitesimal evolution of Markov processes, see, e.g., [17, Chapter 4]. We show that the generator of \(Z^\varepsilon \) converges to the generator L defined in (2). The generator of the process \(Z^\varepsilon \) is spelled out in Lemma 3.4, and the convergence of this generator to L is treated in Lemma 3.5.
The following result is standard for transition kernels \(U^\varepsilon \) on compact state spaces, see, e.g., [17, Chapter 8.3]:
Lemma 3.4
The process \((Z_t^\varepsilon )_{t\ge 0}\) is Feller, and its generator is given by
for all \(f\in C(M)\), the continuous realvalued functions on M. Here, \((U^\varepsilon f)(x) = \frac{1}{\omega _m}\int _{S_x} f({\text {Ret}}_x (\varepsilon \sqrt{m}\, {\bar{v}})) \text {d}{\bar{v}}\) and \(\omega _m\) is the volume of the unit \((m1)\)sphere.
Notice that Lemma 3.4 additionally states that the process \(Z^\varepsilon \) satisfies the Feller property, i.e., that its semigroup is a contraction semigroup that also fulfills a rightcontinuity property. Therefore, in particular, the process \(Z^\varepsilon \) is Markov.
In the following, \(C^{(2,\alpha )}(M)\), \(0<\alpha \le 1\) denotes the space of two times differentiable functions on M with \(\alpha \)Hölder continuous second derivative.
Lemma 3.5
For any \(f\in C^{(2,\alpha )}(M)\) with \(0 < \alpha \le 1\),
as \(\varepsilon \rightarrow 0\), where \(L^\varepsilon \) is the operator from Lemma 3.4 and L as defined in (2).
Proof
Let \(P^\varepsilon \) be the transition kernel of the geodesic random walk, i.e.,
For \(x\in M\) and \(v\in T_x M\), define
and
Then, since \(Lf =\frac{1}{2} \Delta _g f+ {{\tilde{G}}}f\),
Proposition 2.3 in [21] proves that the second summand on the righthand side converges to 0 as \(\varepsilon \rightarrow 0\). Let \({{\tilde{\varepsilon }}}:= \sqrt{m}\,\varepsilon \). Then,
Now, consider the Taylor expansions of the functions \(f\circ {\text {Ret}}_x\) and \(f\circ {\text {Exp}}_x\) in \({{\tilde{\varepsilon }}}\) at 0. To this end, notice that for any \(x\in M\) and \(v\in T_xM\), one has
where \(v_\tau = \frac{\textrm{d}}{\textrm{d}\tau } {\text {Ret}}_x (\tau v)\). Hence, we obtain that
where \({\text {Hess}} f = \nabla d f\) is the Riemannian Hessian of f. Using that \(v_0 = v\), the desired Taylor expansion takes the form
where \(\xi _1\in [0,{{\tilde{\varepsilon }}}]\). Defining \(u_\tau = \frac{\textrm{d}}{\textrm{d}\tau } {\text {Exp}}_{x}(\tau v)\), we similarly obtain that
with \(\xi _2 \in [0,{{\tilde{\varepsilon }}}]\) and
where for the last equality we used the fact that \( \frac{\textrm{D}}{\textrm{d}\tau } u_{\tau } = 0. \) Since \(R_{{\text {Ret}}}^{x,v} (0) = R_{{\text {Exp}}}^{x,v} (0) + 2(Gf)(x,v)\), we obtain that
Since \({\text {Hess}} f\) is Höldercontinuous by assumption and \({\text {Exp}}\) and \({\text {Ret}}\) are smooth, both \(R_{{\text {Exp}}}\) and \(R_{{\text {Ret}}}\) are Höldercontinuous as products and sums of Höldercontinuous functions. Therefore, there exists a constant \(C>0\) such that one has the uniform bound
Hence,
which proves the lemma. \(\square \)
Using the above results, we can finish the proof of Theorem 3.1.
Proof of Theorem 3.1
By the proof of Theorem 2.1 and the remark on page 38 in [21], the space \(C^{(2,\alpha )}(M)\) is a core for the differential operator L. For a sequence of Feller processes \((Z^\varepsilon )_{\varepsilon > 0}\) with generators \(L^\varepsilon , \varepsilon >0\) and another Feller process Z with generator L and core D, the following are equivalent [22, Theorem 19.25]:

(i)
If \(f\in D\), there exists a sequence of \(f_\varepsilon \in {\text {Dom}}(L^\varepsilon )\) such that \(\Vert f_\varepsilon f\Vert _\infty \rightarrow 0\) and \(\left\Vert L^\varepsilon f_\varepsilon Lf\right\Vert _\infty \rightarrow 0\) as \(\varepsilon \rightarrow 0\).

(ii)
If the initial conditions satisfy \(Z^\varepsilon _0 \rightarrow Z_0\) in distribution as \(\varepsilon \rightarrow 0\) in M, then \(Z^\varepsilon \rightarrow Z\) as \(\varepsilon \rightarrow 0\) in distribution in the Skorokhod space \({\mathcal {D}}_M([0,\infty ))\).
In our case, \(D=C^{(2,\alpha )}(M)\) and \(f_\varepsilon = f\in D\) satisfy (i) due to Lemma 3.5. This concludes the proof of Theorem 3.1 since the Ldiffusion is also a Feller process on a compact manifold.\(\square \)
From an analytical point of view, it is perhaps not surprising that convergence of generators, see Lemma 3.5, implies convergence of stationary measures as stated in Theorem 3.2. For the proof, we here follow standard arguments of the theory of Feller processes.
Proof of Theorem 3.2
The family of (unique) stationary measures \(\left( \mu _\varepsilon \right) _{\varepsilon > 0}\) is a tight family of measures since all measures are supported on the same compact manifold M. By Prokhorov’s theorem, which provides equivalence of tightness and relative compactness, any subsequence of the family \((\mu _{\varepsilon _n})_{n\in {\mathbb {N}}}\) with \(\varepsilon _n \rightarrow 0\) as \(n\rightarrow \infty \) has a convergent subsubsequence. But the uniform convergence of the generators that we showed in Lemma 3.5, see also (i) in the proof of Theorem 3.1, is also equivalent to the convergence of the associated semigroups with respect to the supremum norm in space, see again [22, Theorem 19.25]. This implies [17, Chapter 4, Theorem 9.10] that all subsequential limits must be the unique stationary measure of the Ldiffusion, and therefore, all subsubsequences converge to the same measure. A standard subsubsequence argument then proves the theorem. \(\square \)
4 Retractions on SubRiemannian Manifolds
A possible extension of our work concerns random walks in subRiemannian geometries. Indeed, subRiemannian structures can be used to model lowdimensional noise that lives in a higher dimensional state space [19]. A subRiemannian structure on a smooth manifold M consists of a smoothly varying positive definite quadratic form on a subbundle E of the tangent bundle TM. Similar to the Riemannian setting, socalled normal subRiemannian geodesics arise from a firstorder ODE on the cotangent bundle [28]. Such geodesics are uniquely determined by an initial position \(x\in M\) and a 1form \(\alpha \in T^{*}_{x}M\) and can therefore be approximated by a secondorder Taylor expansion of the respective ODE on the cotangent bundle. Such an approximation would algorithmically resemble our retractionbased approach, when attempting to efficiently simulate a subRiemannian geodesic random walk. However, in general, there exists no canonical choice for sampling a requisite 1form \(\alpha \) in every step in order to construct a subRiemannian geodesic random walk (or an approximation thereof). Indeed, one requires an additional choice of a measure for such sampling, and different choices may lead to different limiting processes, see [4, 9]. Notice that once such a choice has been made, one can seamlessly adapt our algorithm to the subRiemannian setting. We leave for future work a respective convergence analysis. Moreover, there might exist canonical choices of such measures for special subRiemannian structures—e.g., when considering the frame bundle of a smooth manifold M in order to model anisotropic Brownian motion and general diffusion processes [18, 29].
References
Absil, P.A., Mahony, R., Sepulchre, R.: Optimization Algorithms on Matrix Manifolds. Princeton University Press, Princeton, New Jersey (2008)
Absil, P.A., Malick, J.: Projectionlike Retractions on Matrix Manifolds. SIAM J. Optim. 22(1), 135–158 (2012). https://doi.org/10.1137/100802529
Adler, R.L., Dedieu, J.P., Margulies, J.Y., Martens, M., Shub, M.: Newton’s method on Riemannian manifolds and a geometric model for the human spine. IMA J. Numer. Anal. 22, 359–390 (2002). https://doi.org/10.1093/imanum/22.3.359
Agrachev, A., Boscain, U., Neel, R., Rizzi, L.: Intrinsic random walks in Riemannian and subRiemannian geometry via volume sampling. ESAIM Control Optim. Calc. Var. 24(3), 1075–1105 (2018). https://doi.org/10.1051/cocv/2017037
Armstrong, J., Brigo, D.: Intrinsic stochastic differential equations as jets. Proc. R. Soc. A 474, 20170559 (2018). https://doi.org/10.1098/rspa.2017.0559
Armstrong, J., King, T.: Curved schemes for stochastic differential equations on, or near, manifolds. Proc. R. Soc. A 478, 20210785 (2022). https://doi.org/10.1098/rspa.2021.0785
Averina, T.A., Rybakov, K.A.: A Modification of Numerical Methods for Stochastic Differential Equations with First Integrals. Numer. Analys. Appl. 12, 203–218 (2019). https://doi.org/10.1134/S1995423919030017
Bonnabel, S.: Stochastic gradient descent on Riemannian manifolds. IEEE Trans. Automat. Contr. 58(9), 2217–2229 (2013). https://doi.org/10.1109/TAC.2013.2254619
Boscain, U., Neel, R., Rizzi, L.: Intrinsic random walks and subLaplacians in subRiemannian geometry. Adv. Math. 314, 124–184 (2017). https://doi.org/10.1016/j.aim.2017.04.024
Brubaker, M., Salzmann, M., Urtasun, R.: A Family of MCMC Methods on Implicitly Defined Manifolds. In: Lawrence, N.D., Girolami, M. (eds.) Proceedings of the Fifteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 22, pp. 161–172 (2012)
Byrne, S., Girolami, M.: Geodesic Monte Carlo on Embedded Manifolds. Scand. J. Stat. 40(4), 825–845 (2013). https://doi.org/10.1111/sjos.12036
Cantarella, J., Shonkwiler, C.: The Symplectic Geometry of Closed Equilateral Random Walks in 3space. Ann. Appl. Probab. 26(1), 549–596 (2016). https://doi.org/10.1214/15aap1100
Dembo, A., Peres, Y., Rosen, J.: Brownian Motion on Compact Manifolds: Cover Time and Late Points. Electron. J. Probab. 8, 1–14 (2003). https://doi.org/10.1214/EJP.v8139
Dembo, A., Peres, Y., Rosen, J., Zeitouni, O.: Cover times for Brownian motion and random walks in two dimensions. Ann. Math. 160, 433–464 (2004). https://doi.org/10.4007/annals.2004.160.433
do Carmo, M.P.: Riemannian Geometry. Birkhäuser, Boston (1992)
Emery, M.: Stochastic Calculus in Manifolds. Springer, Berlin Heidelberg (1989)
Ethier, S.N., Kurtz, T.G.: Markov Processes. John Wiley & Sons, New York (1986)
Grong, E., Sommer, S.: Most Probable Paths for Anisotropic Brownian Motions on Manifolds. Found. Comput. Math. (2022). https://doi.org/10.1007/s10208022095944
Habermann, K.: Geometry of subRiemannian diffusion processes. PhD thesis, University of Cambridge (2017)
Hsu, E.P.: Stochastic Analysis on Manifolds. Graduate Studies in Mathematics, vol. 38. American Mathematical Society, Providence, RI (2002)
Jørgensen, E.: The Central Limit Problem for Geodesic Random Walks. Z. Wahrscheinlichkeitstheorie verw. Gebiete 32, 1–64 (1975)
Kallenberg, O.: Foundations of Modern Probability. Springer, New York (2002)
Kloeckner, B.R.: Empirical measures: regularity is a countercurse to dimensionality. ESAIM: Prob. Stat. 24, 408–434 (2020). https://doi.org/10.1051/ps/2019025
Lee, Y.T., Vempala, S.: Geodesic Walks in Polytopes. SIAM J. Comput. 51(2), 400–488 (2022). https://doi.org/10.1137/17M1145999
Leimkuhler, B., Matthews, C.: Molecular Dynamics. Springer, Cham (2015)
Ma, T., Matveev, V.S., Pavlyukevich, I.: Geodesic Random Walks, Diffusion Processes and Brownian Motion on Finsler Manifolds. J. Geom. Anal. 31, 12446–12484 (2021). https://doi.org/10.1007/s1222002100723z
Mangoubi, O., Smith, A.: Rapid mixing of geodesic walks on manifolds with positive curvature. Ann. Appl. Probab. 28(4), 2501–2543 (2018). https://doi.org/10.1214/17AAP1365
Montgomery, R.: A Tour of Subriemannian Geometries, Their Geodesics and Applications. Mathematical Surveys and Monographs, vol. 91. American Mathematical Society, Providence, RI (2002)
Sommer, S., Svane, A.M.: Modelling anisotropic covariance using stochastic development and subRiemannian frame bundle geometry. J. Geom. Mech. 9(3), 391–410 (2017). https://doi.org/10.3934/jgm.2017015
Vempala, S.: Geometric random walks: a survey. In: Goodman, J.E., Pach, J., Welzl, E. (eds.) Combinatorial and Computational Geometry. Mathematical Sciences Research Institute Publications, vol. 52, pp. 573–612 (2005)
Yanush, V., Kropotov, D.: Hamiltonian MonteCarlo for Orthogonal Matrices (2019) arXiv:1901.08045 [stat.ML]
Zhang, R.: Newton retraction as approximate geodesics on submanifolds (2020) arXiv:2006.14751 [math.NA]
Acknowledgements
We acknowledge support by the Collaborative Research Center 1456 funded by the Deutsche Forschungsgemeinschaft (DFG, German Research Foundation). We thank Pit Neumann for pointing out an error in a previous version of the manuscript and the anonymous referees for improving the quality of our manuscript.
Funding
Open Access funding enabled and organized by Projekt DEAL.
Author information
Authors and Affiliations
Corresponding author
Additional information
Communicated by CarolaBibiane Schönlieb.
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Schwarz, S., Herrmann, M., Sturm, A. et al. Efficient Random Walks on Riemannian Manifolds. Found Comput Math (2023). https://doi.org/10.1007/s10208023096356
Received:
Revised:
Accepted:
Published:
DOI: https://doi.org/10.1007/s10208023096356