Abstract
We prove a functional limit theorem for vector-valued functionals of the fractional Ornstein–Uhlenbeck process, providing the foundation for the fluctuation theory of slow/fast systems driven by both long- and short-range-dependent noise. The limit process has both Gaussian and non-Gaussian components. The theorem holds for any \(L^2\) functions, whereas for functions with stronger integrability properties the convergence is shown to hold in the Hölder topology, the rough topology for processes in \(C^{\frac{1}{2}+}\). This leads to a ‘rough creation’ / ‘rough homogenization’ theorem, by which we mean the weak convergence of a family of random smooth curves to a non-Markovian random process with non-differentiable sample paths. In particular, we obtain effective dynamics for the second-order problem and for the kinetic fractional Brownian motion model.
1 Introduction
The functional limit theorem we study here lays the foundation for the fluctuation problem for a slow/fast system with the fast variable given by a family of non-strong mixing stochastic processes, this will be discussed in [19], see [18] for the preliminary version. A pivot theorem for obtaining effective dynamics for the slow moving particles in a fast turbulent environment are scaling limit theorems for the convergence of the following functionals
with weak convergence in \({\mathcal {C}}^\alpha ([0,T],{\mathbf {R}})\), where T is some finite fixed time horizon, \(\alpha (\varepsilon )\) a suitable scaling and \(G_k : {\mathbf {R}}\rightarrow {\mathbf {R}}\). If \(y_t^\varepsilon =y_{\frac{t}{\varepsilon }}\) and \(y_t\) is a strong mixing process, \(\alpha (\varepsilon )=\frac{1}{ \sqrt{\varepsilon }}\) and the limit is a Markov process, for details see, e.g. the book [23] and the references therein. For stochastic processes whose auto-correlation function does not decay sufficiently fast at infinity there is no reason to have the \(\sqrt{\varepsilon }\) scaling or to obtain a diffusive limit. Furthermore, the scaling limit and the limit function may depend on the individual functions \(G_k\).
In this article, we take \(y_t^\varepsilon \) to be the stationary and rescaled fractional Ornstein–Uhlenbeck process with Hurst parameter H, which, for \(H> \frac{1}{2}\), exhibits long-range dependence (LRD) and is not strong-mixing. Our interest for long range dependent / non-strong mixing noise comes the time series data of the river Nile. In a study of water flows of the Nile river, Hurst and his colleagues [20] observed long-range time dependence and found that the time dependence varies proportionally to \(t^H\) where \(H\sim 0.73\), by contrast, Brownian motions and stable processes have independent increments. Fractional Brownian motions (fBM) were then proposed by Benoit Mandelbrot and John Van Ness [28] for modelling the Hurst phenomenon. A fBM is a continuous mean zero Gaussian process with stationary increments and covariance \({\mathbf {E}}(B_t-B_s)^2=|t-s|^{2H}\), it is self-similar with similarity exponent H, and distinguished by the Gaussian property and stationary, but dependent increments. See, e.g. [26] and the reference therein for stochastic calculus for fBM’s.
Self-similar processes appeared also in mathematically rigorous descriptions of critical phenomena and in renormalisation theory. In [38], Sinai constructed non-Gaussian self-similar fields; while Dobrushin [12] studied self-similar fields subordinated to self-similar Gaussian fields (multiple Wiener integrals). Those self-similar stochastic processes with stationary increments are a particular interesting class. When normalized to begin at 0, to have mean 0 and to have variance 1 at \(t=1\), they necessarily have the covariance \(\frac{1}{2}(t^{2H}+s^{2H}-|t-s|^{2H})\). Those of Gaussian variety are fBMs. Hermite processes are non-Gaussian self-similar processes with stationary increments and the above-mentioned covariance. They appeared as scaling limits of functionals of long-range dependent Gaussian processes, see [36]. Jona-Lasinio was also concerned with the construction of a systematic theory of limit distributions for sums of ‘strongly dependent’ random variables for which the classical central limit theorems do not hold, [22], see also the book [13].
Let us first consider the convergence of one single component, the scalar case. The scaling constant \(\alpha _k(\varepsilon )\) depends on the function \(G_k\), and is a reflection of the self-similarity exponents of the limiting process. If \(\alpha _k(\varepsilon )=\frac{1}{\sqrt{\varepsilon }}\), the limit of \(X^{k, \varepsilon }\) is a Wiener process and the functional central limit theorem is expected to hold. Let \(\mu \) denote the centred and normalized Gaussian measure on \({\mathbf {R}}\) and \(m_k\) denote the Hermite rank of a centred function \(G_k \in L^2(\mu )\), which is the smallest nonzero term in its chaos expansion. Let \((H,m)\mapsto H^*(m)\) denote the function given by (1.3), which decreases with m. Then, the relevant scaling constants are given as below:
See Lemma 2.4 for a preliminary computation indicating the scales. In the past, the limit theorems for the non-Gaussian limits had been called non-central limit theorems, we use the terminology ‘functional central limit theorems’ for all cases.
The intuition for this comes from its counterpart for sequences. If \(Y_n\) is a mean zero, stationary, and strong mixing sequence, such that \(\sigma _n^2={\mathbf {E}}(\sum _{i=1}^n Y_i)^2\rightarrow \infty \), \( {\mathbf {E}}(\sum _{i=1}^n Y_i)^4=O(\sigma _n^4)\), then \( \frac{1}{\sigma _n} \sum _{i=1}^n Y_i {\longrightarrow } N(0,1)\). If \(Y_n\) is not strong mixing, this CLT may fail. Indeed, if \(X_n\) is a stationary mean zero variance 1 Gaussian sequence with auto-correlation \(r(n)\sim c {n^{-\gamma }}\) for some \(\gamma \in (0, 1)\) and \(c\in {\mathbf {R}}\) (allowing for negative correlations), G a centred function with Hermite rank \(m\ge 1\), and A(n) a sequence such that
Then, \(z_n=\frac{1}{A(n)} \sum _{k=1}^{[nt]} G(X_n)\) is expected to converge in finite dimensional distributions. The scaling constant A(n) is of the order \(n^{1-\frac{1}{2} \gamma m}\) in the long-range-dependent case, of order \(\sqrt{n}\) in the short-range-dependent case, and of order \(\sqrt{n \ln n}\) for the borderline case, see [6, 8]. By long-range dependence, we mean \(\sum _{n=1}^\infty |r(n)|= + \infty \).
The limit process, \(\lim _{n\rightarrow \infty } z_n\), is a Wiener process for fast decaying correlations, i.e. in case \(\gamma \in (\frac{1}{m}, 1)\), [6]. In the borderline case, \(\gamma =\frac{1}{m}\), the scaling limit is also a Wiener process. However, if \(\gamma \in (0, \frac{1}{m})\) the correlations fail to decay sufficiently fast, the scaling limit is a Hermite process in the m-th chaos, [8, 12]. The first convergence to a non-Gaussian similar process was shown in [36] where the aim was to construct a not strong mixing sequence of random variables, he achieved this by showing the sequence of random variables has a non-Gaussian scaling limit which is now known as the Rosenblatt process. In [8], vector valued combinations of short- and long-range-dependent sums were studied, however, the limit of each component is assumed to be moment determinate (which can only happen when they are in the \(L^2\) chaos expansion of order less or equal to 2). This is due to a restriction in the asymptotic independence result in [33], which was extended in [29].
We return to the continuous functional limit theorems. For the scalar case, the continuous version CLT for \(\gamma \in ( \frac{1}{m} ,1 )\) was obtained in [4], the borderline case \(\gamma = \frac{1}{m}\) in [2]. These are shown for the convergence in finite dimensional distribution and for G to be a centred \(L^2\) function. They also obtained uniform convergences in the continuous topology for a restrictive class of functions G (assuming sufficiently fast decay of the coefficients in the Wiener chaos expansion). This was extended in [30] to vector valued \(X^\varepsilon \), when each component of \(X^\varepsilon \) falls in the Brownian case, with convergence understood in the sense of finite dimensional distributions. The result in [30] was improved in [11], where the fast chaos decay restrictions on \(G_k\), for \(G_k \in L^p\) for \(p>2\), are removed with techniques from Malliavin calculus. In the continuous long-range-dependent case Taqqu, [40], obtained convergence in the continuous topology. These results, although fragmented (in some regimes these are only known for scalar valued processes or only at the level of sequences), provide a fairly good picture of what is going on.
There exists however no vector valued functional limit theorem with joint convergence, when the scaling limit of the components are mixed, in this article we provide a complete description for the joint convergence of \(\{X^{k,\varepsilon }\}\) for \(G_k \in L^2(\mu )\). We have a functional limit theorem for vector valued processes whose components may exhibit both short and long-range dependence. For \(G_k\) satisfying a stronger integrability condition, we can also show weak convergence in the \({\mathcal {C}}^\alpha ([0,T],{\mathbf {R}}^d)\)-topology and for each fixed time in \(L^2\) for the low Hermite rank case, which already have interesting applications. Furthermore, they are the basis for the convergence in a suitable rough topology, which due to the change of the nature of the problem will appear in [18] where rough path theory is used to study slow/fast systems, leading to ‘rough creation’ / ‘rough homogenization’ in which the effective limit is not necessarily a Markov process.
Application. Consider the second-order equation on \({\mathbf {R}}\):
Taking \(\varepsilon \rightarrow 0\), does \(x_t^\varepsilon \) converge? In case \(H=\frac{1}{2}\) and \(f=1\), this is essentially the Kramer-Smolouchowski limit (this is also called the kinetic Brownian motion model). For \(H\not =\frac{1}{2}\) and for \(f=1\) this was shown in [1, 7, 43] to converge to a fBM, see also [14] for the case with a magnetic field. Given \(H>\frac{1}{3}\) and \(f\in C_b^3\) (for any H if \(f=1\)), we can show \(x_t^\varepsilon \) converges to the solution of the equation \(\dot{x}_t=f(x_t) \;d B_t^H \) with initial value \(x_0\) where the integral in the differential equation is interpreted as a Riemann-Stieltjes integral. Furthermore, we obtain the following bound in \(C^{\gamma '}\) where \(0<\gamma '<\gamma < H\):
This computation is straightforward, see Propositions 3.1 and 3.2 for detail.
With the functional limit theorem below, Theorem A, we can conclude also the convergence of solutions of the equations, for \(h \in {\mathcal {C}}^{2}_b({\mathbf {R}}^d,{\mathbf {R}}^d)\) and \(g \in {\mathcal {C}}_b({\mathbf {R}},{\mathbf {R}})\),
We show that \( x_t^\varepsilon \) converges in \({\mathcal {C}}^\gamma ([0,T],{\mathbf {R}})\) for \(\gamma \in (0, H^*(m) \vee \frac{1}{2} - \frac{1}{p})\) either to the solution of the equation
where \(Z_t^{H^*(m),m}\) is a Hermite process, or to the solution to the Stratonovich stochastic differential equation
where \(W_t\) is a standard Wiener process (given enough integrability on G). Here, c is a specific constant (c.f. Eq. (3.6) depending on G arising from the homogenization procedure. For the above, we follow [11] and use Malliavin calculus to obtain suitable moment bounds on \(\int _0^t G(y^{\varepsilon }_s) ds\). These results appeared in the previous version of the current paper [18]. Equations driven by fractional Brownian motions are also studied for the averaging regime, see [21] and [16]. A fluctuation theorem around the effective average was obtained [3].
Main Results We denote our underlying probability space by \((\Omega , \mathcal {F},\P )\). Let \(\mu \) denote the standard Gaussian distribution and we choose \(\sigma \) such that the stationary scaled fOU process, to be defined below, satisfies \(y^{\varepsilon }_t \sim \mu \). Let \(\{H_m, m\ge 0\}\) be the orthogonal Hermite polynomials on \(L^2( \mu )\), such that they have leading coefficient 1 and \(L^2(\mu )\) norm \(\sqrt{m!} \). Given \(G\in L^2(\mu )\), then it posses an expansion of the form \( G(x)=\sum _{k=0}^\infty c_k H_k(x)\), where \( c_k=\frac{1}{k!}\langle G, H_k\rangle _{L^2(\mu )}\). A function G is centred if and only if \(c_0=0\). The smallest m with \(c_m\not =0\) is called the Hermite rank of G. In case the correlations of \(y^{\varepsilon }_t\) do not decay sufficiently fast the path integral \(\alpha (\varepsilon )\int _0^t G( y_s^\varepsilon )ds\) ought to be approximated by that of the first term of its Wiener chaos expansion. By orthogonality of the \(H_m\)’s, it is sufficient to study the asymptotics of \(\alpha (\varepsilon )\int _0^t H_m( y_s^\varepsilon )ds\) to deduce \(\alpha (\varepsilon )\).
Although the solutions to the fOU equation converge exponentially fast to each other, their autocorrelation function decays only algebraically. The indicator for the behaviour of \(\alpha (\varepsilon ) \int _0^t H_m(y_s^\varepsilon ) ds\) turns out to be
and the self-similarity exponent of the limiting process is determined by \(\alpha (\varepsilon ,H^*(m))\). For large m, the limit will be a Wiener process, and otherwise the limit \(Z_t\) should have the scaling property: \(\varepsilon ^{H^*(m)} Z_{\frac{t}{\varepsilon }} \sim Z_t\). Indeed, \(Z_t\) are the self-similar Hermite processes. To state the functional limit theorem concisely, we make the following convention,
Convention 1.1
Given a collection of functions \((G_k \in L^2(\mu ), k\le N)\), we will label the high rank ones first, so the first n functions satisfy \(H^*(m_k) \le \frac{1}{2}\), where \(n \ge 0\), and the remaining satisfy \(H^*(m_k) > \frac{1}{2}\).
Theorem A
Let \(y^\varepsilon \) be the stationary solution to the scaled fractional Ornstein–Uhlenbeck equation (2.4) with standard Gaussian distribution \(\mu \). Let \(G_k:{\mathbf {R}}\rightarrow {\mathbf {R}}\) be centred functions in \(L^2(\mu )\) with Hermite ranks \(m_k\). Write
Set
Then, the following holds:
-
1.
-
(a)
There exist stochastic processes \(X^W=( X^1, \dots , X^n)\) and \(X^Z=(X^{n+1}, \dots , X^N)\) such that on every finite interval [0, T],
$$\begin{aligned} (X^{W,\varepsilon }, X^{Z,\varepsilon }) \longrightarrow (X^W, X^Z), \end{aligned}$$in the sense of finite dimensional distributions. Furthermore, for any \(t >0\)
$$\begin{aligned} \lim _{\varepsilon \rightarrow 0} \Vert X^{Z,\varepsilon }_t \rightarrow X^Z_t\Vert _{L^2(\Omega )}=0. \end{aligned}$$ -
(b)
If furthermore each \(G_k\) satisfies Assumption 1.2 below, the convergence is weakly in \({\mathcal {C}}^{\gamma }([0,T],{\mathbf {R}}^N)\) for every \(\gamma < \frac{1}{2} - \frac{1}{\min _{k \le n } p_k}\), if there is at least one component converging to a Wiener process. Otherwise they converge in \({\mathcal {C}}^{\gamma }([0,T],{\mathbf {R}}^N)\) for every \(\gamma < \min _{k>n} H^*(m_k) - \frac{1}{p_k} \).
-
2.
We now describe the limit \(X= (X^W,X^Z)\).
-
(1)
\(X^W \in {\mathbf {R}}^n\) and \(X^Z \in {\mathbf {R}}^{N-n}\) are independent.
-
(2)
\( X^W = U {\hat{W}}_t\) where \( {\hat{W}}_t\) is a standard Wiener process and U is a square root of the matrix \((A^{i,j})\),
$$\begin{aligned} A^{i,j}=\int _0^{\infty } {\mathbf {E}}\left( G_i(y_s) G_j(y_0) \right) ds = \sum _{q=m_i\vee m_j}^{\infty } c_{i,q}\; c_{j,q} \; (q!) \, \int _0^\infty \rho (r)^q\, dr \end{aligned}$$and \(\rho (r)={\mathbf {E}}(y_ry_0)\). In other words, \({\mathbf {E}}\left( X^i_t X^j_s\right) = 2 (t \wedge s) A^{i,j}\) for \(i,j\le n\).
-
(3)
Let \(Z_t^{H^*(m_k),m_k}\) be the Hermite processes, represented by (2.1). Then,
$$\begin{aligned} X^Z=(c_{n+1,m_{n+1}} Z_t^{n+1} , \dots , c_{N,m_{N}} Z_t^{N}), \end{aligned}$$where,
$$\begin{aligned} Z_t^{k}= \frac{m_k!}{K(H^*(m_k),m_k)} Z_t^{H^*(m_k),m_k}. \end{aligned}$$(1.4)We emphasize that the Wiener process \(W_t\) defining the Hermite processes is the same for every k (c.f. Eq. (2.1)), which is in addition independent of \({\hat{W}}_t\).
-
(1)
-
(a)
Assumption 1.2
(Functional Limit \({\mathcal {C}}^\gamma \) assumptions) Let \(G_k\in L^{2}(\mu )\) with Hermite rank \(m_k\ge ~1\).
-
High rank case. If \(H^*(m_k) \le \frac{1}{2}\), assume \(G_k \in L^{p_k}(\mu )\) where \(\frac{1}{2} - \frac{1}{p_k} > \frac{1}{3}\) (i.e. \(p_k >6\)).
-
Low rank case. If \(H^*(m_k) > \frac{1}{2}\), assume \(G_k \in L^{p_k}(\mu )\) where \( H^*(m_k) - \frac{1}{p_k} > \frac{1}{2}\).
Remark 1.3
The case \(H=\frac{1}{2}\) is classical, and is not of interest here. In this case the result is independent of the Hermite rank and the scaling is given by \(\alpha (\varepsilon )=\frac{1}{\sqrt{\varepsilon }}\), due to the exponential decay of correlations.
An immediate application is the following rough homogenisation theorem for a toy model:
Theorem B
Let \(H \in (\frac{1}{3},1) \setminus \{ \frac{1}{2} \}\), \(f\in {\mathcal {C}}_b^3({\mathbf {R}}^d, {\mathbf {R}}^d)\), \(h \in {\mathcal {C}}^{2}_b({\mathbf {R}}^d;{\mathbf {R}}^d)\), \(G \in {\mathcal {C}}({\mathbf {R}},{\mathbf {R}})\) satisfying Assumption 1.2 and \(g \in {\mathcal {C}}_b({\mathbf {R}};{\mathbf {R}})\). Let \(\alpha (\varepsilon ) = \alpha (\varepsilon ,H^*(m))\). Fix a finite time T and consider
-
1.
If \(H^*(m) > \frac{1}{2}\), \(x_t^\varepsilon \) converges weakly in \({\mathcal {C}}^{\gamma }([0,T],{\mathbf {R}}^d)\) to the solution to the Young differential equation \(d{\bar{x}}_t = c f({\bar{x}}_t) \,dZ_t^{H^*(m),m} +\bar{g} h(\bar{x}_t) dt\) with initial value \(x_0\) for \(\gamma \in (0, H^*(m)- \frac{1}{p})\).
-
2.
If \(H^*(m) \le \frac{1}{2}\), \(x_t^\varepsilon \) converges weakly in \({\mathcal {C}}^\gamma ([0,T],{\mathbf {R}})\) to the solution of the Stratonovich stochastic differential equation \(d{\bar{x}}_t = c f({\bar{x}}_t) \circ \,dW_t + \bar{g} h(\bar{x}_t) dt\) with \( {\bar{x}}_0=x_0\), where \(\gamma \in (0, \frac{1}{2}- \frac{1}{p})\).
We also take the liberty to point out an intermediate result on the joint convergence of stochastic processes in finite \(L^2\) chaos, for it maybe of service. The proof is a slight modification of results in [29, 41].
Proposition 5.1
Let \(q_1 \le q_2 \le \dots \le q_n \le p_1 \le p_2 \le \dots \le p_m\). Let \(f_i^\varepsilon \in L^2({\mathbf {R}}^{p_i})\), \( g_i^\varepsilon \in L^2({\mathbf {R}}^{q_i})\), \(F^{\varepsilon }=\left( I_{p_1}(f^{\varepsilon }_1), \dots , I_{p_m}(f^{\varepsilon }_m)\right) \) and \(G^{\varepsilon }=\left( I_{q_1}(g^{\varepsilon }_1), \dots , I_{q_n}(g^{\varepsilon }_n)\right) \), where \(I_q\) denotes the Wiener integral of order q. Suppose that for every i, j, and any \(1 \le r \le q_i\):
Then, \(F^\varepsilon \rightarrow U\) and \(G^{\varepsilon } \rightarrow V\) weakly imply that \((F^{\varepsilon },G^{\varepsilon }) \rightarrow (U,V)\) jointly, where U and V are taken to be independent random variables.
2 Preliminaries
We take the Hermite polynomials of degree m to be
Thus, \(H_0(x)=1\), \(H_1(x)=x\). Let \({\hat{H}}\) be the inverse of \(H^*(m)=m(H-1)+1\):
2.1 Hermite Processes
The rank 1 Hermite processes \(Z^{H, 1}\) are fractional BMs, the formulation (2.1) below is exactly the Mandelbrot-Van Ness representation for a fBM.
Definition 2.1
Let \(m\in {\mathbf {N}}\) with \({\hat{H}}(m)>\frac{1}{2}\). The class of Hermite processes of rank m is given by the following mean-zero processes,
where the constant K(H, m) is chosen so their variances are 1 at \(t=1\).
The integration in (2.1) is understood as a multiple Wiener-Itô integral (over the region \({\mathbf {R}}^n\) without the diagonal). Note \({\hat{H}}(1)=H\).
By the properties of Wiener integrals, two Hermite processes \(Z^{H, m}\) and \(Z^{H',m'}\), defined by the same Wiener process, are uncorrelated if \(m \not = m'\). The Hermite processes have stationary increments and finite moments of all orders with covariance
Therefore, using Kolmogorov’s theorem 2.6, one can show that the Hermite processes \(Z_t^{H,m}\) have sample paths of Hölder regularity up to H. They are also self-similar with exponent H which means \( \lambda ^H Z^{H,m} _{\frac{\cdot }{\lambda }} \sim Z^{H,m}_.\).
Remark 2.2
In some literature, see, e.g. [27] where further details on Hermite processes can also be found, the Hermite processes are defined with a different exponent as below:
The two notions are related by
We refer to [10, 35, 37] for detailed studies of fBM’s which are used in this article.
2.2 Fractional Ornstein–Uhlenbeck Processes
Let us normalise the fractional Brownian motion, so that \(B^H_0=0\) and \({\mathbf {E}}(B^H_1)^2=1\). Disjoint increments of \(B_t^H\) have a covariance of the form:
We define the stationary fractional Ornstein–Uhlenbeck processes to be \(y_t = \sigma \int _{-\infty }^t e^{-\sigma (t-s) } dB^H_s\), where \(B^H_t\) is a two-sided fractional BM, \(\sigma \) is chosen such that \(y_t\) is distributed as \(\mu =N(0,1)\). It is the solution of the following Langevin equation:
We take \(y_t^\varepsilon \), the fast or rescaled fOU, to be the stationary solution of
Observe that \( y_\cdot ^\varepsilon \) and \( y_{\frac{\cdot }{\varepsilon }} \) have the same distributions, \(y^\varepsilon _t=\frac{\sigma }{\varepsilon ^H}\int _{-\infty }^t e^{-\frac{1}{\varepsilon }(t-s) } dB^H_s\). In particular, both \(y_t\) and \(y^{\varepsilon }_t\) are Hölder continuous with Hölder exponents \(\gamma \in (0,H)\). Let us denote their correlation functions by \(\rho \) and \(\rho ^\varepsilon \):
Let \(\rho (s)={\mathbf {E}}(y_0y_s)\) for \(s\ge 0\) and extended to \({\mathbf {R}}\) by symmetry, then \(\rho (s,t)=\rho (t-s)\) and similarly for \(\rho ^{\varepsilon }\). For \(H>\frac{1}{2}\), the set of functions for which Wiener integrals are defined include \(L^2\) functions and so \(\rho \) posses an analytical expression. Indeed, since
we have
which is integrable, and therefore we may use the Wiener isometry to compute the covariances
For \(u>0\), we set
Using this, the following correlation decay was shown in [10].
Lemma 2.3
Let \(H\in (0, \frac{1}{2})\cup (\frac{1}{2}, 1)\). Then, \(\rho (s)=2\sigma ^2 H(2H-1) s^{2H-2} +O(s^{2H-4})\) as \( s \rightarrow \infty \). In particular, for any \(t\not =s\),
Hence, \( \int _0^\infty \rho ^m(s) ds\) is finite if and only if \( H^*(m)<\frac{1}{2}\), or \(H=\frac{1}{2}\) and \(m\in {\mathbf {N}}\), as in the latter the usual OU process admits exponential decay of correlations.
Lemma 2.4
Let \(H\in (0,1) \setminus \{ \frac{1}{2}\}\), fix a finite time horizon T, then for \(t \in [0,T]\) the following holds uniformly for \(\varepsilon \in (0,\frac{1}{2}]\):
Note, if \(H=\frac{1}{2}\), for and any \(m \in {\mathbf {N}}\), the bound is always \( \sqrt{\frac{t}{\varepsilon }} \int _0^\infty \rho ^m(s) ds\). In particular,
Proof
We first observe that
By a change of variables and using estimate (2.5) on the decay of the auto-correlation function,
For the case \(H^*(m)=\frac{1}{2}\), we use
To complete the proof, we observe that by a simple change of variables,
\(\square \)
Next, we recall the Garsia-Rodemich-Romsey-Kolmogorv inequality.
Lemma 2.5
(Garsia-Rodemich-Romsey-Kolmogorov inequality) Let \(T>0\). Let \(\theta : [0,T]\rightarrow {\mathbf {R}}^d\). For any positive numbers \(\gamma , p\), there exists a constant \(C(\gamma , p)\) such that
See [17, section A.2] in Appendix, as well as [39], for a proof. As a consequence of this inequality, one obtains the following theorem.
Theorem 2.6
(Kolmogorov’s Theorem) Let \(\theta \) be a stochastic process. Suppose that for \(s,t\in [0, T]\), \(p>1\) and \(\delta >0\),
where \(c_p\) is a constant. Then, for \(\gamma <\frac{\delta }{p}\), \(\theta \in {\mathcal {C}}^{\gamma }([0,T],{\mathbf {R}})\), and in particular
where right hand side is finite when \(\gamma \in (0, \frac{\delta }{p})\).
Lemma 2.7
For any \(\gamma \in (0, H)\), \(p>1\), the following estimates hold:
Proof
We use the fractional Ornstein–Uhlenbeck equation
to obtain \({\mathbf {E}}|y_s-y_r|^2 \lesssim (s-r)^2 {\mathbf {E}}\vert y_1 \vert ^2+q|s-r|^{2H}\). Using the stationarity of \(y_t\), one also has \( {\mathbf {E}}|y_s-y_r|^{2} \le 2{\mathbf {E}}|y_1|^2=2\). Since for Gaussian random variables the \(L^2\) norm controls the \(L^p\) norm we have
Thus, by symmetry and a change of variables, \(\int _0^T \int _0^T \frac{ {\mathbf {E}}|y_s-y_r|^p}{ {|s-r|}^{ \gamma p+2}}dsdr\lesssim T\) and application of Kolmogorov’s theorem 2.6 concludes the proof. \(\square \)
3 Applications
3.1 The Second-Order Problem
If x is a stochastic process, we write \(x_{s,t}=x_t-x_s\).
Proposition 3.1
Let \(H\in (0,1)\), \(\gamma \in (0,H)\), \(p>1\) and fix a finite time T. Let \(X^{\varepsilon }_t = \varepsilon ^{H-1} \int _0^{t } y^\varepsilon _{s} ds\), then,
for any \(\gamma '<\gamma <H\) and for \(t \in [0,T]\).
Proof
Set \(v_t^\varepsilon =\varepsilon ^{H-1} y^\varepsilon _{t}\), then, \(v_t^\varepsilon \) solves the following equation
Using the equation for \(v_t^\varepsilon \), we have
Therefore, for any \(p>1\),
In the last step, we used the stationarity of \( y_t^\varepsilon \).
Thus, applying Kolmogorov’s theorem 2.6 to \(X^\varepsilon - \sigma B^H\), we see that the following holds for any \(t \in [0,T]\), \(p>1\) and any \(\gamma '<\gamma \)
hence, the claim follows. \(\square \)
As an application, we consider the following slow/fast system,
To describe its limit, we review the concept of Young integrals. If \(f, g: [0, T]\rightarrow {\mathbf {R}}\) with \(f\in {\mathcal {C}}^\alpha \) and \(g\in {\mathcal {C}}^\beta \) such that \(\alpha +\beta >1\), then the Riemann-Stieljes integral makes sense, and
For details see [42]; this integral is called a Young integral. Since Young integrals have the regularity of its integrand, for \(H>\frac{1}{2}\) the equation \(\dot{x}_t=f(x_t) \;dB^H_t\) makes sense. In [25], it was shown that if \(f \in {\mathcal {C}}_b^3\), then the equation has a unique global solution from each initial value. This type of equations are Young equation, the simplest type of rough differential equation. The notation \({\mathcal {C}}_b^3\) denotes the space of bounded functions such that their first three derivatives are bounded.
Proposition 3.2
Let \(H \in ( \frac{1}{3}, 1)\) and \(f\in {\mathcal {C}}_b^3({\mathbf {R}}^d,{\mathbf {R}}^d)\). Then for any \(\gamma \in (0,H)\), \(\gamma ' < \gamma \), \(x_t^\varepsilon \) converges in \(L^p\) in \({\mathcal {C}}^{\gamma '}([0,T], {\mathbf {R}}^d)\) to the solution of the rough differential equation:
Furthermore, for \(t \in [0,T]\),
Proof
The idea is to consider Eq. (3.1) as a Young/rough differential equation. In case \(H \in (\frac{1}{2},1)\), we can rewrite our equation as
where \(X^{\varepsilon }=\varepsilon ^{H-1} \int _0^{t } y^\varepsilon _{s} ds\) is as in the previous lemma. Young’s continuity theorem, see Theorem 3.6, states that \(x^\varepsilon \) converges weakly provided \(X^\varepsilon \) converges weakly in a Hölder space of regularity greater than \(\frac{1}{2}\).
For \(H \in (\frac{1}{3}, \frac{1}{2})\), we need to rewrite our equation into a rough differential equation,
where \({{\mathbf {X}}}^{\varepsilon }\) is given by \(X^{\varepsilon }\) enhanced with its canonical lift
As we restrict ourselves to one dimension we obtain, by symmetry, \({\mathbb X}^{\varepsilon }_{s,t}= \frac{1}{2} {(X^{\varepsilon }_{s,t})^2}\), hence, \({{\mathbf {X}}}^{\varepsilon }\) converges to a fBm \(\sigma B^H\) enhanced with \(\mathbb {B}^H_{s,t}=\frac{1}{2} { \left( \sigma B^H_{s,t}\right) ^2} \). As the solution map, \(\Phi \), to a RDE satisfies, see [15] or Theorem 3.6 below,
where \(\rho _{\gamma }\) denotes the inhomogeneous rough path norm of regularity \(\gamma \) and \(\mathbf {B}^H = (\sigma B^H, \frac{1}{2} { \left( \sigma B^H_{s,t}\right) ^2} )\). Thus, the \(L^p\) convergence of the solutions follows from the \(L^p\) convergence of the drivers, hence, we can conclude the proof by Proposition 3.1. \(\square \)
Remark 3.3
Krammer-Smoluchowski limits/Kinetic fBM’s are studied in [1, 7, 43]. See also [14,15,16].
Remark 3.4
-
1.
For \(H<\frac{1}{2}\) and \(m=1\), Theorem A appears to contradict with Proposition 3.1; in the first we claim the limit is a Brownian motion, whereas in the second we claim that it is a fractional Brownian motion. Both results are correct and can be easily explained. It lies in the fact that \(\int _{\mathbf {R}}\rho (s)\, ds\) vanishes if \(H<\frac{1}{2}\), and so the Brownian motion limit is degenerate. Since according to [10],
$$\begin{aligned} \rho (s) = \sigma ^2 \frac{\Gamma (2H+1) \sin (\pi H)}{2 \pi } \int _{{\mathbf {R}}} e^{isx} \frac{\vert x \vert ^{1-2H}}{1 + x^2} dx, \end{aligned}$$(3.3)and by the decay estimate from (2.5), \(\rho \) is integrable, \(s(\lambda ) \) is the value at zero of the inverse Fourier transform of \(\rho (s)\), which is up to a multiplicative constant \(\frac{\vert \lambda \vert ^{1-2H}}{1 + \lambda ^2}\). This is also the spectral density of \(y_t\) and has value 0 at 0. This means we have scaled too much and the correct scaling is to multiply the integral \(\int _0^{t} y^{\varepsilon }_s ds\) by \(\varepsilon ^{H-1}\) in which case we obtain a fBm as limit.
-
2.
For \(m>1\) and \(H<\frac{1}{2}\) the Wiener limit is not trivial. Indeed,
$$\begin{aligned} \begin{aligned} \int _{\mathbf {R}}\rho (s)^m ds&=C \int _{\mathbf {R}}{\mathop {\overbrace{ \int _{\mathbf {R}}\dots \int _{{\mathbf {R}}} }}\limits ^{m}} \prod _{k=1}^m e^{isx_k} \frac{\vert x_k \vert ^{1-2H}}{1 + |x_k|^2} dx_1\dots dx_m\, ds\\&= C {\mathop {\overbrace{ \int _{\mathbf {R}}\dots \int _{{\mathbf {R}}}}}\limits ^{m}} \frac{\vert x_2+\dots + x_m \vert ^{1-2H}}{1 + |x_2+\dots + x_m|^2} \prod _{k=2}^m \, \frac{\vert x_k \vert ^{1-2H}}{1 + |x_k|^2} \not =0. \end{aligned} \end{aligned}$$
3.2 The 1d Fluctuation Problem
In this section, we give an application of Theorem A. Given a function \(g \in L^1(\mu )\) we denote \( \bar{g}=\int _{{\mathbf {R}}} g(y) \mu (dy)\).
Lemma 3.5
The stationary Ornstein–Uhlenbeck process is ergodic. Thus, \(\int _0^t g(y^{\varepsilon }_s) ds\rightarrow t {\bar{g}}\) in probability for every \(g \in L^1(\mu )\).
Proof
A stationary Gaussian process is ergodic if its spectral measure has no atom, see [9, 37]. The spectral measure F of a stationary Gaussian process is obtained from Fourier transforming its correlation function and \(\rho (\lambda )=\int _{\mathbf {R}}e^{i \lambda x} dF(x)\). According to [10]:
so the spectral measure is absolutely continuous with respect to the Lebesgue measure with spectral density, up to a nonzero constant, given by \(s(x) = \frac{ \vert x \vert ^{1-2H}}{1+ x^2}\). Since \(\int _0^t g(y^{\varepsilon }_t) dt\) equals \( \varepsilon \int _0^{\frac{t}{\varepsilon }} g(y_s) ds\) in law, the former converges in law to \( t \bar{g}\) by Birkhoff’s ergodic theorem. The claim now follows as weak convergence to a constant implies convergence in probability. \(\square \)
In the following proof we will need the following theorem from Young/rough path theory for details we refer to [15, 17, 24, 25]. We denote the space of rough paths of regularity \(\beta \) by \(\mathscr {C}^{\beta }\).
Theorem 3.6
Let \(Y_0 \in {\mathbf {R}}^d, \beta \in (\frac{1}{3}, 1), \, \gamma \in (\frac{1}{2}, 1) \, f \in {\mathcal {C}}^3_b({\mathbf {R}}^d,{\mathbf {R}}^d) \), \(h \in {\mathcal {C}}_b^{2}({\mathbf {R}}^d,{\mathbf {R}}^d)\), \({{\mathbf {X}}}\in \mathscr {C}^{\beta }([0,T],{\mathbf {R}})\) and \(\mathbf {Z} \in \mathscr {C}^{\gamma }([0,T],{\mathbf {R}})\) such that \(\beta + \gamma >1\). Then, the differential equation
has a unique solution which belongs to \({\mathcal {C}}^{\beta \wedge \gamma }\). Furthermore, the solution map \(\Phi _{f,h}: ~{\mathbf {R}}\times \mathscr {C}^{\beta }([0,T], {\mathbf {R}}) \times \mathscr {C}^{\gamma }([0,T], {\mathbf {R}}) \rightarrow {\mathcal {C}}^{\beta \wedge \gamma }([0,T],{\mathbf {R}})\), where the first component is the initial condition and the second and third components the drivers, is continuous.
Given a centred function \(G \in L^2(\mu ) \), with chaos expansion \(G=\sum _{k=m}^{\infty } c_k H_k\) and let \(c \ge 0\) be given by
Theorem 3.7
Let \(H \in (\frac{1}{3},1) \setminus \{ \frac{1}{2} \}\), \(f\in {\mathcal {C}}_b^3({\mathbf {R}}^d, {\mathbf {R}}^d)\), \(h \in {\mathcal {C}}^{2}_b({\mathbf {R}}^d;{\mathbf {R}}^d)\), \(G \in {\mathcal {C}}({\mathbf {R}},{\mathbf {R}})\) satisfying Assumption 1.2 and \(g \in {\mathcal {C}}_b({\mathbf {R}};{\mathbf {R}})\). Let \(\alpha (\varepsilon ) = \alpha (\varepsilon ,H^*(m))\). Fix a finite time T and consider
-
1.
If \(H^*(m) > \frac{1}{2}\), \(x_t^\varepsilon \) converges weakly in \({\mathcal {C}}^{\gamma }([0,T],{\mathbf {R}}^d)\) to the solution to the Young differential equation \(d{\bar{x}}_t = c f({\bar{x}}_t) \,dZ_t^{H^*(m),m} +\bar{g} h(\bar{x}_t) dt\) with initial value \(x_0\) for \(\gamma \in (0, H^*(m)- \frac{1}{p})\).
-
2.
If \(H^*(m) \le \frac{1}{2}\), \(x_t^\varepsilon \) converges weakly in \({\mathcal {C}}^\gamma ([0,T],{\mathbf {R}})\) to the solution of the Stratonovich stochastic differential equation \(d{\bar{x}}_t = c f({\bar{x}}_t) \circ \,dW_t + \bar{g} h(\bar{x}_t) dt\) with \( {\bar{x}}_0=x_0\), where \(\gamma \in (0, \frac{1}{2}- \frac{1}{p})\).
Proof
As in Proposition 3.2 we can rewrite our equations as Young/rough differential equations and therefore reduce our analysis to the drivers \(\left( \alpha (\varepsilon ) \int _0^t G(y^{\varepsilon }_s) ds, \int _0^t g(y^{\varepsilon }_s) ds\right) \). By Theorem A, \(\alpha (\varepsilon ) \int _0^t G(y^{\varepsilon }_s) ds\) converges in finite dimensional distributions either to a Wiener or a Hermite process. By Lemma 3.5, \(\int _0^t g(y^{\varepsilon }_s) ds\) converges in probability to the deterministic path \(t \bar{g}\). Hence, \(\left( \alpha (\varepsilon ) \int _0^t G(y^{\varepsilon }_s) ds, \int _0^t g(y^{\varepsilon }_s) ds\right) \) converges jointly in finite dimensional distributions. Furthermore, \(\Vert \int _0^t g(y^{\varepsilon }_s) ds \Vert _{\infty } \le t \Vert g \Vert _{\infty }\), this combined with the moment bounds obtained in Theorem A enables us to apply Theorem 3.6 to conclude the proof. \(\square \)
Remark 3.8
The constant c could be 0, for further details see Remark 3.4.
4 Proof of Theorem A
We first establish the \(L^2(\Omega )\) convergence of \(X_t^\varepsilon = \alpha (\varepsilon )\int _0^{t} G(y^{\varepsilon }_s)ds\), where \(G=\sum _{k=m}^\infty c_k H_k\) has low Hermite rank, followed by a reduction theorem. We then prove moment bounds and conclude the proof of Theorem A.
4.1 Preliminary Lemmas
The basic scalar valued functional limit theorem, for low rank Hermite functions, was proved in [40] for \(\varepsilon ^{H^*(m)} \int _0^{\frac{t}{\varepsilon }} G(X_s) ds\) with \(X_t = \int _{{\mathbf {R}}} p(t-\xi ) dW_{\xi }\) a moving average, where in order to prove convergence one uses the self-similarity of a Wiener process leading to weak convergence as this equivalence of course is only in law. Nevertheless, in our case we can choose a properly scaled fast variable and write, \(y_t^{\varepsilon } = \int _{{\mathbf {R}}} \hat{g} (\frac{t-\xi }{\varepsilon }) dW_{\xi }\) for a function \({\hat{g}}\), and thus avoid using self-similarity. The key idea is to write a Wiener integral representation beginning with
and \(c_1(H) = \sqrt{ \int _{-\infty }^0 \left( (1-s)^{H-\frac{1}{2}} - (-s)^{H- \frac{1}{2}} \right) ^2 ds + \frac{1}{2H} }\). This can be obtained by applying the integral representation for fBM’s:
and by repeated applications of integration by parts (to the Young integrals):
One may also use the following, see [34], taking \(f\in L^1\cap L^2\):
Lemma 4.1
Let \(\lambda \) denote the Lebesgue measure, then as \(\varepsilon \rightarrow 0\), \(\varepsilon ^{H^*(m) -1} \int _0^t H_m(y^{\varepsilon }_s) ds\) converges to \( \frac{ m!}{ K(H,m)} Z_t^{H^*(m),m}\) in \(L^2(\Omega )\). Equivalently,
Proof
This can be shown by applying [40, Theorem 4.7], where weak convergence is obtained. With a small modification and using [40, Lemma 4.5, Lemma 4.6] directly we obtain the \(L^2(\Omega )\) convergence:
using the Wiener integral representation of the Hermite processes this is equivalent, by a multiple Wiener-Itô isometry, to
Examining Taqqu’s proof, we note that in fact the \(L^2\) convergence of (4.4) is obtained under the following conditions.
-
1
\(\int _{{\mathbf {R}}} p(s)^2 ds < \infty \).
-
2
\( \vert p(s) \vert \le C s^{H- \frac{3}{2}} L(u)\) for almost all \(s>0\).
-
3
\(p(s) \sim s^{H - \frac{3}{2} } L(s)\) as \(s \rightarrow \infty \).
-
4
There exists a constant \(\gamma \) such that \(0<\gamma < (1-H)\wedge (H- (1-\frac{1}{2m}))\) such that \(\int _{-\infty }^0 \vert p(s) p(xy+s)\vert ds = o(x^{2H-2} L^2(x)) y^{2H-2-2\gamma }\) as \(x \rightarrow \infty \) uniformly in \(y \in (0,t]\).
where L denotes a slowly varying function (for every \(\lambda >0\) \(\lim _{x \rightarrow \infty } \frac{L(\lambda x)}{L(x)} ) =1\)). Set
then,
We are now in Taqqu’s framework and it is only left the check \(\hat{g}\) defined by (4.5) satisfies these conditions. To increase readability, we suppress the constant \(\frac{\sigma }{c_1(H)}\) in the following computations. For \(s<1\),
We calculate for \(s>1\) via integration by parts
This of course implies that \(\hat{g}\) is \(L^2(\lambda )\) integrable. Finally, observe that \(\int _{-\infty }^0 \vert \hat{g}(s) \hat{g}(xy+s)\vert ds = 0\) as \(\hat{g}(s)=0\) for \(s<0\). With these we apply [40, Theorem 4.7] to conclude the \(L^2\) convergence of the kernels. \(\square \)
Lemma 4.2
Let \(G\in L^2(\mu )\) be a centred function with Hermite rank m satisfying \(H^*(m)> \frac{1}{2}\). Let \(H\in ( \frac{1}{2} ,1)\). Then, the following statements hold for the stationary scaled fOU process \(y^{\varepsilon }_s\). Fix \(t>0\), then,
Proof
For \(G= H_m\), the claim has already been shown in Lemma 4.1. To conclude the claim in the case of a general G, we compute,
as \(\sum _{k=m+1}^{\infty } c_k^2 \sqrt{k!} < \infty \) as \(G \in L^2(\mu )\). This finishes the proof. \(\square \)
The fact that only the first term in the chaos expansion gives a contributions is in the literature often called a reduction lemma. In the high Hermite rank case however it is not possible to restrict one’s analysis to a pure Hermite polynomial, but as the next lemma shows finite linear combinations are indeed sufficient. To make the application later on easier, we directly prove it in the multi-dimensional case.
Lemma 4.3
(Reduction Lemma) Fix \(H \in (0,1)\setminus \{\frac{1}{2}\}\). For \(M\in {\mathbf {N}}\), define the truncated functions:
If for every \(M \in {\mathbf {N}}\), \((X^{1,\varepsilon }_{M}, \dots , X^{N,\varepsilon }_{M}){\mathop {\longrightarrow }\limits ^{(\varepsilon \rightarrow 0)}}(X^{1}_{M}, \dots , X^{N}_{M})\) in finite dimensional distributions, then,
in finite dimensional distributions.
Proof
Firstly,
Using properties of the Hermite polynomials, we obtain
As \(\sum _{j=m}^\infty (c_{k,j})^2 j! < \infty \) we obtain \( \sum _{j=M+1}^\infty (c_{k,j})^2 j!\rightarrow 0\) as \(M \rightarrow \infty \). Thus,
Let \( \{t_{\gamma _{k,l}}, k\le N, l\le A \} \) be a sequence of positive numbers. Now, by the triangle inequality,
With Theorem 3.2 in [5], this proves the claim. \(\square \)
4.2 Moment Bounds
We will use some results from Malliavin Calculus. Let \(x_s\) be a stationary Gaussian process with \(\beta (s) = {\mathbf {E}}\left( x_s x_0 \right) \), such that \(\beta (0)=1\). As a real separable Hilbert space we use \(\mathscr {H} = L^2({\mathbf {R}}_+,\nu )\) where for a Borel-set A we have \(\nu (A) = \int _{{\mathbf {R}}_+} \mathbf {1}_{A} (s) d\beta _s\). Let \({{\mathscr {H}}}^{\otimes q}\) denote the q-th tensor product of \({{\mathscr {H}}}\). For \(h\in {{\mathscr {H}}}\), we may define the Wiener integrals \(W(h)=\int _0^\infty h_s dx_s\) by \(W([a,b])=x(b)-x(a)\) (where \(a, b\ge 0\)), linearity and the Wiener isometry (\(\langle \mathbf {1}_{[0,t]}, \mathbf {1}_{[0,s]}\rangle =\beta (t-s)\)). Iterated Wiener integrals are defined similarly and by its values on indictor functions: \(I_m(\mathbf {1}_{A_1\times \dots \times A_m})=\prod _{i=1}^m W (A_i)\) where \(A_i\) are pairwise disjoint Borel subsets of \({\mathbf {R}}_+\). If \({{\mathcal {F}}}\) denotes the \(\sigma \)-field generated by x, then any \({{\mathcal {F}}}\)-measurable \(L^2(\Omega )\) function F has the chaos expansion: \(F={\mathbf {E}}F+ \sum _{m=1}^\infty I_m(f_m)\) where \(f_m\in L^2( {\mathbf {R}}_+^m)\). This is due to the fact that \(L^2(\Omega )=\bigoplus _{m=0}^{\infty } {{\mathscr {H}}}_m\) where \({{\mathscr {H}}}_m\) is the closed linear space generated by \(\{ H_m(W(h)): \Vert h \Vert _{L^2}=1\}\), \(H_m\) are the m-th Hermite polynomials, and \({{\mathscr {H}}}_m=I_m(L_{\hbox {sym}}^2({\mathbf {R}}_{+}^m))\). The last fact is due to \(H_m(W(h))= I_m( h^{\otimes ^{m} })\). In the following \({\mathbb {D}^{k,p}(\mathscr {H}^{\otimes m})}\) denotes the closure of Malliavin smooth random variables under the following norm \(\Vert u \Vert _{\mathbb {D}^{k,p}(\mathscr {H}^{\otimes m})} = \left( \sum _{j=0}^{k} {\mathbf {E}}\left( \Vert D^j u \Vert _{\mathscr {H}^{\otimes m}}^p \right) \right) ^{\frac{1}{p}}\).
Lemma 4.4
(Meyer’s inequality) [32] Let \(\delta \) denote the divergence operator. Then for \(u \in \mathbb {D}^{k,p}(\mathscr {H}^{\otimes m})\),
Lemma 4.5
[11] If \(G:{\mathbf {R}}\rightarrow {\mathbf {R}}\) is a function of Hermite rank m, then G has the following multiple Wiener-Itô-integral representation:
where \(G_m\) has the following properties:
-
(1)
\(\Vert G_m \Vert _{L^p(\mu )} \lesssim \Vert G \Vert _{L^p(\mu )}\),
-
(2)
\(G_m(x_1)\) is m times Malliavin differentiable and its \(k^{th}\) derivative, denoted by \(G_m^{(k)}(x_1)\mathbf {1}_{[0,1]}^{\otimes k}\), satisfies \( \Vert G_m^{(k)} \Vert _{L^p(\mu )} \lesssim \Vert G \Vert _{L^p(\mu )}\).
In the lemma below we estimate the moments of \(\int _0^{t} G(x_{\frac{r}{\varepsilon }}) dr\), where we need the multiple Wiener-Itô-integral representation above to transfer the correlation function to \(L^2\) norms of indicator functions. We use an idea from [11] for the estimates below.
Lemma 4.6
Let \(x_t=W([0,t])\) be a stationary Gaussian process with correlation \(\beta (t)=\big | {\mathbf {E}}(x_t x_0)\big |\), stationary distribution \(\mu \) and \({{\mathscr {H}}}\) the \(L^2\) space over \({\mathbf {R}}_{+}\) with measure \(\beta (r) dr\). If G is a function of Hermite rank m and \( G \in L^p(\mu )\), for \(p>2\), then,
For the stationary scaled fractional OU process \(y^{\varepsilon }_t\), we have
in particular,
Proof
We first use Lemma 4.5 and then apply Meyer’s inequality from Lemma 4.4 to obtain
Here for \(G_m\) is as given in Lemma 4.5 and \(G_m^{(k)}\) denotes its k-th order Malliavin derivative. We estimate the individual terms using the linearity of the inner product and the isometry \(\langle \mathbf {1}_{[0,r]}, \mathbf {1}_{[0,s]}\rangle _{{\mathscr {H}}}={\mathbf {E}}(x_rx_s)=\beta (r-s)\),
Using Minkowski’s inequality, we obtain
We then estimate \({\mathbf {E}}|G_m^{(k)} (x_r) G_m^{(k)} (x_u)|^{\frac{p}{2}}\) by Hölder’s inequality and use the fact that \(x_t\) is stationary. Since the right hand side is then controlled by
concluding (4.10). We finally apply Lemma 2.4 to conclude (4.11). \(\square \)
Now we are ready to prove our main theorem.
4.3 Concluding the Proof
Proof
Step 1, CLT in the pure Wiener case
We first deal with the high Hermite rank components. For \(k \le n\) we define the truncated functions \(G_{k,M}=\sum _{j=m_k}^{M} c_{k,j} H_j\) and set
Then, by the reduction Lemma 4.3 above, it is sufficient to show the convergence of \((X^{1,\varepsilon }_M, \dots , X^{n,\varepsilon }_M)\) for every M. By [4] and [2] each component alone converges to a Wiener process. Hence, as each \(X^{k,\varepsilon }_M\) belongs to a finite chaos we can make use of the normal approximation theorem from [32, Theorem 6.2.3]: if each component of a family of mean zero vector valued stochastic processes, with components of the form \(I_{q_i}(f_{i,n}) \), where \(f_{i,n}\) are symmetric \(L^2\) functions in \(q_i\) variables, converges in law to a Gaussian process, then they converge jointly in law to a vector valued Gaussian process, provided that their covariance functions converge. Furthermore, the covariance functions of the limit distribution are \(\lim _{\varepsilon \rightarrow 0} {\mathbf {E}}[ X^{i,\varepsilon }(t) X^{j,\varepsilon }(s)] \). Let \(m= \min (m_i,m_j)\) we use
to obtain, for \(s \le t\),
By Lemma 2.4, we obtain, for \(\varepsilon \rightarrow 0\),
Hence,
proving the finite chaos case. We now prove that the correlations of the limit converge as \( M \rightarrow \infty \). Indeed,
As \(G_{i,M} \rightarrow G_i\) in \(L^2(\mu )\), and similarly for j, this proves the joint convergence of the high Hermite rank components.
Step 2, CLT in the pure Hermite case
In this step, we focus on the vector component whose entries satisfy \(H^*(m_k)> \frac{1}{2}\). Recall, this implies \({H> \frac{1}{2}}\). By Lemma 4.2, evaluations of each component of \((X^{n+1}, \dots , X^N)\) converge in \(L^2(\Omega )\). Hence, they converge as well jointly in \(L^2(\Omega )\). Now, choose finitely many points \(t_{k,j} \in [0,T]\) and constants \( a_{k,j} \in {\mathbf {R}}\), then, \(\sum _{k,j} a_{k,j} X^{k,\varepsilon }_{t_{k,j}}\) converges in \(L^2(\Omega )\) to \(\sum _{k,j} a_{k,j} X^{k}_{t_{k,j}}\) and thus we may conclude joint convergence in finite dimensional distributions by an application of the Cramer-Wold theorem.
Step 3, Joint convergence
We have already shown that \(X^{W,\varepsilon } \rightarrow X^W\) and \(X^{Z,\varepsilon } \rightarrow X^Z\) in finite dimensional distributions, it is only left to prove their joint convergence. By Lemma 4.3 and Eq. (4.6), we may again reduce the problem to
Now, we can rewrite \(H_m(y^{\varepsilon }_s)= I_m(f^{m,\varepsilon }_s)\), where \(I_m\) denotes a m-fold Wiener-Itô integral and a function \(f^{m,\varepsilon }_s \in L^2({\mathbf {R}}^m,\mu )\). Now, for \(1 \le i \le n\) we obtain,
where
Similarly for \(j>n\),
Hence, we only need to show that the collection of stochastic processes of the form \(I_{m_k}(\hat{f}_t^{m_k,\varepsilon })\) converges jointly in finite dimensional distribution. It is thus sufficient to show that for every finite collection of times, \(t_{1}, \dots , t_Q \in [0,T]\), the vector, \(\left\{ I_{k}(\hat{f}^{k,\varepsilon }_{t_{l}}), k=m, \dots , M, l =1,\dots , Q \right\} \) converges jointly, where \(m = \min _{k=1, \dots , N} m_k\). Let \(n_0\) denote the smallest natural number such that \(H^*(n_0) < \frac{1}{2}\). For \(k > n_0\), the collection \(I_k(\hat{f}^{k,\varepsilon }_{t_{l}})\) converges to a normal distribution, hence, by the moment bounds in Lemma 4.6\( \Vert (k!) I_k(\hat{f}^{k,\varepsilon }_{t_{l}})\Vert _{H^{\otimes k}}= \sqrt{ {\mathbf {E}}\left( I_{k}(\hat{f}^{k,\varepsilon }_{t_l})^2\right) } \) converges to a constant. The convergence of such a sequence is equivalent to the convergence of the contractions, as defined in Eq. (5.1), of their kernels. Indeed, by a generalised fourth moment theorem [31, Theorem 1], we have
By Cauchy-Schwarz, we obtain for \(r =1, \dots , k_1\),
for all \( t_{l_1},t_{l_2} \in [0,T], \, 1 \le k_1 \le n_0 < k_2 \le M\). We can now apply an asymptotic independence result, Proposition 5.1 in Appendix, to conclude the joint convergence in finite dimensional distributions of \(X^\varepsilon \) to \((X^W,X^Z)\). Furthermore, \(X^W\) is independent of \(X^Z\).
The correlations between \(X^i_t\) and \(X^j_{t'}\), where \(i,j>n\), are 0 if \(m_i \not = m_j\), otherwise given by the \(L^2\) norm of their integrands, which follows from the Wiener-Itô isometry and are given by
Step 4, Convergence in the Hölder norms.
We first choose \(\gamma '\in \Big ( \gamma , (H^*(m_k) \wedge \frac{1}{2}) - \frac{1}{p_k}\Big )\). Then, using the Markov-Chebyshev inequality, we obtain that, as M tends to \( \infty \),
since \( \Vert |\theta |_{ {\mathcal {C}}^{\gamma '}}\Vert _{p_k} < \infty \) by Lemma 4.6 and an application of Kolmogorov’s theorem 2.6. Furthermore, since \({\mathcal {C}}^{\gamma '([0,T];{\mathbf {R}})} \) is compactly embedded in \({\mathcal {C}}^\gamma ([0,T];{\mathbf {R}})\) for \(1>\gamma ' > \gamma \), by an type Arzela-Ascoli argument, the sets \( \{ X^{k,\varepsilon }_t : \vert X^{k,\varepsilon } \vert _{{\mathcal {C}}^{\gamma '}([0,T],{\mathbf {R}})} \le M \} \) are sequentially compact in \({\mathcal {C}}^{\gamma }([0,T],{\mathbf {R}})\). Hence, \(\{X^{k,\varepsilon }_t\)} is tight in \({\mathcal {C}}^{\gamma }([0,T],{\mathbf {R}})\). See, e.g. [17]. As tightness in each component implies joint tightness, we obtain that \(X^{\varepsilon }\) is tight in \({\mathcal {C}}^{\gamma }([0,T],{\mathbf {R}}^N)\), where \(\gamma \in (0, \frac{1}{2} - \frac{1}{\min _{k \le n} p_k}) \) in case \(0<n\) and \(\gamma \in ( 0, \min _{k>n} H^*(m_k) - \frac{1}{p_k} ) \) otherwise. By the above discussion, any limit of a converging subsequence has the same finite dimensional distributions, hence, every subsequence converges to the same limit. This concludes the proof for the convergence in the Hölder norm. \(\square \)
5 Appendix: Joint Convergence by Asymptotic Independence
For the proof in the previous section, we need the following which modifies results from [33] and [29]. Let \(I_p(f)\) denote the \(p^{th}\) iterated Itô-Wiener integral of a symmetric function f of p variables,
If \(f\in L^2({\mathbf {R}}^p)\) and \(g\in L^2({\mathbf {R}}^q)\) are symmetric functions and \(p, q\ge 1\), their \(r^{th}\)-contraction is given by
where \(r \le p \wedge q\). If \(f\otimes _1g=\int _{{\mathbf {R}}} f(x_1, \dots , x_{p-1}, s) g(y_1, \dots , y_{q-1}, s) ds\) vanishes, so do all higher order contractions.
Proposition 5.1
Let \(q_1 \le q_2 \le \dots \le q_n \le p_1 \le p_2 \le \dots \le p_m\). Let \(f_i^\varepsilon \in L^2({\mathbf {R}}^{p_i})\), \( g_i^\varepsilon \in L^2({\mathbf {R}}^{q_i})\), \(F^{\varepsilon }=\left( I_{p_1}(f^{\varepsilon }_1), \dots , I_{p_m}(f^{\varepsilon }_m)\right) \) and \(G^{\varepsilon }=\left( I_{q_1}(g^{\varepsilon }_1), \dots , I_{q_n}(g^{\varepsilon }_n)\right) \). Suppose that for every i, j, and any \(1 \le r \le q_i\):
Then \(F^\varepsilon \rightarrow U\) and \(G^{\varepsilon } \rightarrow V\) weakly imply that \((F^{\varepsilon },G^{\varepsilon }) \rightarrow (U,V)\) jointly, where U and V are taken to be independent random variables.
These results benefit from the insights of Üstünel-Zakai [41] on the independence of two iterated integrals \(I_p(f)\) and \(I_q(g)\). They are independent if and only if the 1-contraction between f and g vanishes almost surely with respect to the Lebesgue measure. An asymptotic independence result follows as below,
Lemma 5.2
[29, Thm. 3.1] Let \(F^{\varepsilon }=I_p(f^{\varepsilon })\) and \(G^{\varepsilon }=I_q(g^{\varepsilon })\), where \(f^\varepsilon \in L^2({\mathbf {R}}^p)\) and \(g^\varepsilon \in L^2({\mathbf {R}}^q)\) . Then,
is equivalent to \( \Vert f^{\varepsilon } \otimes _r g^{\varepsilon }\Vert \rightarrow 0\), for \(1\le r \le p \wedge q\).
It is also known that if two integrals \(I_p(f)\) and \(I_q(g)\) are independent, then their Malliavin derivatives are orthogonal, see [41]. This explains why Malliavin calculus comes into prominent play, which has been developed to its perfection in [29, Lemma 3.2]. Given a smooth test function \(\phi \) we define,
where the sum runs over multi-indices \(k=(k_1, \dots , k_m)\). Let \(L=-\delta D\) and throughout this section \(f_i:{\mathbf {R}}^{p_i}\rightarrow {\mathbf {R}}\) and \(g:{\mathbf {R}}^q\rightarrow {\mathbf {R}}\) denote symmetric functions.
Lemma 5.3
[29] Let \(q\le p_i\), \(g\in L^2({\mathbf {R}}^{q})\), \(G=I_q(g)\), \(f_i\in L^2({\mathbf {R}}^{p_i})\), and \(F_i= I_{p_i}(f_i)\) with \({\mathbf {E}}(F_i^2) = 1\). Set \(F=(F_1, \dots , F_m)\) and let \(\theta \) be a smooth test function. Then,
where c is a constant depending on \(\Vert F \Vert _{L^2}\), \(\Vert G \Vert _{L^2} \), and q, m, \(p_1, \dots , p_m\).
The final piece of the puzzle is the observation that the defect in being independent is quantitatively controlled by the covariance of the squares of the relative components. The following is from [29], our only modification is to take G to be vector valued. Let \(g_i:{\mathbf {R}}^{q_i}\rightarrow {\mathbf {R}}\) be symmetric functions.
Lemma 5.4
Given \(F=\left( I_{p_1}(f_1), \dots I_{p_m}(f_m)\right) \) and \(G=\left( I_{q_1}(g_1), \dots , I_{q_n}(g_n)\right) \) such that \(p_k \ge q_l\) for every pair of k, l. Then, for all test functions \(\phi \) and \( \psi \), the following holds for some constant c, depending on \(\Vert F \Vert _{L^2}\), \(\Vert G \Vert _{L^2} \), and m, n, \(p_1, \dots , p_m\), \(q_1, \dots , q_n\),
Proof
Define \(L^{-1} (\sum _{k=0}^\infty I_k(h_m))=\sum _{k=1}^\infty \frac{1}{k} I_k(h_m) \in {\mathbb D}^{2,2}\). The key equality is \(-DL^{-1}=(I-L)^{-1}D\). As in [29],
Multiplying both sides by \(\psi (G)\), taking expectations and using integration by parts we obtain
To conclude, apply to each summand Lemma 5.3 with \(\theta = \partial _j \phi \) and \(G=G_i\). \(\square \)
Lemma 5.5
Let \(F^{\varepsilon }=\left( I_{p_1}(f^{\varepsilon }_1), \dots I_{p_m}(f^{\varepsilon }_m)\right) \) and \(G^{\varepsilon }=\left( I_{q_1}(g^{\varepsilon }_1), \dots , I_{q_n}(g^{\varepsilon }_n)\right) \) with \(q_1 \le q_2, \dots , q_n \le p_1 \le p_2\le \dots \le p_m\). Then for every \(i \le m ,j \le n\),
implies that for any smooth test functions \(\phi \) and \(\psi \),
Proof
Just combine Lemmas 5.4 and 5.2. \(\square \)
Finally, we finish the proof of Proposition 5.1.
Proof
Since \((F^{\varepsilon },G^{\varepsilon })\) is bounded in \(L^2(\Omega )\) it is tight. Now choose a weakly converging subsequence \((F^n,G^n)\) with limit denoted by (X, Y). Let \(\phi \) and \(\psi \) be smooth test functions, then by Lemma 5.5 and the bounds on \(\phi , \psi \), we pass to the limit under the expectation sign and obtain
Thus, every limit measure is the product measure determined by U and V, hence, \((F^{\varepsilon },G^{\varepsilon })\) converges as claimed. \(\square \)
References
Al-Talibi, H., Hilbert, A.: Differentiable approximation by solutions of Newton equations driven by fractional Brownian motion. Preprint (2012)
Buchmann, B., Chan, N.H.: Integrated functionals of normal and fractional processes. Ann. Appl. Probab. 19(1), 49–70 (2009)
Bourguin, S., Gailus, S., Spiliopoulos, K.: Typical dynamics and fluctuation analysis of slow-fast systems driven by fractional Brownian motion (2019)
Hariz, S.B.: Limit theorems for the non-linear functional of stationary Gaussian processes. J. Multivariate Anal. 80(2), 191–216 (2002)
Billingsley, P.: Convergence of probability measures. Wiley Series in Probability and Statistics: Probability and Statistics. Wiley, New York, second edition (1999)
Breuer, P., Major, P.: Central limit theorems for nonlinear functionals of Gaussian fields. J. Multivariate Anal. 13(3), 425–441 (1983)
Boufoussi, B., Tudor, C.A.: Kramers-Smoluchowski approximation for stochastic evolution equations with FBM. Rev. Roumaine Math. Pures Appl. 50(2), 125–136 (2005)
Bai, S., Taqqu, M.S.: Multivariate limit theorems in the context of long-range dependence. J. Time Series Anal. 34(6), 717–743 (2013)
Cornfeld, I. P., Fomin, S. V., Sinai, Ya. G.: Ergodic theory volume 245 of Grundlehren der Mathematischen Wissenschaften [Fundamental Principles of Mathematical Sciences]. Springer, New York. Translated from the Russian by A. B. Sosinskii (1982)
Cheridito, P., Kawaguchi, H., Maejima, M.: Fractional Ornstein-Uhlenbeck processes. Electron. J. Probab. 8(3), 14 (2003)
Campese, S., Nourdin, I., Nualart, D.: Continuous Breuer-major theorem: tightness and nonstationarity. Ann. Probab. 48(1), 147–177 (2020)
Dobrushin, R.L.: Gaussian and their subordinated self-similar random generalized fields. Ann. Probab. 7(1), 1–28 (1979)
Embrechts, P., Maejima, M.: Selfsimilar Processes. Princeton Series in Applied Mathematics. Princeton University Press, Princeton, NJ (2002)
Friz, P., Gassiat, P., Lyons, T.: Physical Brownian motion in a magnetic field as a rough path. Trans. Am. Math. Soc. 367(11), 7939–7955 (2015)
Friz, P. K., Hairer, M.: A Course on Rough Paths. Universitext. Springer, Cham, With an introduction to regularity structures (2014)
Fannjiang, A., Komorowski, T.: Fractional Brownian motions in a limit of turbulent transport. Ann. Appl. Probab. 10(4), 1100–1120 (2000)
Friz, P. K., Victoir, N. B.: Multidimensional stochastic processes as rough paths, volume 120 of Cambridge Studies in Advanced Mathematics. Cambridge University Press, Cambridge, Theory and applications (2010)
Gehringer, J., Li, X.-M.: Homogenization with fractional random fields. arXiv:1911.12600. This is now improved and split into ‘Functional limit theorem for fractional OU’ and Diffusive and rough homogenisation in fractional noise field (2019)
Gehringer, J., Li, X.-M.: Diffusive and rough homogenisation in fractional noise field. This is an improved version of arXiv:1911.12600 (part II) (2020)
Hurst, H.E., Black, R.P., Simaika, Y.M.: Long-term storage: an experimental study. Constable, London (1965)
Hairer, M., Li, X.-M.: Averaging dynamics driven by fractional Brownian motion. Ann. Probab. 48(4), 1826–1860 (2020)
Jona-Lasinio, G.: Probabilistic approach to critical behavior. In: New developments in quantum field theory and statistical mechanics (Proc. Cargèse Summer Inst., Cargèse, 1976), pp. 419–446. NATO Adv. Study Inst. Ser., Ser. B: Physics, 26 (1977)
Komorowski, T., Landim, C., Olla, S.: Fluctuations in Markov processes, volume 345 of Grundlehren der Mathematischen Wissenschaften [Fundamental Principles of Mathematical Sciences]. Springer, Heidelberg, Time symmetry and martingale approximation (2012)
Lyons, T. J., Caruana, M., Lévy, T.: Differential equations driven by rough paths, volume 1908 of Lecture Notes in Mathematics. Springer, Berlin, Lectures from the 34th Summer School on Probability Theory held in Saint-Flour, July 6–24, 2004, With an introduction concerning the Summer School by Jean Picard (2007)
Lyons, T.: Differential equations driven by rough signals. I. An extension of an inequality of LC Young. Math. Res. Lett. 1(4), 451–464 (1994)
Mishura, Y.S.: Stochastic Calculus for Fractional Brownian Motion and Related Processes. Lecture Notes in Mathematics, vol. 1929. Springer, Berlin (2008)
Maejima, M., Tudor, C.A.: Wiener integrals with respect to the Hermite process and a non-central limit theorem. Stoch. Anal. Appl. 25(5), 1043–1056 (2007)
Mandelbrot, B.B., Van Ness, J.W.: Fractional Brownian motions, fractional noises and applications. SIAM Rev. 10, 422–437 (1968)
Nourdin, I., Nualart, D., Peccati, G.: Strong asymptotic independence on Wiener chaos. Proc. Am. Math. Soc. 144(2), 875–886 (2016)
Nourdin, I., Nualart, D., Zintout, R.: Multivariate central limit theorems for averages of fractional Volterra processes and applications to parameter estimation. Stat. Inference Stoch. Process. 19(2), 219–234 (2016)
Nualart, D., Peccati, G.: Central limit theorems for sequences of multiple stochastic integrals. Ann. Probab. 33(1), 177–193 (2005)
Nourdin, I., Peccati, G.: Normal approximations with Malliavin calculus, volume 192 of Cambridge Tracts in Mathematics. Cambridge University Press, Cambridge, From Stein’s method to universality (2012)
Nourdin, I., Rosinski, J.: Asymptotic independence of multiple wiener-itô integrals and the resulting limit laws. Ann. Probab. 42(2), 497–526 (2014)
Pipiras, V., Taqqu, M.S.: Integration questions related to fractional Brownian motion. Probab. Theory Related Fields 118(2), 251–291 (2000)
Pipiras, V., Taqqu, M.S.: Long-Range Dependence and Self-Similarity. Cambridge Series in Statistical and Probabilistic Mathematics, vol. 45. Cambridge University Press, Cambridge (2017)
Rosenblatt, M.: Independence and dependence. In: Proc. 4th Berkeley Sympos. Math. Statist. and Prob., Vol. II, pp. 431–443. Univ. California Press, Berkeley, Calif (1961)
Samorodnitsky, G.: Long range dependence. Found. Trends Stoch. Syst. 1(3), 163–257 (2006)
Sinaı, Y.G.: Self-similar probability distributions. Theory Probab. Appl 21(1), 63–80 (1976)
Stroock, D. W., Srinivasa Varadhan,S. R.: Multidimensional Diffusion Processes. Classics in Mathematics. Springer, Berlin, Reprint of the 1997 edition (2006)
Taqqu, M.S.: Convergence of integrated processes of arbitrary Hermite rank. Z. Wahrsch. Verw. Gebiete 50(1), 53–83 (1979)
Üstünel, A.S., Zakai, M.: On independence and conditioning on Wiener space. Ann. Probab. 17(4), 1441–1453 (1989)
Young, L.C.: An inequality of the Hölder type, connected with Stieltjes integration. Acta Math. 67(1), 251–282 (1936)
Zhang, S.: Smoluchowski-Kramers approximations for stochastic equations with Lévy noise. PhD thesis, Purdue University (2008)
Acknowledgements
This research is partially funded by an EPSRC Roth studentship.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Gehringer, J., Li, XM. Functional Limit Theorems for the Fractional Ornstein–Uhlenbeck Process. J Theor Probab 35, 426–456 (2022). https://doi.org/10.1007/s10959-020-01044-7
Received:
Revised:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10959-020-01044-7
Keywords
- Passive tracer
- Fractional noise
- Multi-scale
- Mixed functional central and non-central limit theorems
- Rough creation
- Rough homogenization
- Rough topology