Abstract
We prove the almost sure invariance principle (ASIP) with close to optimal error rates for nonuniformly hyperbolic maps. We do not assume exponential contraction along stable leaves, therefore our result covers in particular slowly mixing invertible dynamical systems as Bunimovich flowers, billiards with flat points as in Chernov and Zhang (Stoch Dyn 5:535–553, 2005a, Nonlinearity 18:1527–1553, 2005b) and Wojtkowski’ (Commun Math Phys 126:507–533, 1990) system of two falling balls.For these examples, the ASIP is a new result, not covered by prior works for various reasons, notably because in absence of exponential contraction along stable leaves, it is challenging to employ the so-called Sinai’s trick (Sinai in Russ Math Surv 27:21–70, 1972; Bowen, Lecture Notes in Math vol. 470 (1975)) of reducing a nonuniformly hyperbolic system to a nonuniformly expanding one. Our strategy follows our previous papers on the ASIP for nonuniformly expanding maps, where we build a semiconjugacy to a specific renewal Markov shift and adapt the argument of Berkes et al. (Ann Probab 42:794–817, 2014). The main difference is that now the Markov shift is two-sided, the observables depend on the full trajectory, both the future and the past.
Avoid common mistakes on your manuscript.
1 Introduction
Let \(T :M \rightarrow M\) be a nonuniformly hyperbolic dynamical system in the sense of Young [32]. Notable examples of such systems are Axiom A or Hénon-type attractors and various chaotic billiards. Suppose that T preserves an ergodic physical (Sinai–Ruelle–Bowen) probability measure m. Let \(\varphi :M \rightarrow {\mathbb {R}}\) be a Hölder observable with \(\int \varphi \, dm = 0\) and let \(S_n = \sum _{j =0}^{n-1} \varphi \circ T^j\) be its corresponding Birkhoff sums.
It is common to expect that \(S_n\), considered as a random process on the probability space (M, m), behaves like a Brownian motion. A standard result, proved for a large class of nonuniformly hyperbolic maps under nonrestrictive assumptions, is the Almost Sure Invariance Principle (ASIP), see Melbourne and Nicol [23] or Gouëzel [17]. It holds if, without changing the distribution, the process \((S_n)_{n \ge 0}\) can be defined on a probability space supporting a Brownian motion \((W_t)_{t \ge 0}\) with variance \(c^2\), such that with some \(\delta \in (0, 1/2)\),
In this case we say that the process \(S_n\) satisfies the ASIP with rate \(o(n^\delta )\) and variance \(c^2\).
The ASIP has a range of useful implications, such as the (functional) central limit theorem or the (functional) law of iterated logarithm, see Berkes and Philipp [4]. The error rates received a significant amount of attention, see e.g. our paper [11] for an overview. See also Zaitsev [33] for a historical overview for sums of independent random vectors.
There are natural nonuniformly hyperbolic dynamical systems where the ASIP is expected but is not covered by existing proofs for one reason or another. Prime examples are Bunimovich flowers [7] and Wojtkowski’ system of two falling balls [31].
The nonuniformly hyperbolic structure for these maps, under certain natural assumptions, is established by Bálint, Borbély and Varga [1] and Chernov and Zhang [9]. Although the ASIP has not been proven, other statistical properties such as the functional central limit theorem or (iterated) moment bounds are known, see Fleming-Vázquez [15] and Melbourne and Varandas [24].
Another prototypical example is the intermittent baker’s map [24, Example 4.1]. It is defined as a transformation of the unit square \(M = [0,1] \times [0,1]\) by
where \(g(x) = x ( 1 + 2^\alpha x^\alpha )\) and \(\alpha \in (0,1)\).
In this paper we prove the ASIP for a class of nonuniformly hyperbolic maps, including (1.2) and (1.3), with rates close to optimum (see Theorem 3.1). Applied to the above examples, our main result is:
Theorem 1.1
Suppose that \(T :M \rightarrow M\) is either (1.3) or one of the maps (1.2) under the assumptions of [1, 9]. Let m be the unique physical invariant probability measure. Let \(v :M \rightarrow {\mathbb {R}}\) be Hölder with \(\int v \, dm = 0\) and \(S_n = \sum _{j < n} v \circ T^j\). Then, for each \({\varepsilon }> 0\),
-
(a)
\(S_n\) satisfies the ASIP with rate \(o(n^{1/3} (\log n)^{4/3 + {\varepsilon }})\) for both maps (1.2). (Possibly, the logarithmic factor can be reduced, see Theorem 3.1 and Remark 3.3.)
-
(b)
\(S_n\) satisfies the ASIP with rate \(o(n^\alpha (\log n)^{\alpha + {\varepsilon }})\) for the map (1.3) when \(\alpha < 1/2\).
Our proofs are based on:
- \(\bullet \):
-
A construction of an extension of T, which is similar to a Young tower in the sense that it is (topologically) Markov with a tower structure and that the semiconjugacy map is Lipschitz, but with the difference that our extension is also Markov from the measure-theoretical point of view (i.e. can be described by a matrix of transition probabilities).
- \(\bullet \):
-
A representation of the process \(( v \circ T^j)_{ j \ge 0}\), without changing its distribution (but on another probability space), as
$$\begin{aligned} \big ( \psi ( \ldots , g_{k-1}, g_k, g_{k+1}, \ldots ) \big )_{k \in {\mathbb Z}} , \end{aligned}$$(1.4)where \((g_n)\) is a particular Markov chain and \(\psi \) is a sufficiently regular function of its trajectories. A key tool in the study of the Markov chain is the meeting time which is, informally, the time when two independent copies of the Markov chain first meet. Moments of the meeting time are used to control approximations of \(\psi ( \ldots , g_{j-1}, g_j, g_{j+1}, \ldots )\) by functions of finitely many coordinates.
- \(\bullet \):
-
An adapted argument of Berkes, Liu and Wu [3] which allows to deal with functions of the whole trajectory of a Markov chain instead of a sequence of independent identically distributed random variables as in [3].
This is a continuation and extension of our previous works on nonuniformly expanding maps [11, 12].
Remark 1.2
As in [11, 12], our proof of ASIP happens on the level of a Markov chain (1.4). At the same time, in [11, 12] we had to deal with functions \(\psi \) which depend only on future trajectories, i.e. \(\sum _{j=0}^{n-1} \psi (g_j, g_{j+1}, \ldots )\). Working with nonuniformly hyperbolic dynamical systems, we necessarily have to deal with functions of the whole past and future trajectories. This creates problems which go beyond simple modifications of our previous arguments both in the reduction to the Markov chain (Sect. 2) and in the probabilistic part (Sect. 3). This can be contrasted with the proof of the ASIP in [3] which is written for a one-sided Bernoulli shift rather than for a two-sided one “to simplify the notation”.
Remark 1.3
It is curious that the degenerate case of zero variance is not covered by our general arguments and requires a special and very different treatment (Sect. 3.6).
Remark 1.4
There are prior ideas which may be useful in finding alternative proofs of the ASIP for dynamical systems such as (1.2) and (1.3), including the generalization to vector-valued observables. Notably:
- \(\bullet \):
-
Melbourne and Nicol [22, 23] prove the ASIP for vector-valued Hölder observables on nonuniformly hyperbolic dynamical systems with uniform contraction along stable leaves. Their proofs work using the so-called Sinai’s trick (after Sinai [30] and Bowen [6], see [24, Introduction]), representing a Hölder observable \(v :M \rightarrow {\mathbb {R}}\) as \(v = u + \chi \circ T - \chi \), where \(\chi \) is bounded and u is Hölder and constant on stable leaves. Then one can work with Birkhoff sums of u on a quotient system which is nonuniformly expanding, and can be in turn reduced to a Gibbs–Markov map by inducing. For the ASIP, Melbourne and Nicol adapt the results of Philipp and Stout [26] and Kuelbs and Philipp [20]. In the absence of uniform contraction along stable leaves, as is the case for the maps (1.2) and (1.3), Sinai’s trick works but \(\chi \) and u lose regularity and become challenging to control. Nevertheless, some control is possible. Melbourne and Varandas [24] prove that \(\chi ,u \in L^p\) with some \(p > 1\), and furthermore \(u = m + \psi \circ T - \psi \) where again \(m,\psi \in L^p\) and Birkhoff sums of m are a reverse martingale.
- \(\bullet \):
-
Gouëzel [17] proves the ASIP for vector-valued random variables under a condition of certain exponential memory loss for characteristic functions. This condition does not hold for the maps (1.2) or (1.3), but for instance Gouëzel [17, Theorem 2.4] uses it to prove the ASIP for a class of \(L^p\) observables on Gibbs–Markov maps, and as a corollary on nonuniformly expanding maps via inducing.
However, using the above techniques, the error rates would be \(O(n^{-1/4} (\log n)^\alpha )\) for some \(\alpha >0\) at best, whereas our approach is the only currently available one with better rates.
The paper is organized as follows. In Sect. 2 we state the assumptions on the nonuniformly hyperbolic maps and prove that they can be modelled using Markov chains of a particular type. We call such systems Markov shift towers. In Sect. 3 we prove the ASIP for Hölder observables on nonuniformly hyperbolic maps through that on Markov shift towers. Theorem 1.1 is a corollary of the more general Theorem 3.1.
We use the notation \({\mathbb {N}}= \{0,1,\ldots \}\) and \({\mathbb {N}}_0 = \{1,2,\ldots \}\).
2 Nonuniformly Hyperbolic Maps as Markov Shift Towers
In this section we give a standard definition of nonuniformly hyperbolic maps, introduce Markov shift towers and show that every nonuniformly hyperbolic map \(T :M \rightarrow M\) with invariant measure m has an extension \(f :\Delta \rightarrow \Delta \) with invariant measure \({{\,\mathrm{\mathbb {P}}\,}}\) which is a Markov shift tower, and the semiconjugacy \(\pi :\Delta \rightarrow M\) is Lipschitz.
We also quantify the return time tails of f depending on those of T. This result is stated formally in Theorem 2.4, but first we need a few pages of notation.
2.1 Notation and result
2.1.1 Nonuniformly hyperbolic maps
Let \((M,d_0)\) be a bounded metric space with a Borel probability measure m such that (M, m) is a Lebesgue space. Let \(T :M \rightarrow M\) be measurable for the Borel sigma-field. We assume that T preserves m, is ergodic and is nonuniformly hyperbolic in the following sense:
- \(\bullet \):
-
There is a measurable subset \(Y \subset M\), \(m(Y) >0\), with an at most countable partition \(\alpha \) and a return time \(\tau :Y \rightarrow \{1,2,\ldots \}\) that is constant on each \(a \in \alpha \) with value \(\tau (a)\) and \(T^{\tau (y)}(y) \in Y\) for all \(y \in Y\). We denote by \(F :y \mapsto T^{\tau (y)}(y)\) the induced map. There is an F-invariant probability measure \(\mu \) on Y such that \(\int \tau \, d\mu < \infty \) and which agrees with m in the usual for induced maps sense:
$$\begin{aligned} m = \Bigl ( \int \tau \, d\mu \Bigr )^{-1} \sum _{a \in \alpha } \sum _{k=0}^{\tau (a) - 1} (T^k)_* \mu _a , \end{aligned}$$where \(\mu _a\) is the restriction of \(\mu \) on a.
- \(\bullet \):
-
Coding of orbits under F by elements of \(\alpha \) is non-pathological in the sense that the set of \((\ldots , a_{-1}, a_0, a_1, \ldots ) \in \alpha ^{\mathbb {Z}}\), for which there exists \((\ldots , y_{-1}, y_0, y_1, \ldots ) \in Y^{\mathbb {Z}}\) such that \(y_n \in a_n\) and \(F(y_n) = y_{n+1}\), is measurable in \(\alpha ^{\mathbb {Z}}\) with the product topology and Borel sigma-algebra (\(\alpha \) is equipped with the discrete topology).
- \(\bullet \):
-
For \(y,z \in Y\) we define the separation time s(y, z) as the least \(n \ge 0\) such that \(F^n(y)\) and \(F^n(z)\) belong to different elements of \(\alpha \), or \(\infty \) if such n does not exist. There are constants \(K\ge 1\) and \(\gamma \in (0,1)\) such that:
- \(\bullet \):
-
If \(y,z \in Y\), then for all \(j \ge 0\),
$$\begin{aligned} d_0(F^j(y),F^j(z)) \le K (\gamma ^{s(y,z) - j} + \gamma ^{j}) . \end{aligned}$$(2.1) - \(\bullet \):
-
If \(y,z \in a\), \(a \in \alpha \), then for all \(0 \le j \le \tau (a)\),
$$\begin{aligned} d_0(T^j y,T^j z) \le K \bigl [d_0(y,z) + d_0(F(y), F(z)) \bigr ] \end{aligned}$$(2.2) - \(\bullet \):
-
There is a partition \({\mathcal {W}}^s\) of Y into “stable leaves”. The stable leaves are invariant under F meaning that \(F(W^s_y) \subset W^s_{F(y)}\) for all \(y \in Y\), where \(W^s_y\) is the stable leaf containing y. Each \(a \in \alpha \) is a union of stable leaves, i.e. \({\mathcal {W}}_s\) is a refinement of \(\alpha \). Let \({\bar{Y}}= Y / \sim \), where \(y \sim z\) if \(y \in W^s_z\), and let \({\bar{\pi }}:Y \rightarrow {\bar{Y}}\) be the natural projection. Since the stable leaves are invariant under F and \({\mathcal {W}}^s\) is a refinement of \(\alpha \), we obtain a well defined quotient map \({\bar{F}}:{\bar{Y}}\rightarrow {\bar{Y}}\) with a partition \({\bar{\alpha }}\) of \({\bar{Y}}\) and separation time s. We suppose that the probability measure \({\bar{\mu }}= {\bar{\pi }}_* \mu \) on \({\bar{Y}}\) is invariant under \({\bar{F}}\). Moreover, for each \(a \in {\bar{\alpha }}\), the restriction \({\bar{F}}:a \rightarrow {\bar{Y}}\) is a bijection (modulo \({\bar{\mu }}\) zero measure), and its inverse Jacobian \( \displaystyle \zeta _a= \frac{d{\bar{\mu }}}{d{\bar{\mu }}\circ {\bar{F}}}\) satisfies
$$\begin{aligned} \bigl |\log \zeta _a(y)-\log \zeta _a(z)\bigr | \le K\gamma ^{s(y,z)} \quad \text {for all } y,z \in a . \end{aligned}$$In other words, the quotient map \({\bar{F}}\) is full branch Gibbs–Markov.
2.1.2 Markov shift towers
A closely related class of nonuniformly hyperbolic maps is what we call a Markov shift tower. The main difference is that the induced map, denoted by \(f_X\) below, is a Bernoulli shift, both topologically and measure-theoretically.
We say that \(f :\Delta \rightarrow \Delta \) is a Markov shift tower if it has the following structure.
- \(\bullet \):
-
We are given a finite or countable probability space \(({\mathcal {A}}, {{\,\mathrm{\mathbb {P}}\,}}_{\mathcal {A}})\), an integrable function \(h_{\mathcal {A}}:{\mathcal {A}}\rightarrow \{1,2,\ldots \}\) and a constant \(\xi \in (0, 1)\). These define the rest of the construction.
- \(\bullet \):
-
Let \((X, {{\,\mathrm{\mathbb {P}}\,}}_X) = ({\mathcal {A}}^{\mathbb {Z}}, {{\,\mathrm{\mathbb {P}}\,}}_{\mathcal {A}}^{\mathbb {Z}})\) be the product probability space and let \(f_X :X \rightarrow X\) be the left shift
$$\begin{aligned} f_X(\ldots , a_{-1}, a_0, a_1, \ldots ) = (\ldots , a_0, a_1, a_2, \ldots ) . \end{aligned}$$Define \(h :X \rightarrow \{1,2,\ldots \}\) by \(h(\ldots , a_{-1}, a_0, a_1, \ldots ) = h_{\mathcal {A}}(a_0)\).
- \(\bullet \):
-
The map \(f :\Delta \rightarrow \Delta \) is a suspension over \(f_X :X \rightarrow X\) with a roof function h, i.e.
$$\begin{aligned} \begin{aligned} \Delta&= \{(x,\ell ) \in X \times {\mathbb {Z}}: 0 \le \ell< h(x)\} \\ f(x,\ell )&= {\left\{ \begin{array}{ll} (x, \ell + 1), &{} \quad \ell < h(x) - 1 \\ (f_X (x), 0), &{} \quad \ell = h(x) - 1 \end{array}\right. } . \end{aligned} \end{aligned}$$(2.3) - \(\bullet \):
-
Let \({{\,\mathrm{\mathbb {P}}\,}}\) be the probability measure on \(\Delta \) defined by
$$\begin{aligned} {{\,\mathrm {\mathbb {P}}\,}}(A \times \{\ell \})=&{} \biggl ( \int h \, d {{\,\mathrm {\mathbb {P}}\,}}_X \biggr )^{-1} {{\,\mathrm {\mathbb {P}}\,}}_X(A) \quad \text{ for } \text{ all } \quad \ell \ge 0 \quad \text{ and } \quad \\ {}{}&{} \quad A \subset \{x \in X : h(x) \ge \ell + 1\} . \end{aligned}$$Since \({{\,\mathrm{\mathbb {P}}\,}}_X\) is \(f_X\)-invariant, note that \({{\,\mathrm{\mathbb {P}}\,}}\) is \(f\)-invariant.
- \(\bullet \):
-
Define a distance \(d\) on \(X\) by \(d(x,y) = \xi ^{s(x,y)}\), where \(s :X \times X \rightarrow \{0,1,\ldots \}\) is the separation time,
$$\begin{aligned}{} & {} s((\ldots , a_{-1}, a_0, a_1, \ldots ), (\ldots , b_{-1}, b_0, b_1, \ldots ) ) \\{} & {} \qquad = \inf \{ j \ge 0 : a_j \ne b_j \text { or } a_{-j} \ne b_{-j} \} . \end{aligned}$$Let \(d\) also denote the related distance on \(\Delta \):
$$\begin{aligned} d((x,k),(y,j)) = {\left\{ \begin{array}{ll} 1, &{} \quad k \ne j \\ d(x,y), &{} \quad k=j \end{array}\right. }. \end{aligned}$$(2.4)
Remark 2.1
As a part of the definition of a Markov shift tower, \(f :\Delta \rightarrow \Delta \) is Markov both topologically (as a suspension over a Bernoulli shift, it has a natural Markov partition) and probabilistically (the invariant measure \({{\,\mathrm{\mathbb {P}}\,}}\) is Markov, it can be described by transition probabilities). Also \(\Delta \) is a metric space with metric d.
Remark 2.2
We used similar Markov shift towers in our previous works [11, 12, 19]. The main difference is that they were one-sided: the base map was the left shift \(f_X :X \rightarrow X\) on \(X = {\mathcal {A}}^{\mathbb {N}}\) instead of \(X = {\mathcal {A}}^{\mathbb {Z}}\) here. A two-sided Markov shift tower is the natural extension [28, 29] of the corresponding one-side tower.
Remark 2.3
An important characteristic of both nonuniformly hyperbolic maps and Markov shift towers is tail of the return times, i.e. the asymptotics of \(\mu (\tau \ge n)\) and \({{\,\mathrm{\mathbb {P}}\,}}_X(h \ge n)\) respectively.
2.1.3 Main result
Theorem 2.4
Suppose that \(T :M \rightarrow M\) is a nonuniformly hyperbolic map as in Sect. 2.1.1. Then there exists a Markov shift tower \(f :\Delta \rightarrow \Delta \) and a map \(\pi :\Delta \rightarrow M\), defined \({{\,\mathrm{\mathbb {P}}\,}}\)-almost surely on \(\Delta \), such that:
-
\(\pi \) is a semiconjugacy, i.e. \(\pi \circ f = T \circ \pi \),
-
\(\pi \) is measure preserving, i.e. \(\pi _* {{\,\mathrm{\mathbb {P}}\,}}= m\),
-
\(\pi \) is Lipschitz.
Moreover, the return time tails of f are closely related to those of T:
-
if \(\mu (\tau \ge n) = O(n^{-\beta } L(n) )\) with \(\beta > 0\) and L a slowly varying function at infinity, then \({{\,\mathrm{\mathbb {P}}\,}}_{\mathcal {A}}(h_{\mathcal {A}}\ge n) = O(n^{-\beta } L(n) )\);
-
if \(\int \tau ^\beta \, d\mu < \infty \) with \(\beta > 0\), then \(\int h_{\mathcal {A}}^\beta \, d{{\,\mathrm{\mathbb {P}}\,}}_{\mathcal {A}}< \infty \);
-
if \(\int e^{\beta \tau ^\delta } \, d\mu < \infty \) with \(\beta > 0\) and \(\delta \in (0,1]\), then \(\int e^{\beta ' h_{\mathcal {A}}^\delta } \, d{{\,\mathrm{\mathbb {P}}\,}}_{\mathcal {A}}< \infty \) for some \(0 < \beta ' \le \beta \).
2.2 Proof of Theorem 2.4
Our strategy is to reduce the proof to an application of [19, Theorem 3.4]. There, a one-sided Markov shift tower is constructed for a nonuniformly expanding map. Then T can be considered via the natural extension of such a map, and the corresponding natural extension of the one-sided Markov shift in [19] is a two-sided Markov shift, as required for Theorem 2.4.
Let \(\sigma :\alpha ^{\mathbb {Z}}\rightarrow \alpha ^{\mathbb {Z}}\) be the left shift. We supply \(\alpha ^{\mathbb {Z}}\) with the product topology and the Borel sigma-algebra. For \(x=(x_k)_{k \in {\mathbb {Z}}}\) and \(y=(y_k)_{k \in {\mathbb {Z}}}\) in \(\alpha ^{\mathbb {Z}}\), define the separation time and distance by
Here \(\gamma \) is the constant in (2.1).
2.2.1 Tower over full shift
Our first step is to model \(T :M \rightarrow M\) as a suspension tower over \(\sigma :\alpha ^{\mathbb {Z}}\rightarrow \alpha ^{\mathbb {Z}}\). It is natural that this can be done, yet this requires some care.
Proposition 2.5
There exists a probability measure \(\mu _\alpha \) on \(\alpha ^{\mathbb {Z}}\) and a \(\mu _\alpha \)-almost everywhere defined map \(\chi _\alpha :\alpha ^{\mathbb {Z}}\rightarrow Y\) such that on the domain of \(\chi _\alpha \):
-
\(\chi _\alpha \) is Lipschitz: \(d_0(\chi _\alpha (x), \chi _\alpha (y)) \le 2 K d(x,y)\),
-
\((\chi _\alpha )_* \mu _\alpha = \mu \),
-
\(F \circ \chi _\alpha = \chi _\alpha \circ \sigma \),
-
\(\chi _\alpha (\ldots , a_{-1}, a_0, a_1, \ldots ) \in a_0\).
Proof
In order to construct the natural extension [28, 29] of \(F :Y \rightarrow Y\), we assume that \(F(Y) = Y\). This comes without loss of generality: otherwise instead of Y we work with \(\cap _{n \ge 0} F^{-n}(Y)\). This is a full measure subset of Y on which F is surjective, and the proof goes through with straightforward and minor changes.
Let \({\breve{F}}:{\breve{Y}}\rightarrow {\breve{Y}}\) be the natural extension [28, 29] of \(F :Y \rightarrow Y\):
and \({\breve{F}}\) is the left shift. The topology on \({\breve{Y}}\) is generated by open cylinders of the type \(\{y_i \in E\}\) with open \(E \subset Y\) (in the induced topology on Y as a subspace of M), and we consider \({\breve{Y}}\) with the Borel sigma-algebra. Let \({\breve{\pi }}_\ell :{\breve{Y}}\rightarrow Y\), \((\ldots , y_{-1}, y_0, y_1, \ldots ) \mapsto y_\ell \) be the natural projections, set \({\breve{\pi }}= {\breve{\pi }}_0\), and let \({\breve{\mu }}\) be the unique \({\breve{F}}\)-invariant probability measure on \({\breve{Y}}\) such that \({\breve{\pi }}_* {\breve{\mu }}= \mu \).
Define \(\iota :{\breve{Y}}\rightarrow \alpha ^{\mathbb {Z}}\) by \((\ldots , y_{-1}, y_0, y_1, \ldots ) \mapsto (\ldots , a_{-1}, a_0, a_1, \ldots )\) where \(y_n \in a_n\). Then \(\iota \) is a measurable and injective map. Set \(\mu _\alpha = \iota _* {\breve{\mu }}\). Using (2.1), for \(0 \le |\ell | \le n\),
where \([a_{-n}, \ldots , a_n] \subset \alpha ^{\mathbb {Z}}\) is a cylinder. Hence \({\breve{\pi }}_\ell \circ \iota ^{-1}\) is Lipschitz with Lipschitz constant \(2 K \gamma ^{-|\ell |}\). Since \(\iota ^{-1}(x) = ( ({\breve{\pi }}_\ell \circ \iota ^{-1}) (x) )_{\ell \in {\mathbb {Z}}}\) and each \({\breve{\pi }}_\ell \circ \iota ^{-1}\) is continuous, \(\iota ^{-1} :\iota ({\breve{Y}}) \rightarrow {\breve{Y}}\) is continuous. Recall that, as a part of the definition of T as a nonuniformly hyperbolic map we assumed that \(\iota ({\breve{Y}})\) is measurable in \(\alpha ^{\mathbb {Z}}\). Then \(\iota ^{-1}\) is measurable.
Observe that \(\iota ^{-1}\) is a conjugacy between \({\breve{F}}\) and \(\sigma \). Set \(\chi _\alpha = {\breve{\pi }}\circ \iota ^{-1}\).
Then \(\chi _\alpha \) is a measure preserving semiconjugacy between \(\sigma \) and F with the required properties. \(\square \)
In the setup of Proposition 2.5, consider the suspension map over \(\sigma \) with roof function \(\tau _\alpha = \tau \circ \chi _\alpha \), i.e. the space
with the transformation
As in (2.4), we supply \(M_\alpha \) with the metric
Let \(m_\alpha = \mu _\alpha \times \text {counting} / \text {normalization}\) be the probability measure on \(M_\alpha \), and let \(\pi _\alpha :M_\alpha \rightarrow M\),
Remark 2.6
Observe that:
-
\(T_\alpha :M_\alpha \rightarrow M_\alpha \) is a nonuniformly hyperbolic map with the same tails of return times as T, namely \(\mu _\alpha (\tau _\alpha \ge n) = \mu (\tau \ge n)\) for all n.
-
\(\pi _\alpha \) is defined \(m_\alpha \)-almost surely, but possibly not on the whole \(M_\alpha \).
-
\(\pi _\alpha :M_\alpha \rightarrow M\) is a measure preserving semiconjugacy between \(T_\alpha \) and T (modulo zero measure).
-
\(\pi _\alpha :M_\alpha \rightarrow M\) is Lipschitz (on its domain). Indeed,
$$\begin{aligned} d_0(\pi _\alpha (x, \ell ), \pi _\alpha (y, k)) \le C d((x,\ell ), (y, k)) \end{aligned}$$holds:
-
with \(C = {{\,\textrm{diam}\,}}M\) when \(\ell \ne k\) by construction,
-
with \(C = 2K\) when \(\ell = k = 0\) by Proposition 2.5,
-
with \(C = 2 K^2 ( 1 + \gamma ^{-1})\) when \(\ell = k > 0\). Indeed, by (2.2),
$$\begin{aligned} d_0(\pi _\alpha (x, \ell ), \pi _\alpha (y, \ell )) \le K \big [ d_0 ( \chi _\alpha (x), \chi _\alpha (y)) + d_0 ( F \circ \chi _\alpha (x), F \circ \chi _\alpha (y)) \big ] \, . \end{aligned}$$Now, by Proposition 2.5, \(F \circ \chi _\alpha = \chi _\alpha \circ \sigma \) and \(\chi _\alpha \) is Lipschitz. Then
$$\begin{aligned} d_0(\pi _\alpha (x, \ell ), \pi _\alpha (y, \ell )) \le 2K^2 \big [ d ( x,y) + d ( \sigma (x), \sigma (y)) \big ] \, . \end{aligned}$$To conclude, we use the fact that since \(s(x,y) \le 1 + s(\sigma (x),\sigma (y)) \), \(d ( \sigma (x), \sigma (y)) \le \gamma ^{-1} d ( x,y)\).
-
2.2.2 Proof of Theorem 2.4 for \(T_\alpha \)
We prove Theorem 2.4 for \(T_\alpha \) instead of T, and Remark 2.6 guarantees that the result carries over to T. We construct a number of intermediate spaces.
Let \({\mathcal {A}}\) denote the set of all finite words in the alphabet \(\alpha \), not including the empty word. For \(w = a_0 \cdots a_{n-1} \in {\mathcal {A}}\) let \(h_{\mathcal {A}}(w) = \tau (a_0) + \cdots + \tau (a_{n-1})\).
Define \(\pi _{\mathcal {A}}:{\mathcal {A}}^{\mathbb {Z}}\rightarrow \alpha ^{\mathbb {Z}}\) by \(\pi _{\mathcal {A}}(\ldots , w_{-1}, w_0, w_1, \ldots ) = \cdots w_{-1} w_0 w_1 \cdots \), which is a concatenation with the first letter of \(w_0\) at index 0. Similarly define \({\bar{\pi }}_{\mathcal {A}}:{\mathcal {A}}^{\mathbb {N}}\rightarrow \alpha ^{{\mathbb {N}}}\) by \({\bar{\pi }}_{\mathcal {A}}(w_0, w_1, \ldots ) = w_0 w_1 \cdots \).
Let \(\pi _\alpha ^+ :\alpha ^{\mathbb {Z}}\rightarrow \alpha ^{\mathbb {N}}\) be the projection on nonnegative coordinates. Using that \(\tau _\alpha \) depends only on the “future” coordinates, define \({\bar{\tau }}_\alpha :\alpha ^{{\mathbb {N}}} \rightarrow {\mathbb {N}}_0\) so that \(\tau (x) = {\bar{\tau }}_\alpha (\pi _\alpha ^+(x))\).
Let \({\bar{M}}_\alpha = \{ (x, \ell ) \in \alpha ^{{\mathbb {N}}} \times {\mathbb {Z}}: 0 \le \ell < {\bar{\tau }}_\alpha (x) \}\), and let
Then \({\bar{T}}_\alpha :{\bar{M}}_\alpha \rightarrow {\bar{M}}_\alpha \) is a nonuniformly expanding system. It preserves the probability measure \({\bar{m}}_\alpha = {\bar{\mu }}_\alpha \times \text {counting} / \text {normalization}\), where \({\bar{\mu }}_\alpha = (\pi _\alpha ^+)_* \mu _\alpha \) is the projection of \(\mu _\alpha \) on nonnegative coordinates.
Let \(\psi :M_\alpha \rightarrow {\bar{M}}_\alpha \) be the natural projection,
Remark 2.7
\(T_\alpha \) is the natural extension of \({\bar{T}}_\alpha \). In particular, \(\psi \) is a semiconjugacy, and \({\bar{m}}_\alpha \) is the unique \({\bar{T}}_\alpha \)-invariant probability measure so that \(\psi _* m_\alpha = {\bar{m}}_\alpha \).
To use the same notations as in [19], let \(\xi = \gamma \). By [19, Theorem 3.4, Section 4.1], there exists a probability measure \({{\,\mathrm{\mathbb {P}}\,}}_{\mathcal {A}}\) on \({\mathcal {A}}\), such that the one-sided Markov shift tower (see Remark 2.2) \({\bar{f}}:{\bar{\Delta }}\rightarrow {\bar{\Delta }}\), defined by \(\bigl \{ ({\mathcal {A}}, {{\,\mathrm{\mathbb {P}}\,}}_{\mathcal {A}})\), \(h_{\mathcal {A}}\), \(\xi \bigr \}\), is an extension of \({\bar{T}}_\alpha \) with the measure preserving and Lipschitz semiconjugacy \({\bar{\pi }}_\alpha :{\bar{\Delta }}\rightarrow {\bar{M}}_\alpha \),
Next let \({\bar{{{\,\mathrm{\mathbb {P}}\,}}}}\) denote be the \({\bar{f}}\)-invariant probability measure on \({\bar{\Delta }}\).
Now let \(f :\Delta \rightarrow \Delta \) be the two-sided (i.e. as in Sect. 2.1) Markov shift tower defined by the same objects \(\bigl \{ ({\mathcal {A}}, {{\,\mathrm{\mathbb {P}}\,}}_{\mathcal {A}})\), \(h_{\mathcal {A}}\), \(\xi \bigr \}\), with the invariant probability measure \({{\,\mathrm{\mathbb {P}}\,}}\).
Let \({\bar{\pi }}_\Delta :\Delta \rightarrow {\bar{\Delta }}\) be the natural projection,
where \({\bar{\pi }}^+_{\mathcal {A}}: {\mathcal {A}}^{\mathbb {Z}}\rightarrow \alpha ^{\mathbb {N}}\) is defined by \({\bar{\pi }}^+_{\mathcal {A}}(\ldots , w_{-1}, w_0, w_1, \ldots ) =w_0 w_1 w_2 \cdots \). Observe that \({\bar{\pi }}_\Delta \) is a measure preserving and Lipschitz semiconjugacy between f and \({\bar{f}}\).
Remark 2.8
\({\bar{\pi }}_\alpha \circ {\bar{\pi }}_\Delta :\Delta \rightarrow {\bar{M}}_\alpha \) is a measure preserving semiconjugacy, in particular
Define \(\pi _\alpha :\Delta \rightarrow M_\alpha \),
Observe that \(\pi _\alpha \) is a measure preserving semiconjugacy too.
Now, f (\( \Delta \rightarrow \Delta \)) preserving the probability measure \({{\,\mathrm{\mathbb {P}}\,}}\), with the semiconjacy \(\pi _\alpha \) to \(T_\alpha \) (\( M_\alpha \rightarrow M_\alpha \)), is the Markov shift tower we are after. It remains to verify that it has the required properties: that \(\pi _\alpha \) is measure preserving and Lipschitz, and to bound the return time tails.
Proposition 2.9
\(\pi _\alpha \) is measure preserving: \((\pi _\alpha )_* {{\,\mathrm{\mathbb {P}}\,}}= m_\alpha \).
Proof
We have constructed four dynamical systems: \(f :\Delta \rightarrow \Delta \) with invariant measure \({{\,\mathrm{\mathbb {P}}\,}}\), \({\bar{f}}:{\bar{\Delta }}\rightarrow {\bar{\Delta }}\) with invariant measure \({\bar{{{\,\mathrm{\mathbb {P}}\,}}}}\), \(T_\alpha :M_\alpha \rightarrow M_\alpha \) with invariant measure \(m_\alpha \), and \({\bar{T}}_\alpha :{\bar{M}}_\alpha \rightarrow {\bar{M}}_\alpha \) with invariant measure \({\bar{m}}_\alpha \). They are connected by semiconjugacies \(\pi _\alpha :\Delta \rightarrow M_\alpha \), \({\bar{\pi }}_\alpha :{\bar{\Delta }}\rightarrow {\bar{M}}_\alpha \), \({\bar{\pi }}_\Delta :\Delta \rightarrow {\bar{\Delta }}\) and \(\psi :M_\alpha \rightarrow M_\alpha \), where \(\psi \) is the natural projection.
The left diagram commutes, and we have to justify the dashed arrow in the right diagram:
By Remark 2.7, \(m_\alpha \) is the unique probability measure on \(M_\alpha \) so that \(\psi _* m_\alpha = {\bar{m}}_\alpha \). On the other hand, \(\psi \circ \pi _\alpha = {\bar{\pi }}_\Delta \circ {\bar{\pi }}_\alpha \), and therefore \(\psi _* \bigl ( (\pi _\alpha )_* {{\,\mathrm{\mathbb {P}}\,}}\bigr ) = {\bar{m}}_\alpha \), so \((\pi _\alpha )_* {{\,\mathrm{\mathbb {P}}\,}}= m_\alpha \) as required. \(\square \)
By the same argument as in [19, Proposition 4.18], \(\pi _\alpha \) is Lipschitz.
The return time tails on \(\Delta \) and \({\bar{\Delta }}\) are equal by construction, and [19, Theorem 3.4] proves that on \({\bar{\Delta }}\) they are as required for Theorem 2.4 except for the more general than in [19] polynomial case \(\mu _\alpha (\tau _\alpha \ge n) = O(n^{-\beta } L(n))\) with \(\beta > 0\) and L a slowly varying function at infinity. In turn, the relation between the tails in [19] is taken from [18, Section 4]. To extend the results as we require, it is sufficient to notice that, by taking into account the properties of slowly varying functions (see for instance [5, Chapter 1]), the following version of [18, Proposition 4.4] holds:
Proposition 2.10
(in notation of [18]). Suppose that there exist \(C_\tau > 0\), \(\beta > 0\), L a slowly varying function at infinity and \(\ell _0 \ge 1\), such that \(m(\tau \ge \ell ) \le C_\tau \ell ^{-\beta } L( \ell )\) for all \(\ell \ge \ell _0\). Then \({{\,\mathrm{\mathbb {P}}\,}}(t \ge \ell ) \le C \ell ^{-\beta }L( \ell )\).
The proof of Theorem 2.4 for \(T_\alpha \) is complete, and following Remark 2.6 we recover the full result.
3 ASIP for Hölder Observables of Nonuniformly Hyperbolic Maps
Theorem 1.1 is an application of the following general result, which is the goal of this section:
Theorem 3.1
Let \(T :M \rightarrow M\) be an ergodic, measure-preserving transformation defined on a bounded metric space (M, d) with Borel probability measure m. Let \(\varphi :M \rightarrow {\mathbb {R}}\) be Hölder continuous and centered, i.e. \(\int \varphi \, dm = 0\). Let \(S_n( \varphi ) = \sum _{k =0}^{ n-1} \varphi \circ T^k\). Suppose that T is nonuniformly hyperbolic (in the sense of Sect. 2.1.1) with uniformly hyperbolic induced map F, return time \(\tau \) and F-invariant measure \(\mu \) associated with a subset Y of M. Assume that \(\int \tau ^2 \, d \mu < \infty \). Then the limit \(c^2= \lim _{n \rightarrow \infty } n^{-1}\int |S_n( \varphi )|^2 d \mu \) exists. In addition,
-
(a)
If \( \mu ( \tau \ge n) = O(n^{-\beta } (\log n)^\gamma )\), with \(\beta > 2\) and \(\gamma \in {\mathbb R}\), then for each \({\varepsilon }> 0\) the process \(S_n\) satisfies the ASIP with variance \(c^2\) and rate \(o(n^{1/\beta } (\log n)^{( \gamma + 1)/\beta + {\varepsilon }})\).
-
(b)
If \(\int \tau ^\beta \, d \mu < \infty \) with \(\beta > 2\), then \(S_n\) satisfies the ASIP with variance \(c^2\) and rate \(o(n^{1/\beta })\).
-
(c)
If \(\int e^{\beta \tau ^\delta } \, d \mu < \infty \) with \(\beta > 0\) and \(\delta \in (0,1]\), then \(S_n\) satisfies the ASIP with variance \(c^2\) and rate \(O(\log n)^{1 + 1/\delta }\).
Remark 3.2
The dependence of the ASIP on \({\varepsilon }\) in item (a) should be understood as follows: for every \({\varepsilon }> 0\), there exists a Brownian motion \((W_t)_{t >0}\) with variance \(c^2\) such that \(S_n = W_n + o(n^{1/\beta } (\log n)^{( \gamma + 1)/\beta + {\varepsilon }})\) almost surely.
For the flower billiard map and the two falling balls map defined in (1.2), it has been proved in [9] that they are examples of nonuniformly hyperbolic maps with tails of the return times of the form \(O ( n^{-3} ( \log n)^3 )\). On another hand, for the intermittent Baker’s map as described in (1.3), it has been proved in [24] that the associated return times have tails \(\sim n^{-\beta }\) with \(\beta =1/\alpha \). These considerations together with Theorem 3.1 prove Theorem 1.1 of the introduction.
Remark 3.3
It is expected [2] that for the maps (1.2) the tails of the return times can be improved to \(O(n^{-3})\), dropping the logarithmic factor, similarly to what was done for Buminovich stadia and related systems [10]. This would improve our result.
Let \(f :\Delta \rightarrow \Delta \) be a Markov shift tower constructed for T as in Theorem 2.4, built with a probability space \(({\mathcal {A}}, {{\,\mathrm{\mathbb {P}}\,}}_{\mathcal {A}})\), roof function \(h_{\mathcal {A}}\) and the metric constant \(\xi \in (0,1)\). Recall the associated notations, in particular that \((\Delta , d)\) is a metric space and that f preserves the probability measure \({{\,\mathrm{\mathbb {P}}\,}}\).
By Theorem 2.4, the process \(( \varphi \circ T^k)_k\) defined on (M, m) and the process \(( v \circ f^k)_k\), where \(v = \varphi \circ \pi \), defined on \((\Delta ,{{\,\mathrm{\mathbb {P}}\,}})\) have the same law. Moreover, if \( \varphi \) is Hölder continuous then so is v since \(\pi \) is Lipschitz. Hence Theorem 3.1 will follow from its equivalent result stated for Markov shift towers as given in Proposition 3.4 below.
3.1 ASIP for Markov shift towers
Proposition 3.4
Suppose that \(v :\Delta \rightarrow {\mathbb {R}}\) is Hölder continuous and centered, i.e. \(\int v \, d{{\,\mathrm{\mathbb {P}}\,}}= 0\). Let \(S_n = \sum _{k < n} v \circ f^k\). Assume that \(\int h_{\mathcal {A}}^2 \, d{{\,\mathrm{\mathbb {P}}\,}}_{\mathcal {A}}< \infty \). Then \(c^2= \lim _{n \rightarrow \infty } n^{-1}\int |S_n( \varphi )|^2 d \mu \) exists. In addition,
-
(a)
If \({{\,\mathrm{\mathbb {P}}\,}}_{\mathcal {A}}(h_{\mathcal {A}}\ge n) = O(n^{-\beta } (\log n)^\gamma )\), with \(\beta > 2\) and \(\gamma \in {\mathbb R}\), then \(S_n\) satisfies the ASIP with variance \(c^2\) and rate \(o(n^{1/\beta } (\log n)^{ (\gamma + 1)/\beta + {\varepsilon }})\) for each \({\varepsilon }> 0\).
-
(b)
If \(\int h_{\mathcal {A}}^\beta \, d{{\,\mathrm{\mathbb {P}}\,}}_{\mathcal {A}}< \infty \) with \(\beta > 2\), then \(S_n\) satisfies the ASIP with variance \(c^2\) and rate \(o(n^{1/\beta })\).
-
(c)
If \(\int e^{\beta h_{\mathcal {A}}^\delta } \, d{{\,\mathrm{\mathbb {P}}\,}}_{\mathcal {A}}< \infty \) with \(\beta > 0\) and \(\delta \in (0,1]\), then \(S_n\) satisfies the ASIP with variance \(c^2\) and rate \(O(\log n)^{1 + 1/\delta }\).
The rest of this section is devoted to the proof of Proposition 3.4.
3.2 The associated Markov chain
It is convenient to represent the dynamics on a Markov shift tower as a Markov chain in the conventional sense. For this, let \(G = \{(a, \ell ) \in {\mathcal {A}}\times {\mathbb {Z}}: 0 \le \ell < h_{\mathcal {A}}(a)\}\) and let \({\mathcal {G}}\subset G^{\mathbb {Z}}\) be the set of admissible symbolic trajectories:
Let now \((g_n)_{n\ge 0}\) be a Markov chain with state space G and transition probabilities
The Markov chain has a unique invariant probability measure \(\nu \) with respect to which the Markov chain \((g_n)_{n \ge 0}\) is stationary. By the Kolmogorov existence theorem, there exists a stationary Markov chain indexed by \({\mathbb {Z}}\) that we still denote by \((g_n)_{n \in {\mathbb {Z}}}\) with transition probabilities given above and which defines a probability measure \({{\,\mathrm{\mathbb {P}}\,}}_{{\mathcal {G}}}\) on the space \({\mathcal {G}}\).
Define \(\pi _G :\Delta \rightarrow {\mathcal {G}}\) by \(\pi _G(x) = (a_n, \ell _n)_n\) such that \(f^n(x) \in ([a_n], \ell _n)\) for all n, where \([a_n] \subset {\mathcal {A}}^{\mathbb {Z}}\) is the cylinder \(\{ x \in {\mathcal {A}}^{\mathbb {Z}}: x_0 = a_n \}\). Note that \(\pi _G\) is a bijection; moreover, it is a measure preserving conjugacy. It follows that if we define
then setting for all \(k \in {\mathbb {Z}}\),
the stationary process \((X_k)_{k \in {\mathbb {Z}}} \) has the same law as \(( v \circ f^k)_{k \in {\mathbb {Z}}} \). To give the regularity properties of \(\psi \), we need to equip \({\mathcal {G}}\) with a suitable metric.
3.3 Metric on \({\mathcal {G}}\)
Since \(\pi _G :\Delta \rightarrow {\mathcal {G}}\) is a bijection and \(\Delta \) is a metric space, it is natural to define a metric on \({\mathcal {G}}\) so that \(\pi _G\) is an isometry. For \(g, g' \in {\mathcal {G}}\), let
Let \(G_0 =\{(a, 0) : a \in {\mathcal {A}}\} \subset G\),
Let
Then s is a kind of separation time (cf. the separation time on X defined in Sect. 2.1). Recall that \(\xi \in (0,1)\) and define
This is the metric on \({\mathcal {G}}\) which agrees with that on \(\Delta \) in the sense that \(d(x,y) = d(\pi _G(x), \pi _G(y))\) for all \(x,y \in \Delta \).
This allows to infer that if v is Hölder with index \(\eta \in ]0,1]\) then so is \(\psi \) defined by (3.1).
Next, to prove the ASIP for the partial sums associated with \((X_k)_{k \in {\mathbb {Z}}}\) defined by (3.2) with \(\psi \) bounded and Hölder continuous, it is convenient, as in [11, 12], to represent the Markov chain with the help of its innovations and to introduce a particular meeting time.
3.4 Innovations and meeting time of the underlying Markov chain
Extending the approach in [11, Section 3], without changing the distribution of \((g_n)_{n \in {\mathbb {Z}}}\), we assume that there is a sequence of iid random elements \(({\varepsilon }_n)_{n \in {\mathbb {Z}}}\) sampled from \(({\mathcal {A}}, {{\,\mathrm{\mathbb {P}}\,}}_{\mathcal {A}})\) such that each \({\varepsilon }_{n+1}\) is independent from \(({\varepsilon }_\ell , g_\ell )_{\ell \le n}\), and \(g_{n+1}=U(g_n,{\varepsilon }_{n+1})\), where
We refer to \(({\varepsilon }_n)_{n \in {\mathbb {Z}}}\) as innovations.
For \(g_0, g'_0 \in G\) and a sequence \(({\varepsilon }_n)_{n \in {\mathbb {Z}}}\) of random elements in \({\mathcal {A}}\), let \(T_{g_0, g'_0} (({\varepsilon }_n))\) be the meeting time of two Markov chains with respective initial states \(g_0, g'_0\) and common innovations \(({\varepsilon }_n)_{n \in {\mathbb {Z}}}\):
where \(g_{n+1} = U(g_n, {\varepsilon }_{n+1})\) and \(g'_{n+1} = U(g'_n, {\varepsilon }_{n+1})\).
Proposition 3.5
Suppose that \(g'_0\) is an independent copy of \(g_0\), also independent from \((g_n, {\varepsilon }_n)_{n \ge 0}\). Set \(T = T_{g_0, g'_0} (({\varepsilon }_n))\).
-
1.
Assume that there exists a sequence \({\varepsilon }(n)\) of positive reals tending to 0 as \(n \rightarrow \infty \), such that \({{\,\mathrm{\mathbb {P}}\,}}_{\mathcal {A}}( h_{\mathcal {A}}\ge n) \ge \exp (-n {\varepsilon }(n))\). If in addition \({{\,\mathrm{\mathbb {P}}\,}}_{\mathcal {A}}( h_{\mathcal {A}}\ge n) = O ( L(n) n^{-\beta }) \) with \(\beta > 1\) and L(n) a slowly varying function at infinity, then \({{\,\mathrm{\mathbb {P}}\,}}(T \ge n ) = O ( L(n) n^{-(\beta -1)})\).
-
2.
If \({{\,\mathrm{\mathbb {P}}\,}}_{\mathcal {A}}( h_{\mathcal {A}}\ge n) = O ( L(n) n^{-\beta }) \) with \(\beta > 1\) and L(n) a slowly varying function at infinity, then \( {{\,\mathrm{\mathbb {E}}\,}}( g_\beta (T) ) <\infty \) for any \(\eta >1\) where \(g_\beta (x) = x^{\beta -1} (\log (1+x))^{- \eta }/ L(x)\).
-
3.
If \({{\,\mathrm{\mathbb {E}}\,}}_{\mathcal {A}}( h^\beta _{\mathcal {A}}) < \infty \) for some \(\beta >1\), then \({{\,\mathrm{\mathbb {E}}\,}}( T^{\beta -1} ) < \infty \).
-
4.
If \({{\,\mathrm{\mathbb {P}}\,}}_{\mathcal {A}}( h_{\mathcal {A}}\ge n) = O ( \textrm{e}^{ - c n^\delta } )< \infty \) for some \(c >0\) and \(\delta \in (0,1]\) , then there exists \(\kappa >0\) such that \( {{\,\mathrm{\mathbb {P}}\,}}(T \ge n ) = O ( \textrm{e}^{ - \kappa n^\delta } )\).
Items 2 and 3 are proved in [11] whereas Item 4 is proved in [12]. Item 1 of Proposition 3.5 follows straightforwardly from Lemma 3.6 below.
Lemma 3.6
Suppose that \(g'_0\) is an independent copy of \(g_0\), also independent from \((g_n, {\varepsilon }_n)_{n \ge 0}\). Set \(T = T_{g_0, g'_0} (({\varepsilon }_n))\). Suppose that there exists a sequence \({\varepsilon }(n)\) of positive reals tending to 0 as \(n \rightarrow \infty \), such that \({{\,\mathrm{\mathbb {P}}\,}}_{\mathcal {A}}( h_{\mathcal {A}}\ge n) \ge \exp (-n {\varepsilon }(n))\). Then
Proof
As in Lindvall [21] (see also Rio [27, Proposition 9.6]), let \(\Lambda _0\) be the class of nondecreasing functions \(\psi \) from \({\mathbb N}\) to \([1, \infty [\) such that \(\log (\psi (n))/n\) decreases to 0 as \(n \rightarrow \infty \). Let
Proceeding exactly as in the proof of [11, Lemma 3.1] and applying the result of [21], we infer that
This last assertion is in fact equivalent to
Now, assume that the conclusion of Lemma 3.6 is not true. Then there exists an increasing sequence \((n_k)_{k \ge 1}\) such that \({{\,\mathrm{\mathbb {P}}\,}}( T \ge n_k) \ge k {{\,\mathrm{\mathbb {E}}\,}}_{\mathcal {A}}( (h_{\mathcal {A}}-n_k)_+)\). Define then the function \(\varphi \) as follows: \(\varphi (i)=0\) if \(i\notin \{n_k, k \ge 1\}\) and \(\varphi (n_k)= k^{-3/2}/{{\,\mathrm{\mathbb {E}}\,}}_{\mathcal {A}}( (h_{\mathcal {A}}-n_k)_+) \). For such a \(\varphi \) it is clear that the sum on left hand in (3.3) is finite, while the selection of \(n_k\) implies that the sum on right hand is \(+\infty \). This leads to a contradiction and proves the Lemma, provided we check that the function \(\psi \) defined by \(\psi (k) =\psi (0) + \varphi (1)+ \cdots + \varphi (k)\) belongs to \(\Lambda _0\). Hence, it remains to check that \(\log (\psi (n_k))/n_k\) tends to 0 as \(k\rightarrow \infty \). Now, by definition of \(\varphi \), \(\psi (n_k)-\psi (0) \le C\sqrt{k}/{{\,\mathrm{\mathbb {E}}\,}}_{\mathcal {A}}( (h_{\mathcal {A}}-n_k)_+)\). On another hand, the assumption on \(h_{\mathcal {A}}\) implies that \({\mathbb E}_{\mathcal {A}}( (h_{\mathcal {A}}-n_k)_+) \ge \exp (-n_k {\varepsilon }'(n_k))\) for some sequence \({\varepsilon }'(n)\) of positive reals tending to 0 as \(n \rightarrow \infty \). This implies that \(\log (\psi (n_k))/n_k\) tends to 0 as \(k\rightarrow \infty \), and the proof is complete. \(\square \)
3.5 Proof of Proposition 3.4 when \(c^2 > 0\)
By the previous considerations, it suffices to prove the ASIP for the partial sums associated with \((X_k)_{k \in {\mathbb {Z}}}\) defined by (3.2) with \(\psi \) bounded and Hölder continuous, by taking into account Proposition 3.5.
Without loss of generality, one can assume that \(v :\Delta \rightarrow {\mathbb {R}}\) is Lipschitz and that \(\Vert v\Vert _{{{\,\textrm{Lip}\,}}} \le 1\), where \(\Vert v\Vert _{{{\,\textrm{Lip}\,}}} = \sup _x |v(x)| + \sup _{x \ne y} |v(x) - v(y)| / d(x,y)\). Recall that \(\psi :{\mathcal {G}}\rightarrow {\mathbb {R}}\), \(\psi = v \circ \pi _G^{-1}\). Then \(\psi \) is also Lipschitz, in particular
We assume also in the rest of this section that the Markov chain \((g_n)_{n \in {\mathbb {Z}}}\) is aperiodic. If it is not the case, we modify the proof as in [11, Appendix A].
The proof follows the lines of [11, Section 4.2] or [12, Section 3] with the same notations, but with the following changes. [11, Proposition 3.2] and [12, Proposition 2.3] are replaced by Proposition 3.7 below whereas [11, Inequality 4.10] is replaced by our Proposition 3.8, whereas [11, Lemma 3.3] is replaced by our Lemma 3.9. Moreover in the case where \({{\,\mathrm{\mathbb {P}}\,}}_{\mathcal {A}}(h_{\mathcal {A}}\ge n) = O(n^{-\beta } (\log n)^\gamma )\), with \(\beta > 2\) and \(\gamma \in {\mathbb R}\), we take \(m=m_{\ell } = [3^{\ell /\beta } \ell ^{\kappa }]\) with \(\kappa > (1+\gamma )/\beta \). In case \(\int h_{\mathcal {A}}^\beta \, d{{\,\mathrm{\mathbb {P}}\,}}_{\mathcal {A}}< \infty \) with \(\beta > 2\), we select \(m=m_{\ell } = [3^{\ell /\beta } ]\), and if \(\int e^{\beta h_{\mathcal {A}}^\delta } \, d{{\,\mathrm{\mathbb {P}}\,}}_{\mathcal {A}}< \infty \) with \(\beta > 0\) and \(\delta \in (0,1]\), we take \(m=m_{\ell } = [\kappa \ell ^{1/\delta } ] \), for a suitable \(\kappa \).
To state the key propositions 3.7 and 3.8, we need to introduce some notations.
Fix \(k \in {\mathbb {Z}}\) and \(m > 0\), and let \(({\varepsilon }'_n)_n\) be a copy of \(({\varepsilon }_n)_n\) independent of \((g_n, {\varepsilon }_n)_n\). Define
where \({{\,\mathrm{\mathbb {E}}\,}}_g\) is the conditional expectation given \((g_n, {\varepsilon }_n)\).
Let
be the meeting time between \((g_n)_{n \ge 0}\) and \(({\tilde{g}}_{n})_{n \ge 0}\) with the same innovation \(({\varepsilon }_n)_{n \ge 1}\) but starting from two independent starting points \(g_0\) and \(g_0'\).
Proposition 3.7
-
1.
Assume that \({{\,\mathrm{\mathbb {E}}\,}}(T) < \infty \). Then, for every \(r \ge 1\),
$$\begin{aligned} {{\,\mathrm{\mathbb {E}}\,}}| X_{m,k} - X_k | \ll m^{-r/2} + {{\,\mathrm{\mathbb {P}}\,}}( T \ge \lfloor m / r \rfloor ) . \end{aligned}$$ -
2.
Assume that there exist \(\delta >0\) and \(\gamma \in ]0,1]\) such that \({{\,\mathrm{\mathbb {P}}\,}}( T \ge n) = O ( \textrm{e}^{- \delta n^{\gamma }} )\). Then there exists \(\delta ' >0\) such that
$$\begin{aligned} {{\,\mathrm{\mathbb {E}}\,}}| X_{m,k} - X_k | = O ( ( \textrm{e}^{- \delta ' m^{\gamma }} ) . \end{aligned}$$
Proof
See [11, Proposition 3.2] and [12, Proposition 2.3]. \(\square \)
Now define
Proposition 3.8
Let \(\theta ^{-}_u = \sum _{k=0}^{u-1} \textbf{1}_{\{ g_{-k} \in G_0 \}}\) and \(\theta ^{+}_u = \sum _{k=0}^{u-1} \textbf{1}_{\{ g_{k} \in G_0 \}}\). Then
where T is defined by (3.4). Consequently:
-
1.
If \({{\,\mathrm{\mathbb {E}}\,}}(T) < \infty \), then, for every \(r \ge 1\),
$$\begin{aligned} {{\,\mathrm{\mathbb {E}}\,}}| {\widetilde{X}}_{m,k} - X_k | \ll m^{-r/2} + {{\,\mathrm{\mathbb {P}}\,}}( T \ge \lfloor m / r \rfloor ) . \end{aligned}$$ -
2.
If there exist \(\delta >0\) and \(\gamma \in ]0,1]\) such that \({{\,\mathrm{\mathbb {P}}\,}}( T \ge n) = O ( \textrm{e}^{- \delta n^{\gamma }} )\), then there exists \(\delta ' >0\) such that
$$\begin{aligned} {{\,\mathrm{\mathbb {E}}\,}}| {\widetilde{X}}_{m,k} - X_k | = O ( \textrm{e}^{- \delta ' m^{\gamma }} ) . \end{aligned}$$
Proof
Since \(X_{m,k}\) is determined by \((g_n)_{n \le k + m}\), we write \(X_{m,k} = h_m((g_n)_{n \le k + m})\) with some function \(h_m\).
Let \((g'_n, {\varepsilon }'_n)_{n \in {\mathbb {Z}}}\) be an independent copy of \((g_n, {\varepsilon }_n)_{n \in {\mathbb {Z}}}\). Note that \(g_n' = U (g'_{n-1}, {\varepsilon }'_n)\). Define \(( {\hat{g}}_{\ell }) \) by
Let \({\mathcal {B}}_{k,m}\) be the sigma-algebra generated by \(\bigl ( (g_{\ell })_{\ell < k-m} , {\varepsilon }_{k-m}, \ldots , {\varepsilon }_{k+m} \bigr )\). By the properties of conditional expectation, we have
and
Set
This is the meeting time of the chains \((g_n)_{n \ge k - m - 1}\) and \(({\hat{g}}'_n)_{n \ge k - m-1}\) with innovations \(( {\varepsilon }_{\ell } )_{\ell \le k - m}\) and independent starting points \(g_{k-m-1}\) and \({\hat{g}}'_{k-m-1}\). It has the same law as T defined in (3.4). Note that
But, from the definition of the separation distance and using \(\Vert v\Vert _{{{\,\textrm{Lip}\,}}} \le 1\), for any \(\ell \) in \(\{0, \dots , [m/2] \}\),
Taking the expectation, we get
The first part of the proposition follows by stationarity and the fact that \( \Vert h_m \Vert _{\infty } \le \Vert \psi \Vert _\infty \). To end the proof, we used the same arguments as in the proofs of [11, Proposition 3.2] and [12, Proposition 2.3]. \(\square \)
Lemma 3.9
-
1.
If \({{\,\mathrm{\mathbb {E}}\,}}(T) < \infty \), for every \(\alpha \ge 1\) and every \(k \ge 1\),
$$\begin{aligned} | \textrm{Cov} (X_0,X_k) | \ll k^{-\alpha /2} + {{\,\mathrm{\mathbb {P}}\,}}( T \ge \lfloor k / (4 \alpha ) \rfloor ) . \end{aligned}$$ -
2.
If there exist \(\delta >0\) and \(\gamma \in ]0,1]\) such that \({{\,\mathrm{\mathbb {P}}\,}}( T \ge n) = O ( \textrm{e}^{- \delta n^{\gamma }} )\), then for any \(k \ge 0\), there exists \(\delta ' >0\) such that
$$\begin{aligned} | \textrm{Cov} (X_0,X_k) | = O ( ( \textrm{e}^{- \delta ' k^{\gamma }} ) . \end{aligned}$$
Proof
For every positive integers m and n, we have
For every \(k \ge 2\), we select \(m= [k/2]\) and \(n = [k/2] -1\). In this case \({\widetilde{X}}_{m,0}\) and \({\widetilde{X}}_{n,k}\) are independent, implying that for any \(k \ge 2\),
This upper bound combined with Proposition 3.8 proves the lemma. \(\square \)
3.6 Proof of Proposition 3.4 when \(c^2=0\)
Like in the nonuniformly expanding case [11], our strategy is to represent \(X_k\) as a coboundary: \(X_k = z_{k-1} - z_k\), where \(z_i\) is a sufficiently nice stationary sequence, and to bound the growth of \(z_k\) almost surely. Our proof follows well established ideas [6, 13, 24, 30] but unlike in the nonuniformly expanding case, we have not found a result that we can conveniently cite.
The representation \(X_k = z_{k-1} - z_k\) is provided by the following general result which is essentially due to Gordin [16] (see also [13, Lemma 7.1] in case of adapted r.v.’s but there the condition c) is actually unnecessary).
Lemma 3.10
Suppose that \((X_k)_{k \in {\mathbb {Z}}}\) is a strictly stationary sequence of real-valued random variables that are in \(L^2\), and let \(({{\mathcal {F}}}_k)_{k \in {\mathbb {Z}}}\) be a stationary filtration. Let \({{\,\mathrm{\mathbb {E}}\,}}_i\) denote the conditional expectation given \({\mathcal {F}}_i\) and set \(S_n = X_1+ \cdots + X_n\). Assume that
-
(a)
\({{\,\mathrm{\mathbb {E}}\,}}_0 (S_n)\) and \( \sum _{k =0}^{n} ( X_{-k} - {{\,\mathrm{\mathbb {E}}\,}}_{0} ( X_{-k}) )\) converge in \(L^1\) as \(n \rightarrow \infty \),
-
(b)
\(\displaystyle \liminf _{n \rightarrow \infty } \frac{{{\,\mathrm{\mathbb {E}}\,}}|S_n| }{\sqrt{n}}=0\).
Then, almost surely, \( X_i = z_{i-1} - z_i \), where \(z_i = g_i-h_i\) with \(g_i = \sum _{k \ge i +1} {{\,\mathrm{\mathbb {E}}\,}}_{i} ( X_k)\) and \(h_i = \sum _{k \le i } ( X_{k} - {{\,\mathrm{\mathbb {E}}\,}}_{i} ( X_{k}) )\). All \(z_i, g_i, h_i\) are in \(L^1\).
Proof
Set \(d_{i} = \sum _{k \in {\mathbb Z}} P_i (X_{k+i}) \) where \(P_i = {{\,\mathrm{\mathbb {E}}\,}}_i - {{\,\mathrm{\mathbb {E}}\,}}_{i-1} \). Then
By assumption, all random variables in the above decomposition are in \(L^1\). Moreover, \((d_{i})_{i \in {\mathbb {Z}}}\) is a stationary martingale difference sequence. Let’s prove that the \(d_i\)’s are almost surely equal to zero. With this aim, we shall proceed as in [14]. By Burkholder’s deviation inequality for the martingale square function [8] (see [14, Lemma 6] for an easy reference), there exists a positive constant C such that, for any \(\lambda >0\),
But, by (3.6), \( {{\,\mathrm{\mathbb {E}}\,}}\big | \sum _{i=1}^n d_i \big | \le {{\,\mathrm{\mathbb {E}}\,}}|S_n| + 2 {{\,\mathrm{\mathbb {E}}\,}}|z_0| \). Using part (b), there exists a subsequence \((n_k)_{k>0}\) tending to infinity such that \(n_k^{-1} \sum _{i=1}^{n_k} d_i^2\) converges to zero in probability as \(k \rightarrow \infty \). On the other hand, by the ergodic theorem, since \((d_i^2)_{i \in {\mathbb Z}}\) is a stationary sequence of nonnegative r.v.’s, \(n^{-1} \sum _{i=1}^n d_i^2\) converges almost surely to \({{\,\mathrm{\mathbb {E}}\,}}(d_0^2 | \mathcal {I} ) \) where \(\mathcal {I}\) is the so-called invariant \(\sigma \)-field. So, overall, \( {{\,\mathrm{\mathbb {E}}\,}}(d_0^2 | \mathcal {I} ) =0\) almost surely and \({{\,\mathrm{\mathbb {E}}\,}}(d_0^2) = {{\,\mathrm{\mathbb {E}}\,}}( {{\,\mathrm{\mathbb {E}}\,}}(d_0^2 | \mathcal {I} ))=0\). This proves that \(d_0=0\) almost surely. \(\square \)
Now we return to the specific process we are interested in: \(X_k = \psi ( (g_{k+\ell })_{\ell \in {\mathbb {Z}}})\) as in (3.2). Let \({\mathcal {F}}_k = \sigma ( {\varepsilon }_i, i \le k)\).
Proposition 3.11
Suppose that \({{\,\mathrm{\mathbb {E}}\,}}(T) < \infty \) and \(c^2 = 0\). Then the assumptions of Lemma 3.10 are satisfied.
Proof
Recall the construction of \(X_{m,k}\) and \({\widetilde{X}}_{m,k} = {{\,\mathrm{\mathbb {E}}\,}}( X_{m,k} | {\varepsilon }_{k-m}, \ldots , {\varepsilon }_{k+m})\). For \(k \ge 0\), let \(X_{k}^*={\widetilde{X}}_{m_{k} ,k} \) with \(m_k = \lfloor k/2 \rfloor \). Clearly \(X_{k}^*\) is independent of \({\mathcal {F}}_{0}\) and is centered. To prove the first part of assumption (a), that is the convergence of \({{\,\mathrm{\mathbb {E}}\,}}_0 (S_n)\), write
and use Proposition 3.8 to prove that the right hand side above converges.
Next we prove the second part of assumption (a). For \(k \ge 0\) let \(X_{-k}^* = {\widetilde{X}}_{\lfloor k/2 \rfloor , - k}\). Clearly \(X_{-k}^*\) is \({\mathcal {F}}_0\)-measurable. Hence
By another application of Proposition 3.8, for any \(r \ge 1\),
Choosing \(r >1\), the second part of assumption (a) follows.
We turn now to the proof of assumption (b). By Lemma 3.9, since \({{\,\mathrm{\mathbb {E}}\,}}(T) < \infty \), \(\sum _{k \ge 0} \bigl | \textrm{Cov} (X_0, X_k) \bigr | < \infty \) and then \(c^2=\lim _{n \rightarrow \infty } n^{-1} {{\,\mathrm{\mathbb {E}}\,}}(S_n^2)\) is well defined and is assumed to be zero. It follows that assumption (b) is satisfied. \(\square \)
At this point, as long as \(c^2 = 0\), we have the representation
with \(z_n\) as in Lemma 3.10. Thus the proof of Proposition 3.4 is reduced to estimating \(z_n\) almost surely, and is completed by:
Proposition 3.12
Suppose \({{\,\mathrm{\mathbb {E}}\,}}(T) < \infty \) and \(c^2 = 0\) and \(z_i\) are as in Lemma 3.10.
-
1.
If \({{\,\mathrm{\mathbb {E}}\,}}(T^{p}) < \infty \) for some \(p>1\), then \( \Vert z_0 \Vert _p < \infty \) and \(S_n = o (n^{1/(p+1)})\) a.s.
-
2.
If \({{\,\mathrm{\mathbb {E}}\,}}(\psi _p(T) ) < \infty \) with \(\psi _p(x) = x^{p} ( \log x )^{ - ( \gamma + 1 + {\varepsilon })}\) for some \({\varepsilon }>0\), then \(S_n = o (n^{1/(p+1)} ( \log n)^{ ( \gamma + 1 + {\varepsilon }) /(p+1) } )\) a.s.
-
3.
If \({{\,\mathrm{\mathbb {P}}\,}}( T \ge n) = O ( \textrm{e}^{- \delta n^{\gamma }} )\) with some \(\delta >0\) and \(\gamma \in ]0,1]\), then \(S_n = O ( ( \log n)^{1/ \gamma } )\) a.s.
Proof
Denote \(M = \Vert X_0\Vert _\infty = \Vert \psi \Vert _\infty \).
We begin with part 1: let \(p > 1\) and suppose that \({{\,\mathrm{\mathbb {E}}\,}}(T^p) < \infty \). By Proposition 3.8,
First we use (3.7) to prove that \(z_0\) is in \(L^p\). Recall that \(z_0 = g_0 - h_0\) where \(g_0 = \sum _{k \ge 1} {{\,\mathrm{\mathbb {E}}\,}}_{0} ( X_k)\) and \(h_0 = \sum _{k \le 0 } ( X_{k} - {{\,\mathrm{\mathbb {E}}\,}}_{0} ( X_{k}) ) \). Observe that for \(k \ge 2\), these summands can be bounded in \(L^1\) by those in (3.7): \(\Vert {{\,\mathrm{\mathbb {E}}\,}}_0 (X_k) \Vert _1 \le \Vert X_k - {\widetilde{X}}_{ [k/2] ,k} \Vert _1\) and \(\Vert X_{-k} - {{\,\mathrm{\mathbb {E}}\,}}_{0} ( X_{-k}) \Vert _1 \le 2 \Vert X_{-k} - {\widetilde{X}}_{ [k/2] ,-k} \Vert _1\). Now, \(g_0\), \(h_0\) and consequently \(z_0\) are in \(L^p\) by an application of Lemma 3.13 with \(V=g_0\) and \(V=h_0\) respectively. (To prove that \(g_0\) is in \(L^p\) we can also use the arguments given in the proof of [13, Proposition 2.1].)
To prove that \(S_n = o (n^{1/(p+1)})\) a.s., it suffices to show that, for every \({\varepsilon }>0\), \( \sum _{\ell>0} {{\,\mathrm{\mathbb {P}}\,}}\bigl ( \max _{k \le 2^{\ell }} |S_k| > {\varepsilon }2^{\ell /(p+1)} \bigr ) < \infty \), which is equivalent to proving that
Let \(k \le n\) and write
Let \(q \ge 1\). Using \({{\,\mathrm{\mathbb {E}}\,}}_k( X_i - {{\,\mathrm{\mathbb {E}}\,}}_{i-q} (X_i) ) =0 \) if \(i -q \ge k\), we have
Then, using \({{\,\mathrm{\mathbb {E}}\,}}_k( {{\,\mathrm{\mathbb {E}}\,}}_{\ell + q } (X_{\ell } ) ) = {{\,\mathrm{\mathbb {E}}\,}}_{\ell + q } (X_{\ell } ) \) for \(\ell \le k-q\),
Overall,
Accordingly, if \(x > 0\) and \(qM \le x\), then
Let
Starting from (3.9) and using Doob’s maximal inequality, as done in the proof of inequality (A.42) in [25], we infer that for any nondecreasing, non negative and convex function \(\varphi \) and any \(x > 0\),
Then, taking \(\varphi (x) =x^p\),
The expression above is finite: \(\Vert S_n\Vert _p \le 2 \Vert z_0\Vert _p < \infty \) and \(\sum _{n >1} n^{p-1} \theta (n) < \infty \) by Proposition 3.8. This verifies (3.8) and completes the proof of part 1.
We now turn to part 2. By part 1, \(\Vert S_n \Vert _r < \infty \) for all \(r \in [1, p[\). Hence, with \(b_n = n^{1/(p+1)} (\log n)^{ ( {\varepsilon }+ \gamma +1 )/(p+1) }\), applying (3.10) with \(\varphi (x) = x^r\), it follows that for any \(\eta >0\),
where c is a positive constant which does not depend on n. Part 2 is proved if we show that \(\sum _{n \ge 1} b_n^{-1} {{\,\mathrm{\mathbb {P}}\,}}(T > c \eta b_n) < \infty \). By a change of variable this is equivalent to \(\sum _{n \ge 2} n^{p-1} (\log n )^{ {\varepsilon }+ \gamma +1} {{\,\mathrm{\mathbb {P}}\,}}(T > n) < \infty \) which holds by Proposition 3.8 because \( {{\,\mathrm{\mathbb {E}}\,}}(\psi _p (T)) < \infty \).
We turn now to the proof of part 3. Let \(K>0\), to be chosen later. By (3.10) applied with with \(x= K(\log n)^{1/\gamma }\) and \(\varphi (u)=u^{ 2}\), using Lemma (3.13), our assumption and part 2 of Proposition 3.8, we see that
We conclude by taking \(K =2M\delta '^{-1/\gamma }\). \(\square \)
Lemma 3.13
Let \((V_k)_{k\in {\mathbb {N}}}\) be a sequence of non negative random variables uniformly bounded by M. Set \(V = \sum _{k\in {\mathbb {N}}}V_k\). Then, for every \(p>1\), we have
Proof
Recall that \(\Vert V\Vert _p=\sup _{\Vert Z\Vert _q=1}{{\,\mathrm{\mathbb {E}}\,}}(ZV)\), where \(q=p/(p-1)\). Let Z be a random variable with \(\Vert Z\Vert _q=1\). Then
and the lemma is proved.\(\square \)
Data availability
There is no associated data.
References
Bálint, P., Borbély, G., Varga, A.N.: Statistical properties of the system of two falling balls. Chaos 22, 026104 (2012)
Bálint, P., Zhang, H.K.: private communication
Berkes, I., Liu, W., Wu, W.: Komlós–Major–Tusnády approximation under dependence. Ann. Probab. 42, 794–817 (2014)
Berkes, I., Philipp, W.: Approximation theorems for independent and weakly dependent random vectors. Ann. Probab. 7, 29–54 (1979)
Bingham, N.H., Goldie, C.M., Teugels, J.L.: Regular Variation. Cambridge University Press, Cambridge (1987)
Bowen, R.: Equilibrium States and the Ergodic Theory of Anosov Diffeomorphisms. Lecture Notes in Math, vol. 470. Springer, Berlin (1975)
Bunimovich, L.A.: The ergodic properties of billiards that are nearly scattering. Sov. Math. Dokl. 14, 1136–1139 (1973)
Burkholder, D.L.: A sharp inequality for martingale transforms. Ann. Probab. 7(5), 858–863 (1979)
Chernov, N., Zhang, H.-K.: Billiards with polynomial mixing rates. Nonlinearity 18, 1527–1553 (2005)
Chernov, N., Zhang, H.-K.: Improved estimates for correlations in billiards. Commun. Math. Phys. 277, 305–321 (2008)
Cuny, C., Dedecker, J., Korepanov, A., Merlevède, F.: Rates in almost sure invariance principle for slowly mixing dynamical systems. Ergod. Theory Dyn. Syst. 40, 2317–2348 (2020)
Cuny, C., Dedecker, J., Korepanov, A., Merlevède, F.: Rates in almost sure invariance principle for quickly mixing dynamical systems. Stoch. Dyn. 20(1), 2050002 (2020)
Dedecker, J., Merlevède, F., Rio, E.: Deviation inequalities for dependent sequences with applications to strong approximations. Stoch. Process. Appl. 174 (2024), Paper No. 104377
Esseen, C.G., Janson, S.: On moment conditions for normed sums of independent variables and martingale differences. Stoch. Process. Appl. 19(1), 173–182 (1985)
Fleming-Vázquez, N.: Functional correlation bounds and optimal iterated moment bounds for slowly-mixing nonuniformly hyperbolic maps. Commun. Math. Phys. 391, 173–198 (2022)
Gordin, M.I.: Central limit theorems for stationary processes without the assumption of finite variance. In: Abstracts of Communications, T.I:A-K. International Conference on Probability Theory and Mathematical Statistics, June 25–30, Vilnius, pp. 173–174 (1973)
Gouëzel, S.: Almost sure invariance principle for dynamical systems by spectral methods. Ann. Probab. 38, 1639–1671 (2010)
Korepanov, A.: Equidistribution for nonuniformly expanding dynamical systems, and application to the almost sure invariance principle. Commun. Math. Phys. 359, 1123–1138 (2018)
Korepanov, A.: Rates in almost sure invariance principle for dynamical systems with some hyperbolicity. Commun. Math. Phys. 363, 173–190 (2018)
Kuelbs, J., Philipp, W.: Almost sure invariance principles for partial sums of mixing B-valued random variables. Ann. Probab. 8, 1003–1036 (1980)
Lindvall, T.: On coupling of discrete renewal processes. Z. Wahrscheinlichkeitstheor. verw. Geb. 48, 57–70 (1979)
Melbourne, I., Nicol, M.: Almost sure invariance principle for nonuniformly hyperbolic systems. Commun. Math. Phys. 260, 131–146 (2005)
Melbourne, I., Nicol, M.: A vector-valued almost sure invariance principle for hyperbolic dynamical systems. Ann. Probab. 37, 478–505 (2009)
Melbourne, I., Varandas, P.: A note on statistical properties for nonuniformly hyperbolic systems with slow contraction and expansion, Stoch. Dyn. 16 (2016)
Merlevède, F., Rio, E.: Strong approximation of partial sums under dependence conditions with application to dynamical systems. Stoch. Process. Appl. 122(1), 386–417 (2012)
Philipp, W., Stout, W.: Almost sure invariance principles for partial sums of weakly dependent random variables. Am. Math. Soc. Mem. 161 (1975)
Rio, A.: Asymptotic theory of weakly dependent random processes. Translated from the 2000 French edition. Probability Theory and Stochastic Modelling, 80. Springer, Berlin, xviii+204 pp (2017)
Rokhlin, V.A.: Exact endomorphisms of a Lebesgue space. Am. Math. Soc. 39, 1–36 (1964)
Sarig, O.: Lecture notes on ergodic theory. https://www.weizmann.ac.il/math/sarigo/lecture-notes
Sinaĭ, Y.G.: Gibbs measures in ergodic theory. Russ. Math. Surv. 27, 21–70 (1972)
Wojtkowski, M.P.: A system of one dimensional balls with gravity. Commun. Math. Phys. 126, 507–533 (1990)
Young, L.-S.: Statistical properties of dynamical systems with some hyperbolicity. Ann. Math. 147, 585–650 (1998)
Zaitsev, A.Y.: The accuracy of strong Gaussian approximation for sums of independent random vectors. Russ. Math. Surv. 68, 721–761 (2013)
Acknowledgements
The authors are grateful to E. Rio for useful discussions concerning the proof of Lemma 3.6. The authors are also grateful to the three referees for their careful reading and comments, which improved the presentation of the paper.
Funding
A.K. has been partially supported by the European Research Council (ERC) under the European Union’s Horizon 2020 research and innovation programme (grant agreement No 787304) and by EPSRC grant EP/V053493/1.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The authors declare that they have no financial interests.
Additional information
Communicated by C. Liverani.
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Cuny, C., Dedecker, J., Korepanov, A. et al. Rates in Almost Sure Invariance Principle for Nonuniformly Hyperbolic Maps. Commun. Math. Phys. 405, 223 (2024). https://doi.org/10.1007/s00220-024-05090-8
Received:
Accepted:
Published:
DOI: https://doi.org/10.1007/s00220-024-05090-8