Abstract
We study a model of multi-excited random walk with non-nearest neighbour steps on \(\mathbb {Z}\), in which the walk can jump from a vertex x to either \(x+1\) or \(x-i\) with \(i\in \{1,2,\dots ,L\}\), \(L\ge 1\). We first point out the multi-type branching structure of this random walk and then prove a limit theorem for a related multi-type Galton–Watson process with emigration, which is of independent interest. Combining this result and the method introduced by Basdevant and Singh (Probab Theory Relat Fields 141:3–4, 2008), we extend their result (w.r.t. the case \(L=1\)) to our model. More specifically, we show that in the regime of transience to the right, the walk has positive speed if and only if the expected total drift \(\delta >2\). This confirms a special case of a conjecture proposed by Davis and Peterson.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
Excited random walk is a model of non-markovian random walk in a cookie environment, in which the walker consumes a cookie (if available) upon reaching a site and makes a jump with transition law dynamically depending on the number of remaining cookies at its current position. The model of nearest-neighbour excited random walks has been extensively studied in recent years. Benjamini and Wilson [3] first studied once-excited random walks with a focus on higher-dimensional integer lattice. Later, Zerner [26] extended this model to multi-excited random walks and established a criterion for recurrence/transience of the model on \(\mathbb {Z}\). There are also notable results for asymptotic behaviour of the multi-excited model including criteria non-ballisticity/ballisticity [2] as well as characterization of the limit distribution in such specific regimes [7, 14, 17]. See also [15, 18, 20, 22]. For a literature review, we refer the reader to [16].
1.1 Description of the Model and the Main Result
We define a non-nearest-neighbour random walk \(\mathbf {X }:= (X_n)_{n\ge 0}\), which describes the position of a particle moving in a cookie environment on the integers \(\mathbb {Z}\) as follows. For any integer n, set \([n] = \{1, 2, \ldots , n\}\). Let M and L be positive integers, \(\nu \) and \((q_j)_{j\in [M]}\) be probability measures on \(\Lambda :=\{-L, -L+1,\dots ,-1,1 \}\). Initially, each vertex in \(\mathbb {Z}\) is assigned a stack of M cookies and we set \(X_0=0\). Suppose that \(X_n =x\) and by time n there are exactly remaining \(M-j+1\) cookie(s) at site x with some \(j\in [M]\). Before the particle jumps to a different location, it eats one cookie and jumps to site \(x+i\), \(i\in \Lambda \), with probability \(q_j(i)\). On the other hand, if the stack of cookies at x is empty then it jumps to site \(x+i\), \(i\in \Lambda \) with probability \(\nu (i)\). More formally, denote by \((\mathcal {F}_n)_{n\ge 0}\) the natural filtration of \(\mathbf {X }\). For each \(i\in \Lambda \)
where is the number of visits to vertex \(x\in \mathbb {Z}\) up to time n, and \(\omega : \mathbb {N}\times \Lambda \rightarrow [0,1]\) is the cookie environment given by
Throughout this paper, we make the following assumption.
Assumption A
-
(i)
The distribution \(\nu \) has zero mean.
-
(ii)
For each \(j\in [M]\), the distribution \(q_j\) has positive mean and \(q_j(1)<1\).
We call the process \(\mathbf {X }\) described above (L, 1) non-nearest neighbors excited random walk ((L, 1)-ERW, for brevity). It is worth mentioning that (L, 1)-ERW is a special case of excited random walks with non-nearest neighbour steps considered by Davis and Peterson in [5] in which the particle can also jump to non-nearest neighbours on the right and \(\Lambda \) can be an unbounded subset of \(\mathbb {Z}\). In particular, Theorem 1.6 in [5] implies that the process studied in this paper is
-
Transient to the right if the expected total drift \(\delta \), defined as
$$\begin{aligned} \delta :=\sum _{j=1}^M\sum _{\ell \in \Lambda } \ell q_j(\ell ) \end{aligned}$$(1)is larger than 1, and
-
Recurrent if \(\delta \in [0,1]\).
Additionally, Davis and Peterson conjectured that the limiting speed of the random walk exists if \(\delta >1\) and it is positive when \(\delta >2\). (see Conjecture 1.8 in [5]).
Recently, a sufficient condition for once-excited random walks with long forward jumps to have positive speed has been shown in [4]. However, the coupling method introduced in [4] seems to not be applicable to models of multi-excited random walks.
In the present paper, we verify Davis-Peterson conjecture for (L, 1)-ERW. More precisely, we show that \(\delta >2\) is a sufficient and necessary condition for (L, 1)-ERW to have positive limiting speed, under Assumption A.
Theorem 1.1
Under Assumption A,
-
(a)
if \(\delta >1\) the speed of (L, 1)-ERW \(\mathbf {X }\) exists, i.e. \(X_n/n\) converges a.s. to a non-negative constant v, and
-
(b)
if \(\delta >2\) we have that \(v >0\). If \(\delta \in (1, 2]\) then \(v =0\).
1.2 Summary of the Proof of Theorem 1.1.
Our proof strategy relies on the connection between non-nearest neighbor excited random walks and multi-type branching processes with migration. The idea can be traced back to the branching structures of nearest-neighbor excited random walks [2] and random walks in a random environment (see e.g. [13] and [10]). In the present paper, we introduce multi-type branching process with emigration and develop techniques from [2] to deal with various higher dimensional issues in our model.
The remaining parts of the paper are organized as follows. We first describe in Sect. 2 the multi-type branching structure of the number of backward jumps. This branching structure is formulated by a multi-type branching process with (random) migration \(\mathbf {Z}\) defined in Proposition 2.1. In Sect. 3, we next demonstrate a limiting theorem for a class of critical multi-type Galton–Watson processes with emigration. We believe that this result is of independent interest. In Sect. 4, we derive a functional equation related to limiting distribution of \(\mathbf {Z}\) (Propositions 4.2 and 4.3). Combining these results together with a coupling between \(\mathbf {Z}\) and a critical multi-type branching process with emigration (which is studied in Sect. 3), we deduce the claim of Theorem 1.1.
It is worth mentioning that the techniques introduced in this paper is unfortunately not applicable to the case of excited random walks having non-nearest-neighbour jumps to the right. We refer the reader to [4] for a recent work studying the speed of once-excited random walks with long forward steps.
2 Multi-branching Structure of Excited Random Walks
For any pair of functions f and g of one real or discrete variable, we write
-
\(f(x) \sim g(x)\) as \(x\rightarrow x_0\) if \(\lim _{x \rightarrow x_0} {f(x)}/{g(x)} =1,\)
-
\(f(x)=O(g(x))\) as \(x\rightarrow x_0\) if \(\limsup _{x\rightarrow x_0}|f(x)/g(x)| <\infty \) and
-
\(f(x)=o(g(x))\) as \(x\rightarrow x_0\) if \(\lim _{x\rightarrow x_0}f(x)/g(x)=0\).
Denote \(\mathbb {Z}_+:=\{0,1,2,\dots \}\) and \(\mathbb {N}=\mathbb {Z}_+\setminus \{0\}\). For any \(m,n \in \mathbb {Z}\), \(m\le n\), set \([m,n]_{\mathbb {Z}}:= \{m, \ldots , n\}\) and \([n]=\{1,2,\dots ,n\}\).
For each \(n\in \mathbb {Z}_+\), let \(T_n=\inf \{ k\ge 0: X_k=n\}\) be the first hitting time of site n. For \(i\le n-1\), define \(V_i^n=(V_{i,1}^n, V_{i,2}^n,\dots , V_{i,L}^n)\) where for \(\ell \in [L]\),
stands for the number of backward jumps from a site in the set \( i + \mathbb {N}\) to site \(i-\ell +1\) before time \(T_n\). Notice that \(T_n\) is equal to the total number of forward and backward jumps before time \(T_n\). In particular, the number backward jumps to site i before time \(T_n\) is equal to \(V_{i,1}^n\). On the other hand, between two consecutive forward jumps from i to \(i+1\), there is exactly one backward jump from \( i + \mathbb {N}\) to \( i - \mathbb {Z}_+\). Furthermore, for \(0\le i\le n-1\), before the first backward jump from \( i + \mathbb {N}\) to \( i - \mathbb {Z}_+\), the walk must have its first forward jump from i to \(i+1\). Therefore the number of forward jumps from i to \(i+1\) before time \(T_n\) is equal to \(1_{\{0\le i\le n-1\}}+\sum _{\ell =1}^LV_{i,\ell }^n\). As a result, we obtain
Assume from now that \((X_n)_{n\ge 0}\) is transient to the right. Notice that the walk spends only a finite amount of time on \(-\mathbb {N}\) and thus
It is worth mentioning that the above hitting time decomposition was mentioned by Hong and Wang [10], in which they studied random walks in random environment with non-nearest-neighbour jumps to the left. The idea can be traced back to the well-known Kesten–Kozlov–Spitzer hitting time decomposition for nearest-neighbour random walks in random environments [13].
Let \((\xi _n)_{n\in \mathbb {N}}\) be a sequence of independent random unit vectors such that the distribution of \(\xi _n=(\xi _{n,1},\xi _{n,2},\dots ,\xi _{n,L+1})\) is given by
where \(\mathbf {e}_{\ell }\) with \(\ell \in [L+1]\) is the standard basis of \(\mathbb {R}^{L+1}.\) If \(\xi _n=\mathbf {e}_{\ell }\) with \(\ell \in [L]\), we say that the outcome of the n-th experiment is an \(\ell \)-th type failure. Otherwise, if \(\xi _n={\mathbf {e}}_{L+1}\), we say that it is a success.
For \(m\in \mathbb {Z}_+\), we define the random vector \(A(m)=(A_1(m),\dots , A_L(m))\) such that for \(\ell \in [L]\),
In other words, the random variable \(A_{\ell }(m)\) is the total number of \(\ell \)-th type failures before obtaining \(m+1\) successes.
Let \((A^{(n)}(m))_{m\in \mathbb {Z}_+}\), with \(n\in \mathbb {N}\), be i.i.d. copies of the process \((A(m))_{m\in \mathbb {Z}_+}\). We define a L-dimensional process \(\mathbf {Z}=(Z_n)_{n\ge 0}=(Z_{n,1},Z_{n,2},\dots ,Z_{n,L})_{n\ge 0}\) such that \(Z_0\in \mathbb {Z}_{+}^L\) is independent of \((A^{(n)}(m))_{n\in \mathbb {N}, m\in \mathbb {Z}_+}\) and for \(n\ge 1\),
where for \(z=(z_1, z_1,\dots ,z_{L})\in \mathbb {R}^{L}\), we denote \(|z|=z_1+z_2+\dots +z_{L}\). Therefore \(\mathbf {Z}=(Z_n)_{n\ge 0}\) is a Markov chain in \(\mathbb {Z}_{+}^L\) and its transition law is given by
for \((k_{1},k_{2},\dots ,k_L), (j_{1},j_{2},\dots ,j_L)\in \mathbb {Z}_+^L\).
Proposition 2.1
Assume that \(Z_{0,\ell }=0\) for all \(\ell \in [L]\). Then for each \(n\in \mathbb {N}\), we have that \((V_{n-1}^n,V_{n-2}^n,\dots , V_{0}^n)\) has the same distribution as \((Z_0,Z_1,\dots ,Z_{n-1})\).
Proof
A backward jump is called \(\ell \)-th type of level i if it is a backward jump from a site in \(i + \mathbb {N}\) to site \(i-\ell +1\). Recall that \(V_{i,\ell }^n\) is the number of \(\ell \)-th type backward jumps of level i before time \(T_n\). Assume that \(\{V_{i}^n=(V_{i,1}^n,\dots ,V_{i,L}^n)=(j_1,j_2,\dots ,j_L)\}\). The number of forward jumps from i to \(i+1\) before time \(T_n\) is thus equals to \(1+\sum _{\ell =1}^L V_{i,\ell }^n=1+\sum _{\ell =1}^L j_\ell \).
For each \(i\in \mathbb {Z}\), denote by \(T_{i}^{(k)}\) the time for k-th forward jump from \(i-1\) to i and also set \(T^{(0)}_i=0\). We have that \(T_{i}^{(1)} = T_i\). Moreover, as the process \(\mathbf {X }\) is transient, we have that only finitely many \((T_i^{(k)})_k\) are finite, and conditioning on \(\{V_{i}^n=(j_1,j_2,\dots ,j_L)\}\), we have that \(T_i^{k} < \infty \) for \(k \le 1+\sum _{s=1}^L j_s\).
Note that \(V_{i-1,\ell }^n\), i.e. the number of \(\ell \)-th type backward jumps of level \(i-1\) before time \(T_n\), is equal to the sum of number of \(\ell \)-th type backward jumps of level \(i-1\) during \([T_{i+1}^{(k-1)},T_{i+1}^{(k)}-1]_{\mathbb {Z}}\) for \(k \in [1+\sum _{k=1}^L j_k]\).
By the definition of \(T_{i}^{(k)}\), the walk will visit i at least once during the time interval \([T_{i+1}^{(k-1)},T_{i+1}^{(k)}-1]_{\mathbb {Z}}\). Whenever the walk visits i, it will make a forward jump from i to \(i+1\) (which corresponds to a success) or a backward jump from i to \(i-\ell \), i.e. a \(\ell \)-th type jump of level \(i-1\), with \(\ell \in [L]\). If the latter happens, then i will be visited again during \([T_{i+1}^{(k-1)},T_{i+1}^{(k)}-1]_{\mathbb {Z}}\). Moreover, an \(\ell \)-th type backward jump of level i is also an \((\ell -1)\)-th type backward jump of level \(i-1\). Thus conditionally on \(\{(V_{i,1}^n,\dots ,V_{i,L}^n)=(j_1,j_2,\dots ,j_L)\}\), the random vector \((V_{i-1,1}^n,V_{i-1,2}^n,\dots , V_{i-1,L}^n)\) has the same distribution as
\(\square \)
Recall that M is the total number of cookies initially placed on each site. By the definition of sequence \((A(m))_{m\in \mathbb {Z}_+}\) given in (3), we can easily obtain the following.
Proposition 2.2
For \(m\ge M-1\), we have
where \((\eta _k)_{k\ge 1}\) are i.i.d. random vectors independent of \(A({M-1})\) with multivariate geometrical law
In the above formula, we use the convention that \(0^0=1\).
Remark 2.1
The multivariate Markov chain \({\mathbf {Z}}\) defined in (4) can be interpreted as a multi-type branching process with (random) migration as follows. Let \(M'\) be a fixed integer such that \(M'\ge M-1\) and suppose that \(Z_{n-1}=j=(j_1,j_2,\dots ,j_L)\). Then
-
If \(|j|=j_1+j_2+\dots +j_L\ge M'\), we have
where \(\widetilde{j}:=(j_2,\dots , j_{L},0)\); \(\eta _k^{(n)}\) with \(k\in \mathbb {N}\) and \(n\in \mathbb {N}\) are i.i.d. random vectors with the multivariate geometrical law defined in (5); and for each \(n\in \mathbb {N}\), \((\eta _k^{(n)})_{k\in \mathbb {N}}\) is independent of \(A^{(n)}(M')\) and \((Z_{k})_{0\le k\le n-1}\). In this case, there is an emigration of \(M'\) particles (each particle of any type has the same possibility to emigrate) while all the remaining \(|j|-M'\) particles reproduce according to the multivariate geometrical law defined in (5). For each \(\ell \in [L]\), there is also an immigration of \(A_{\ell }^{(n)}(M')+j_{\ell +1}\) new \(\ell \)-th type particles (here we use the convention that \(j_{L+1}=0\)).
-
If \(|j|< M'\), we have \(Z_{n}= A^{(n)}({|j|})+\widetilde{j}.\) In this case, for each \(\ell \in [L]\), all \(j_{\ell }\) particles of \(\ell \)-th type emigrate while \(A_{\ell }^{(n)}(|j|)+j_{\ell +1}\) new particles of \(\ell \)-th type immigrate.
Proposition 2.3
The Markov chain \(\mathbf {Z}\) is ergodic.
Proof
Recall from (4) that for each \(n\in \mathbb {Z}\),
Taking L iterations, we have that for \(n\ge L\),
Define \(L'=\max \{\ell \in [L] : \nu (-\ell )>0\}\) and \(d_{\ell }=\text {Card}\big \{k\in [M]:q_k(-\ell )>0\big \}\). Set \(\mathcal {S}=\mathcal {S}_{1}\times \mathcal {S}_{2}\times \cdots \times \mathcal {S}_{L}\) where
It is evident from from (6) that for \(n\ge L\), the support of \(Z_n\) is equal to \({\mathcal {S}}\). In particular, \(\mathbb {P}(Z_{{L}}=j|Z_0=i)>0\) and \(\mathbb {P}(Z_{{L+1}}=j|Z_0=i)>0\) for any \(j\in \mathcal {S}, i\in {\mathbb {Z}}_+^L\). Hence \(\mathbf {Z}\) is irreducible and aperiodic. Using Proposition 2.1, we have that conditional on \(\{Z_0=(0,0,\dots ,0)\}\), \(Z_{n-1}\) has the same distribution as \(V_{0}^n\). As the process \(\mathbf {X }\) is transient, \(V_0^n\) convergences almost surely to \(V_0^{\infty }=(V_{0,1}^{\infty },V_{0,2}^{\infty },\dots , V_{0,L}^{\infty })\) where \(V_{0,\ell }^{\infty }\) is the total number of jumps from a site in \(\mathbb {N}\) to site \(-\ell +1\). Hence \(Z_n\) converges in law to some a.s. finite random vector \(Z_{\infty }\) as \(n\rightarrow \infty \). This implies that \(\mathbf {Z}\) is positive recurrent. Hence \(\mathbf {Z}\) is ergodic. \(\square \)
3 Critical Multi-Type Galton–Watson Branching Process with Emigration
In this section, we prove a limit theorem (see Theorem 3.1 below) for critical multi-type Galton–Watson processes with emigration. This result will be used to solve the critical case \(\delta =2\) in Sect. 4.
Definition 1
Let \(N=(N_1,N_2,...,N_L)\) be a vector of L deterministic positive integers and \((\psi (k,n))_{k,n\in \mathbb {N}}\) be a family of i.i.d. copies of a random matrices \(\psi \) such that \(\psi \) takes values in \(\mathbb {Z}_+^{L\times L}\) and its rows are independent. Let \((U(n))_{n\ge 0}\) be a Markov chain in \(\mathbb {Z}_{+}^L\) defined recursively by
where . We call \((U(n))_{n\ge 0}\) a multi-type Galton–Watson branching process with \((N_1,N_2,\dots , N_L)\)-emigration.
We can interpret the branching process \((U(n))_{n\ge 0}\) defined above as a model of a population with L different types of particles in which \(U_i(n)\) stands for the number of particles of type i in generation n. The number of offsprings of type j produced by a particles of type i has the same distribution as \(\psi _{i,j}\). In generation n, if \(U_i(n)\ge N_i\) for all \(i\in [L]\) then there is an emigration of \(N_i\) particles of type i for \(i\in [L]\), otherwise all the particles emigrate and \(U(n+1)=(0,0,\dots , 0)\).
Let \(|\cdot |\), \(\Vert \cdot \Vert \) and \(\langle \cdot ,\cdot \rangle \) stand for the 1-norm, the Euclidean norm and the Euclidean scalar product on \({\mathbb {R}}^L\) respectively. Denote \(\mathbf {0}=(0,0,\dots , 0)\), \(\mathbf {1}=(1,1,\dots , 1)\).
From now on, we always assume that \((U(n))_{n\ge 0}\) satisfies the following condition.
Assumption B
The branching process \((U(n))_{n\ge 0}\) is critical, i.e. the expected offspring matrix \({\mathbb {E}}[\psi ]\) is positively regular (in the sense that there exists \(n\in \mathbb {N}\) such that \(({\mathbb {E}}[\psi ])^{n}\) is a positive matrix) and \(\lambda =1\) is its largest eigenvalue in modulus.
By Perron-Frobenius theorem, the maximal eigenvalue 1 is simple and has the positive right and left eigenvectors \(u=(u_1,\dots ,u_L)\), \(v=(v_1,\dots ,v_L)\) which are uniquely defined such that \(\langle u,v\rangle =1\) and \(|u|=1\). Furthermore, other eigenvalue than 1 is strictly smaller than 1 in modulus.
Set
We will prove the following theorem, which is a multivariate extension of the limit theorem for critical (one-type) branching processes with emigration obtained in [23, 25] and [12] (see also [24] for a literature review).
Theorem 3.1
Let \(K=(K_1,K_2,\dots , K_L)\) be a deterministic vector such that \(K_i\ge N_i\) for all \(i\in [L]\). Assume that \(U(0)=K\) a.s. and there exists \(\varepsilon >0\) such that
where we denote by \(\lceil \theta \rceil \) the ceiling value of \(\theta \). Then the followings hold true:
-
a.
There exists a constant \(\varrho >0\) such that
$$\begin{aligned} \mathbb {P}(U(n)\ne \mathbf {0})\sim \frac{\varrho }{n^{1+\theta }}\quad \text {as } n\rightarrow \infty . \end{aligned}$$ -
b.
We have
$$\begin{aligned} \liminf _{n\rightarrow \infty }\frac{{\mathbb {E}}[|U_n|\ | \ U(n)\ne \mathbf {0}]}{n}\ge \beta . \end{aligned}$$
For each \(n\in \mathbb {Z}_+\) and \(s=(s_1,s_2,\dots , s_L)\in [0,1]^L\), we set
which stand respectively for the multivariate probability generating functions of U(n) and the random row vectors \((\psi _{1,{\ell }})_{{\ell }=1}^L\), \((\psi _{2,{\ell }})_{{\ell }=1}^L, \dots , (\psi _{L,{\ell }})_{{\ell }=1}^L\).
Let \(f^k=(f^k_1,f^k_2,\dots ,f^k_L)\) be the k-th iteration of f, i.e. \(f^0(s)=s\) and \(f^{k+1}(s)=f(f^{k}(s))\) for \(k\ge 0\). We also set
In order to prove Theorem 3.1, we will need the following lemmas.
Lemma 3.2
Assume that there exists \(\varepsilon >0\) such that
Then there exists a positive constant C such that
and
where \(\Gamma \) is the Gamma function defined by \(\Gamma (x)=\int _0^{\infty } u^{x-1}e^{-u}\mathrm{d}u\) for \(x>0\).
Proof
We denote by \({\mathbf {1}}\) the L-dimentional vector with all entries equal to 1. Note that
(see, e.g., Corollary V.5, p. 191 in [1]). Set \(r_n:=\left\| \mathbf {1}-f^n(\mathbf {0}) -\frac{u }{\beta n}\right\| \). We first show that
Indeed, let \(\mathcal Q(s)=({\mathcal {Q}}_1(s),{\mathcal {Q}}_2(s),\dots , {\mathcal {Q}}_L(s))\) be a vector of quadratic forms with
and set
We have
In virtue of Taylor’s expansion, we have \(\mathbf{1}-f(s)=({\mathbb {E}}[\psi ]-H(s))\cdot ({\mathbf {1}}-s)\) with \(H(s)=O(\Vert \mathbf{1}-s\Vert )\). It follows that
Using (9), we note that \(\Vert H(f^n(\mathbf {0}))\Vert =O(\Vert {\mathbf {1}}-f^n(\mathbf {0})\Vert )=O(1/n)\). Moreover, \({\mathbb {E}}[\psi ]^n=uv^T+O(|\lambda |^{n})\) where \(\lambda \) (with \(|\lambda |<1\)) is the eigenvalue of \({\mathbb {E}}[\psi ]\) with the second largest modulus. In what follows, we denote by \({\textsf {Cst}}\) a positive constant but its value may vary from line to line. Applying inequality (4.11) in [11] to (12), we deduce that
Since \({\mathcal {Q}}(s)\) is Lipschitz, we thus have
As a result, we have
W.l.o.g., we assume that \(\varepsilon \in (0,1)\) (which satisfies (8)). By Taylor’s expansion, there exists a vector function \({\mathcal {E}}(t,s)\) such that \(\mathbf{1}-f(s)={\mathbb {E}}[\psi ]({\mathbf {1}}-s)-{\mathcal {Q}}({\mathbf {1}}-s)+\mathcal E({\mathbf {1}}-s,{\mathbf {1}}-s)\) where we note that
According to inequality (4.42) in [11], we have
By reason of (13), we notice that
Combining (16) with (15)-(17) and using Cauchy-Schwarz inequality, we obtain
Combining (11) with (14) and (18), we get
Consequently,
Hence \(r_n=O(n^{-1-\varepsilon })\) and (10) is thus proved. On the other hand, by Taylor’s expansion, we have
Thus
Since \(\sum _{k=1}^n1/k= \log (n)+O(1)\) as \(n\rightarrow \infty \) and \(\theta =\langle N,u \rangle /{\beta }\), we obtain that \(\gamma _n(\mathbf {0})\sim C n^{\theta }\) for some positive constant C. Furthermore, by Hardy–Littlewood tauberian theorem for power series (see e.g. Theorem 5, Section XIII.5, p. 447 in [8]), we deduce that
\(\square \)
In what follows, for \(x,y \in \mathbb {Z}_+^{L}\) we write \(x\succeq y\) if \(x_{i}\ge y_i\) for all \(i\in [L]\), otherwise we write \(x\nsucceq y\). Set \(\mathcal {S}(N)=\{r\in {\mathbb {Z}_{+}^L\setminus \{\mathbf {0}\}}: r\nsucceq N \}\). For each \(r\in {\mathbb {Z}}_{+}^L\) and \(s=(s_1,s_2,\dots , s_L)\in \mathbb {R}^L\) such that \(s_{\ell }\ne 0\) for all \(\ell \in [L]\), define
For each \(n\in \mathbb {Z}_+\) and \(z\in [0,1)\), set
Lemma 3.3
The generating function of \((\mu _n)_{n\ge 0}\) is given by
in which we define
Proof
From the definition of U(n), we have
Consequently,
Note that \(\mu _n=1-F(\mathbf {0},n)\). Substituting \(s=\mathbf {0}\) into (21), we obtain
Multiplying both sides of the above equation by \(z^n\) and summing over all \(n\ge 0\), we get
This ends the proof of the lemma. \(\square \)
Let \((p_{n})_{n\ge 0}\) be a non-negative sequence such that its generating function \(P(z)=\sum _{n=0}^{\infty }p_nz^n\) has radius of converge 1. For each \(k\in \mathbb {Z}_+\), we define the sequence \((p_n^{(k)})_{n\ge 0}\) recursively by
Denote by \(P^{(k)}(z)=\sum _{n=0}^{\infty }p_{n}^{(k)}z^n\) the generating function of \((p_n^{(k)})_{n\ge 0}\). Assume that \(P^{(j)}(1^-)<\infty \) for all \(0\le j\le k\). By Abel’s theorem, we notice that \(P^{(j)}(1^-)=P^{(j)}(1)=\sum _{n=0}^{\infty }p_{n}^{(j)}\) and
The next lemma is derived directly from Corollary 2 and Lemma 5 in [23].
Lemma 3.4
Let \((p_n)_{n\ge 0}\) be a non-negative non-increasing sequence and \(p_n\rightarrow 0\) as \(n\rightarrow \infty \).
-
i.
Assume that \(p^{(1)}_n\sim n^{-\alpha }\) as \(n\rightarrow \infty \) for some \(\alpha >0\). Then there exists a constant \(\varpi >0\) such that \(p_n\sim \varpi n^{-\alpha -1}\) as \(n\rightarrow \infty \).
-
ii.
Assume that \(\sum _{k=1}^{n}kp_k\sim n^{2-\alpha }\) as \(n\rightarrow \infty \) for some \(0<\alpha <2\). Then \(p_n\sim (2-\alpha )n^{-\alpha }\) as \(n\rightarrow \infty \).
For \(r\in \mathbb {Z}_{+}^L\), \(n\in \mathbb {N}\) and \(k\in \mathbb {Z}_+\) set \(\pi _{r,n}:=\mathbb {P}(U(n-1)=r),\)
Recall that \(\mathcal {S}(N)=\{r=(r_1,r_2,\dots ,r_L)\in \mathbb {Z}_+^L\setminus \{\mathbf {0}\}: \exists \ell \in [L], r_{\ell }<N_{\ell } \}\).
Lemma 3.5
We have \(\sum _{r\in \mathcal {S}(N)}\Pi _r(1^{-}) \le 1.\) Furthermore, if \(Q^{(k)}(1^{-})<\infty \) for some \(k\ge 0\) then \(\sum _{r\in \mathcal {S}(N)}\Pi _r^{(k+1)}(1^{-})<\infty .\)
Proof
Define \(\tau =\inf \{n\ge 0: U(n+1)=\mathbf {0} \}\). For each \(r\in \mathcal {S}(N)\), we have
yielding that
We also have
By induction, we obtain that \(\sum _{r\in \mathcal {S}(N)}\pi _{r,n}^{(k+1)}\le \mu _{n}^{(k)}\) for all \(k\ge 0\). Hence
This ends the proof of the lemma. \(\square \)
In what follows, we set \(\widetilde{\theta }:=\lceil \theta \rceil \), which is the ceiling value of \(\theta \). We denote by \({\textsf {Cst}}\) a generic non-zero constant and its value may vary from line to line.
Lemma 3.6
Assume that the condition (7) is fulfilled. Then:
i. For each \( k\in \{0,1,\dots , \widetilde{\theta }-1\}\), there exist power series \(R_D^{[k]}(z), D_m^{[k]}(z)\) and non-zero constants \(d^{[k]}_m\), \(m\in \{k+1,\dots , \widetilde{\theta }\}\) such that as \(z\rightarrow 1^{-}\),
ii. There exist power series \(R_B(z), B_m(z)\) and \(b_m(z)\) with \(m\in \{1,2,\dots ,\widetilde{\theta }\}\) such that \(|b_m(1^-)|<\infty \),
as \(z\rightarrow 1^-\).
Proof
For \(k\in \mathbb {N}\), \(r\in \mathbb {Z}^L\setminus \{\mathbf {0}\}\) and \(j=(j_1,\dots , j_L)\in \mathbb {Z}_+^L\) with \(|j|\le \widetilde{\theta }+1\), we set
which is well-defined thanks to the condition (7). Using Taylor’s expansion, we have that for \(r=(r_1,\dots , r_L)\in \mathbb {Z}^L\setminus \{\mathbf {0}\}\), \(n\ge \widetilde{\theta }\) and \(k\ge 1\),
By reason of (9), we notice that
Recall from Lemma 3.2 that \(\gamma _n(\mathbf {0})\sim C n^{\theta }.\) Using Hardy–Littlewood tauberian theorem for power series, we thus have that as \(z\rightarrow 1^{-}\),
Part i. Notice that for \(n\in \mathbb {N}\), \(\gamma _{n}(\mathbf {0})-\gamma _{n-1}(\mathbf {0})=\gamma _{n}(\mathbf {0})\big (1-\prod _{\ell =1}^L(f^n_{\ell }(\mathbf {0}))^{N_{\ell }}\big )\) and thus
By induction, we easily obtain that for \( k\in \{0,1,\dots , \widetilde{\theta }-1\},\)
where we use the convention that \(f^{h}(\mathbf {0})=\mathbf {0}\) for \(h\le 0\). Using (31) and (35), we have that for \(k\in \{0, 1,\dots , \widetilde{\theta }-1\},\)
where for \(k\in \mathbb {Z}_{+}\) and \(j\in \mathbb {Z}_{+}^L\), we set
In virtue of Lemma 3.2 and (9), we note that
For \(1\le m\le \widetilde{\theta }+1\) and \(0\le k\le \widetilde{\theta }-1\), set
Note that \(D_{m,n}^{[k]}\sim d_m^{[k]} n^{\theta -m}\) as \(n\rightarrow \infty \). For \(m\in \{k+1,\dots , \widetilde{\theta }\}\), define \(D_m^{[k]}(z):=\sum _{n=\widetilde{\theta }}^{\infty } D_{m,n}^{[k]}z^{n}\). In view of Hardy–Littlewood tauberian theorem for power series, we thus obtain (24), (25) and (26).
Part ii. Recall that
in which we have
Using (31), we obtain that
where we set
Here we notice that for \(s\in \mathbb {R}^L_+\) with \(|s|< 1\) we have
It follows that
By (32) and (37), we also note that \(|\varrho _n^B(1^-)|<\infty \). Furthermore, in virtue of (33), we notice that as \(z\rightarrow 1^{-}\),
For \(m\in \{1,2,\dots , \widetilde{\theta }\}\), set
and define \(B_{m}(z):=\sum _{n=\widetilde{\theta }}^{\infty }B_{m,n}(z)z^{n}.\) By (37), we note that \(|b_m(1^-)|<\infty \). In view of Hardy–Littlewood tauberian theorem and (38), we obtain (27), (28) and (29). \(\square \)
We notice that
Proof of Theorem 3.1
Part a. We adopt an idea by Vatutin (see [23]) as follows. Recall that \(\widetilde{\theta }:=\lceil \theta \rceil \) is the ceiling value of \(\theta \). We will prove that for all \(k\in \{0,1,\dots , \widetilde{\theta }-1\}\),
and as \(z\rightarrow 1^-\),
We first consider the case when \(\theta \) is not an positive integer. By Hardy–Littlewood tauberian theorem for power series, it follows from (40) that \(\mu _{n}^{(\widetilde{\theta })}\sim {\textsf { Cst}}\cdot n^{\widetilde{\theta }-\theta -1}\) as \(n\rightarrow \infty \). If \(\theta \in \mathbb {N}\) then it follows from (41) and Hardy–Littlewood tauberian theorem for power series that \(\sum _{k=1}^n k \mu ^{(\theta )}_k\sim {\textsf { Cst}} \cdot n\) as \(n\rightarrow \infty \). In the latter case, by Lemma 3.4.ii, we obtain that \(\mu ^{(\theta )}_n\sim {\textsf {Cst}}\cdot n^{-1}\) as \(n\rightarrow \infty \). In both cases, using Lemma 3.4.i, we deduce that \(\mu _{n}\sim \varrho n^{-\theta -1}\) for some constant \(\varrho >0\) as \(n\rightarrow \infty \).
Hence, to finish the proof of Theorem 3.1(a), we only have to verify (39), (40) and (41).
Set \(B^{[0]}(z)=B(z)\) and \(B^{[k]}(z)=Q^{(k-1)}(1^-)(1-z)^{k-1}D(z)-B^{[k-1]}(z)\) for \(1\le k\le \widetilde{\theta }.\) By induction on k, we notice that if \(Q^{(k-1)}(1^-)<\infty \) then
Assume that up to some \(k\in \{1,2,\dots \widetilde{\theta }\}\), the power series \(R^{[k-1]}_B(z)\), \(B_m^{[k-1]}(z)\), and \(b_m^{[k-1]}(z)\) are defined for all \(m\in \{k,k+1,\dots , \widetilde{\theta }\}\) such that \(|b_m^{[k-1]}(1^-)|<\infty \) and as \(z\rightarrow 1^{-}\),
yielding that \(Q^{(k-1)}(1^-)={b^{[k-1]}_k(1^-)}/{d^{[k-1]}_k}\) is finite. The above statement holds for \(k=1\) thanks to Lemma 3.6. We next prove that this statement also holds true when replacing k by \(k+1\). Indeed, by reason of Lemma 3.6, notice that as \(z\rightarrow 1^{-}\),
We also have that as \(z\rightarrow 1^{-}\),
where we set \(\widehat{b}_k(z):=(b_{k}^{[k-1]}(1^-)-b_{k}^{[k-1]}(z))/(1-z)\). For \(m\in \{k+1, \dots ,\widetilde{\theta } \}\), set
and
By the recurrence relation of \(b^{[k]}_m(z)\) and Lemma 3.5, we note that \(|b^{[k]}_m(1^-)|<\infty \). Therefore, as \(z\rightarrow 1^-\),
and thus \(Q^{(k)}(1^-)={b_{k+1}^{[k]}(1^-)}/{d^{[k]}_{k+1}}\) is finite. By the principle of mathematical induction, we deduce that (39) holds true for all \(k\in \{0,1,\dots , \widetilde{\theta }-1\}\).
We now have
By Lemma 3.2, we notice that
We also note that as \(z\rightarrow 1^-\),
Hence (40) is verified.
Assume now that \(\theta \in \mathbb {N}\). One can easily show by induction that \(\frac{\text {d}}{\text {d}z} R_{B}^{[k]}(z)\sim {\textsf {Cst}}\cdot (1-z)^{-1}\) as \(z\rightarrow 1^{-}\) for all \(k\in \{0,1,\dots , \theta \}\). Hence
On the other hand, using (34), we have
By reason of (19) and (20), we note that \(1-g(f^n(\mathbf {0}))\sim \theta n^{-1}+O(n^{-(1+\varepsilon )})\) for some \(\varepsilon \in (0,1)\) as \(n\rightarrow \infty \). Hence
where in the last equality we use Hardy–Littlewood tauberian theorem and the fact that \(\gamma _n(\mathbf {0})\sim C n^{\theta } \). Combining (42), (43), (44) and (45), we have
Hence (41) is verified.
Part b. Recall from the proof of Lemma 3.3 that
Differentiating both sides of the above equation at \(s={\mathbf {1}}\), we obtain that
Iterating the above equality and taking the scalar product with u, we have
where we recall that \(U(0)= K\) a.s. and u is the right eigenvector of \({\mathbb {E}}[\psi ]\) w.r.t. the maximal eigenvalue 1. Taking \(n\rightarrow \infty \) in (46) and using (22), we get
Hence we obtain
By reason of (23), \(\sum _{k=n}^{\infty } \sum _{r\in \mathcal {S}(N)}\mathbb {P}(U(k-1)=r)\le \mu _{n-1}=O(n^{-1-\theta })\). On the other hand, we notice that \(\sum _{k=n}^{\infty } \mu _{k}\sim \frac{\varrho }{\theta } n^{-\theta }\) and \(\langle U(n), u\rangle \le |U(n)|\cdot |u|=|U(n)|\). Hence
\(\square \)
4 Phase Transition for the Speed of ERW
We first prove the following result which in particular implies Part (a) of Theorem 1.1.
Theorem 4.1
Under Assumption A, we have that
where \(\varsigma =(2,1,\dots ,1)\).
Proof
In virtue of Proposition 2.3, the distribution of \(Z_{\infty }\) does not depend on \(Z_0\) and
Assume w.l.o.g. that \(Z_0=0\) almost surely. By reason of Proposition 2.1, \(\sum _{i=0}^{n-1}\langle Z_i, \varsigma \rangle \) is equal to \(\sum _{i=0}^{n-1}\langle V^n_i, \varsigma \rangle \) in distribution. Moreover, recall from (2) that \(T_n\sim n+\sum _{i=0}^{n-1} \langle V^n_i, \varsigma \rangle \) as \(n\rightarrow \infty \). Hence
Using an argument by Zerner (see the proof of Theorem 13 in [26]), we next show that \(X_n/n\) and \(\lim _{n\rightarrow \infty }n/T_n\) converge in probability to the same limit. Indeed, for \(n\ge 0\), set \(S_n=\sup \{k: T_k\le n\}\). We have \(T_{S_n}\le n< T_{S_n+1}\). It immediately follows that
Since \(n< T_{S_n+1}\) and before time \(T_{S_n+1}\), the walk is always below level \(S_n+1\), we must have \(X_n\le S_n\). On the other hand \(X_{n}=S_n+X_n-X_{T_{S_n}}\ge S_n -L(n-T_{S_n}).\) It follows that
Using (48), we note that
As a result,
Furthermore, \(X_n/n\) converges almost surely as \(n\rightarrow \infty \) thanks to Proposition 3.1 in [4]. We note that the proof of Proposition 3.1 in [4] generally holds for transient multi-excited random walk with long jumps both to the left and the right assuming that
where \(\mathrm {supp}\,(\mu )\) stands for the support of measure \(\mu \). This ends the proof of the theorem. \(\square \)
Theorem 4.1 also implies that Part (b) of Theorem 1.1 is equivalent to the following theorem.
Theorem 4.2
\({\mathbb {E}}[Z_{\infty ,\ell }]<\infty \) for all \(\ell \in [L]\) if and only if \(\delta >2.\)
To prove the above theorem, we will need some preliminary results. The next proposition is immediate from the proof of Proposition 3.6 in [2].
Proposition 4.1
Suppose that for \(s\in [0,1]\)
where
-
I.
a(s) and b(s) are analytic functions in some neighborhood of 1 such that \(a(1)=1, a'(1)=\delta \) for some \(\delta >1\) and \(b(1)=0\);
-
II.
G is a function defined on [0, 1] such that G is left continuous function at 1, \(G'(1^-)\in (0,\infty ]\) and there exists \(\epsilon \in (0,1)\) such that \(G^{(i)}(s)>0\) for each \(s\in (1-\epsilon ,1)\) and \(i\in \mathbb {N}\).
Then, the following statements hold true:
-
(i)
\(b'(1)=0\).
-
(ii)
If \(\delta >2\) then \(b''(1)>0\) and \(1-G(1-s)=\frac{b''(1)}{2(\delta -2)}s+O(s^{2\wedge (\delta -1)}) \) as \(s\downarrow 0\).
-
(iii)
If \(\delta =2\) and \(b''(1)=0\) then \(G^{(i)}(1^{-})<\infty \) for all \(i\in \mathbb {N}\).
-
(iv)
If \(\delta =2\) and \(b''(1)\ne 0\) then \(1-G(1-s)\sim C s |\ln (s)|\) as \(s\downarrow 0\) for some constant \(C>0\).
In this section, we consider the following function
Notice that \(G'(1)={\mathbb {E}}\Big [ \sum _{\ell =1}^L \ell Z_{\,\infty , \ell }\Big ] \). For \(s_0\in \left( \frac{L-1}{L},1\right] \), we have
It follows that G(s) can be expanded as a power series of \(s-s_0\) with all positive coefficients. As a consequence, \(G^{(i)}(s)>0\) for all \(s\in \left( \frac{L-1}{L},1\right) \) and \(i\in \mathbb {N}\). Hence the condition II of Proposition 4.1 is verified. We next show (in Proposition 4.2 and Proposition 4.3 below) that there exist functions a and b such that the condition I and the functional equation (49) (with G given by (50)) are fulfilled.
Recall that \((\eta _n)_{n\ge 1}\) is a sequence of i.i.d. random vectors with multivariate geometrical law defined in (5). For \(\ell \in [L]\), set \(\rho _{\ell }=\nu (-\ell )/\nu (1)\). Notice that the probability generating function of \(\eta _1\) is given by
for \(s=(s_1,s_2,\dots , s_L)\) such that \(\sum _{\ell =1}^Ls_{\ell }\nu (-\ell )<1\).
Lemma 4.3
For \(\ell \in [L]\),
Proof
Set \(S=\sum _{i=1}^M \xi _{i,L+1}\). We have that \({\mathbb {E}}[S]=\sum _{i=1}^Mq_i(1),\) and
On the other hand, recall that
Hence \(\sum _{i=M+1}^{\gamma _{M-1} } \xi _{i,\ell }\) has the same distribution as \(\sum _{i=1}^{M-S} \eta _{i,\ell }.\) Therefore
\(\square \)
We can combine Proposition 2.2 with Proposition 4.3 to compute \({\mathbb {E}}[A(m)]\) for \(m \ge M\). We also note from (51) that \({\mathbb {E}}[\eta _{1,\ell }]=\rho _{\ell }\) for \(\ell \in [L]\). As a result, for \(\ell \in [L ]\) and \(k\in \mathbb {Z}^L_+\) such that \(|k|=k_1+\dots +k_L\ge M-1\), we have
where we use the convention \(k_{L+1}=0\).
Proposition 4.2
The function G defined by (50) satisfies the functional equation (49) where we define
Proof
We have that
Recall that given \(\{Z_0=(k_1,k_2,\dots , k_L)\}\), the random vector \(Z_1=(Z_{1,1},Z_{1,2},\dots , Z_{1,L})\) has the same distribution as \(\left( A_1(|k|)+k_2,\dots , A_{L-1}(|k|)+k_L , A_L(|k|)\right) \). Using Proposition 2.2, we thus have
On the other hand, using (51) and the fact that \(\sum _{\ell =1}^L\ell \rho _{\ell }=1\), we obtain
Hence
Therefore
\(\square \)
Recall from (1) that the expected total drift \(\delta \) of the cookie environment is given by
We must have that \(\delta >1\) as \(\mathbf {X }\) is assumed to be transient to the right.
Proposition 4.3
The functions a(s) and b(s) defined in Proposition 4.2 satisfies the condition (I) of Proposition 4.1. More specifically,
where
Proof
Using Taylor’s expansion, we have
as \(s\rightarrow 0\), in which the third identity follows from Lemma 4.3 and the fact that \(\sum _{\ell =1}^L\ell \rho _{\ell }=1\). Furthermore,
\(\square \)
Remark 4.4
Note that the functional equation (49) has the same form with the one in [2]. However the coefficient function b defined in Proposition 4.2 is more complicated and strongly depends on the distribution of \(Z_{\infty }\) while the function G defined by (50) is not a probability generating function when \(L\ge 2\) and it will not give us the full information to compute b. Nevertheless, Proposition 4.1 still play a crucial role in the proof of Theorem 1.1.
Let \(\eta =(\eta ^1,\eta ^2,\dots , \eta ^{L})\) be a L-dimensional random vectors with multivariate geometric law defined by
for each \(i=(i_1,i_2,\dots , i_L)\in \mathbb {Z}^L_+\). Recall that the probability generating function of \(\eta \) is given by
In particular, we have \({\mathbb {E}}[\eta ]=(\rho _1,\rho _2,\dots , \rho _{L})\).
Let \((\vartheta (k,n))_{k,n\in \mathbb {N}}=(\vartheta _{i,j}(k,n), i,j\in [L] )_{k,n\in \mathbb {N}}\) be a sequence of i.i.d. \(L\times L\) random matrices such that its rows are i.i.d. copies of \(\eta \). Let \((W(n))_{n\ge 0}=(W_1(n),\dots , W_L(n))_{n\ge 0}\) be a multi-type branching process with \((N_1,N_2,\dots , N_L)\)-emigration such that for each \(n\in \mathbb {N}\) and \(j\in [L]\),
where
(here \(\delta _{i,j}\) stands for the Kronecker delta) and .
Lemma 4.5
Assume \(\nu (-L)>0\). Then \((W(n))_{n\ge 0}\) is a critical multi-type Galton–Watson process with \((N_1, N_2,\dots , N_L)\)-emigration according to Definition 1 and Assumption B.
Proof
We have
and notice that all the entries of
are positive as \(\rho _L=\nu (-L)/\nu (1)>0\). Hence \(\overline{\chi }\) is positively regular.
Applying the determinant formula (see, e.g., Theorem 18.1.1 in [9])
(in which \(\Sigma \) is an invertible matrix, \({\mathbf {x}}\) and \({\mathbf {y}}\) are column vectors) to \({\mathbf {x}}=(\rho _1,\rho _2,\dots ,\rho _L)^{\textsf {T}}\), \(\mathbf{y}=(1,1,\dots , 1)^{\textsf {T}}\),
we obtain
where the last equality follows from the fact that
As a result, \(\Phi (1)=0\) and
yielding that \(\lambda =1\) is a simple eigenvalue of \(\overline{\chi }\). Using the triangle inequality, we have that
Therefore all the eigenvalues of \(\overline{\chi }\) must lie in the closed unit disk. Hence \(\lambda =1\) is also the largest eigenvalue in modulus. This ends the proof of the lemma. \(\square \)
Remark 4.6
The right and left eigenvectors of the maximal eigenvalue \(\lambda =1\) are given respectively by
It is also clear that (7) holds true since \({\mathbb {E}}[\chi _{ij}^k]<\infty \) for all \(k\ge 1\) and \(i,j\in [L]\).
Proposition 4.4
Assume that \(\delta =2\). Then there exists a positive integer \(\kappa \) such that
Proof
Define \(\tau =\inf \{n\ge 1 :\ Z_n=\mathbf {0}\}\). Notice that \({\mathbb {E}}[\tau | Z_0=\mathbf {0}]<\infty \) as \(\mathbf {Z}\) is positive recurrent. Furthermore, for any function \(\pi :\mathbb {Z}^L\rightarrow \mathbb {R}_+\), we have (see e.g. Theorem 1.7.5 in [21])
Let \(L'=\max \{\ell \in [L]:\nu (-\ell )>0\}\) and \(z^*=(z^*_1,\dots ,z^*_L)\) with \(z^*_{\ell }=M\) for \(1\le \ell \le L'\) and \(z^*=0\) for \(L'+1\le \ell \le L\). Let \(\kappa \) be a fixed positive integer that we will choose later. By setting \(\pi (z)=\left( \sum _{\ell =1}^L z_{\ell }\right) ^{\kappa }\), we obtain
Using (6), we note that \(\mathbb {P}\big (Z_L=z^*| \tau>L , Z_0=\mathbf {0}\big )=\mathbb {P}\big (Z_L=z^*| Z_k\ne \mathbf {0} \text { for all }1\le k\le L-1 , Z_0=\mathbf {0}\big )>0\) and thus
We next use a coupling argument to estimate the order of \({\mathbb {E}}[|Z_{n\wedge \tau }|^{\kappa }| Z_0=z_*]\) as \(n\rightarrow \infty \). Recall from Remark 2.1 that
where we denote \(\widetilde{z}=(z_2,z_3,\dots , z_{L+1},0)\) for each \(z=(z_1,z_2,\dots , z_L)\in \mathbb {R}^L\); and \(\eta _k^{(n)}=(\eta _{k,1}^{(n)},\dots , \eta _{k,L}^{(n)})\) with \(k, n\in \mathbb {N}\) are i.i.d. copies of the random vector \(\eta \) which are independent of \(Z_0\). Let \((W(n))_{n\ge 0}\) be the multi-type branching process with N-emigration defined by (52), in which we set \(N=(N_1,N_2,\dots ,N_L)=(M-1,M,\dots ,M)\), \(W(0)=(M,M,\dots , M)\) and \(\vartheta _{i,j}(k,n)=\eta _{(L-1)M+k,j}^{(n)}\) for \(i,j\in [L]\) and \(k, n\in \mathbb {N}\). For the sake of simplicity, we assume from now on that \(L=L'\) (otherwise, by reducing the dimension of W(n), one can easily handle the case \(L'<L\) by the same argument used in the case \(L=L'\)). In this case, we note that \(z^*=(M,M,\dots ,M)\).
Recall that for L-dimensional vectors x and y, we write \(x\succeq y\) if \({x}_{\ell }\ge y_{\ell }\) for all \(\ell \in [L]\). Assuming \(Z_0=z^*=(M,M,\dots ,M)\), we will show that
Indeed, for \(n=1\), we have
Suppose that \(Z_{n-1}{\succeq } W(n-1)\) for some \(n\in \mathbb {N}\). If \(Z_{n-1,\ell }< N_{\ell }\) for some \(\ell \in [L]\) then \(W_{\ell }(n-1)\le Z_{n-1,\ell } < N_{\ell }\) and thus \(W(n)=\mathbf {0}\). On the other hand, if \(Z_{n-1,\ell }\ge N_{\ell }\) for all \(\ell \in [L]\) then \(|Z_{n-1}|\ge ML-1\) and thus for \(\ell \in [L]\),
in which we use the convention that \(Z_{n-1,L+1}=N_{L+1}=0\). Hence \(Z_n \succeq W(n)\). By the principle of mathematical induction, we deduce (55). It implies additionally that on the event \(\{Z_0=z^*\}\), \(W(n)=\mathbf {0}\) for all \(n\ge \tau \). Hence, for all \(n\in \mathbb {N}\),
where the last equality follows from the fact that \((W(n))_{n\ge 0}\) is independent of \(Z_0\).
On the other hand, \((W(n))_{n\ge 0}\) is a critical multi-type Galton–Watson process with \((M-1,M,\dots , M)\)-emigration. By Theorem 3.1, there exist positive constants \(c_1, c_2\) and \(\theta \) such that
Choose \(\kappa =\left\lfloor \theta \right\rfloor +1\). Using Jensen inequality, we thus have
Combining (53), (54), (56) and (57), we obtain that \({\mathbb {E}}[|Z_{\infty }|^{\kappa }]=\infty .\) \(\square \)
We now turn to the proof of our main result.
Proof of Part (b), Theorem 1.1
Remind that this part is equivalent to Theorem 4.2. As the function G defined by (50) satisfies the functional equation (49) and the conditions I, II of Proposition 4.1, it follows from Proposition 4.1(ii) that if \(\delta >2\) then
The above fact and (47) imply that the random walk \(\mathbf {X }\) has positive speed in the supercritical case \(\delta >2\).
Let us now consider the critical case \(\delta =2\). If \(b''(1)\ne 0\) then by the virtue of Proposition 4.1(iv), we must have \(G'(1^-)=\infty \). Hence, to prove that \(G'(1^-)=\infty \), it is sufficient to exclude the case \(b''(1)=0\). Assume now that \(b''(1)=0\). By Proposition 4.1(iii), \(G^{(i)}(1^-)<\infty \) for all \(i\in \mathbb {N}\). On the other hand, by Proposition 4.4, there exists a positive integer \(\kappa \) such that \({\mathbb {E}}[|Z_{\infty }|^{\kappa }]=\infty \) and thus \(G^{(\kappa )}(1^-)=\infty \), which is a contradiction. Hence \(G'(1^-)=\infty \) and thus there exists \(\ell \in [L]\) such that \({\mathbb {E}}[Z_{\infty ,\ell }]=\infty \). It follows that a.s. \(\lim _{n\rightarrow \infty }X_n/n=0.\)
The subcritical case can be solved by showing the monotonicity of the speed as follows. Assume that \(1<\delta <2\). There exist probability measures \(\widehat{q}_1, \widehat{q}_2,\dots , {\widehat{q}}_M\) on \(\{-L,-L+1,\dots ,-1,1\}\) such that \({\widehat{q}}_j(-\ell )\le q_j(-\ell )\) for each \(\ell \in [L]\), \(j\in [M]\) and
Let \(\widehat{\mathbf {X }}=({\widehat{X}}_n)_n\) be the (L, 1)-excited random walk w.r.t the cookie environment \(\widehat{\omega }\) defined by
We thus have that a.s. \(\lim _{n\rightarrow \infty }{\widehat{X}}_n/n=0\). Let \({\mathbf {Z}}\) and \(\widehat{{\mathbf {Z}}}\) be respectively the Markov chains associated with \(\mathbf {X }\) and \(\widehat{\mathbf {X }}\) as defined by (4). Let \(Z_{\infty }\) and \({\widehat{Z}}_{\infty }\) be their limiting distributions. For \(h, \widehat{h}, k\in \mathbb {Z}_+^L\) with \( h \succeq \widehat{h}\), we notice that
Applying Strassen’s theorem on stochastic dominance for Markov chains (see e.g. Theorem 5.8, Chapter IV, p. 134 in [19] or Theorem 7.15 in [6]), we have that \(\widehat{\mathbf{Z }}\) is stochastically dominated by \({\mathbf {Z}}\). In particular, \({\mathbb {E}}[\widehat{Z}_{\infty }]\le {\mathbb {E}}[{Z}_{\infty }]\). Combining the above fact and the speed formula (47), we conclude that a.s.
This ends the proof of our main theorem. \(\square \)
References
Athreya, K.B., Ney, P.E.: Branching processes, Die Grundlehren der mathematischen Wissenschaften, vol. 196. Springer, New York (1972)
Basdevant, A.-L., Singh, A.: On the speed of a cookie random walk. Probab. Theory Relat. Fields 141(3–4), 625–645 (2008)
Benjamini, I., Wilson, D.B.: Excited random walk. Electron. Commun. Probab. 8, 86–92 (2003)
Collevecchio, A., Hamza, K., Nguyen, T.-M.: Long range one-cookie random walk with positive speed. Stoch. Process. Appl. 142, 462–478 (2021)
Davis, B., Peterson, J.: Excited random walks with non-nearest neighbor steps. J. Theoret. Probab. 30(4), 1255–1284 (2017)
den Hollander, F.: Probability Theory: The Coupling Method. Leiden University, Leiden (2012)
Dolgopyat, D., Kosygina, E.: Scaling limits of recurrent excited random walks on integers. Electron. Commun. Probab. 17(35), 14 (2012)
Feller, W.: An Introduction to Probability Theory and Its Applications, vol. II, 2nd edn. Wiley, New York (1971)
Harville, D.A.: Matrix Algebra from a Statistician’s Perspective. Springer, New York (1997)
Hong, W., Wang, H.: Branching Structures Within Random Walks and Their Applications. Branching Processes and Their Applications. Lecture Notes Statistics, vol. 219, pp. 57–73. Springer, Berlin (2016)
Joffe, A., Spitzer, F.: On multitype branching processes with \(\rho \le 1\). J. Math. Anal. Appl. 19, 409–430 (1967)
Kaverin, S.V.: Refinement of limit theorems for critical branching processes with emigration. Teor. Veroyatnost. I Primenen. 35(3), 570–575 (1990)
Kesten, H., Kozlov, M.V., Spitzer, F.: A limit law for random walk in a random environment. Compos. Math. 30, 145–168 (1975)
Kosygina, E., Mountford, T.: Limit laws of transient excited random walks on integers. Ann. Inst. Henri Poincaré Probab. Stat. 47(2), 575–600 (2011)
Kosygina, E., Peterson, J.: Excited random walks with Markovian cookie stacks. Ann. Inst. Henri Poincaré Probab. Stat. 53(3), 1458–1497 (2017)
Kosygina, E., Zerner, M.P.W.: Excited random walks: results, methods, open problems. Bull. Inst. Math. Acad. Sin. (N.S.) 8(1), 105–157 (2013)
Kosygina, E., Zerner, M.P.W.: Excursions of excited random walks on integers. Electron. J. Probab. 19(25), 25 (2014)
Kosygina, E., Mountford, T., Peterson, J.: Convergence of random walks with Markovian cookie stacks to Brownian motion perturbed at extrema. Probab. Theory Relat. Fields 182(1–2), 189–275 (2022)
Lindvall, T.: Lectures on the Coupling Method. Dover Publications Inc, Mineola (2002)
Menshikov, M., Popov, S., Ramírez, A.F., Vachkovskaia, M.: On a general many-dimensional excited random walk. Ann. Probab. 40(5), 2106–2130 (2012)
Norris, J.R.: Markov Chains, Cambridge Series in Statistical and Probabilistic Mathematics, vol. 2. Cambridge University Press, Cambridge (1998)
van der Hofstad, R., Holmes, M.: Monotonicity for excited random walk in high dimensions. Probab. Theory Relat. Fields 147(1–2), 333–348 (2010)
Vatutin, V.A.: A critical Galton-Watson branching process with immigration. Teor. Verojatnost. I Primenen. 22(3), 482–497 (1977)
Vatutin, V.A., Zubkov, A.M.: Branching processes. II. J. Soviet Math. 67(6), 3407–3485 (1993)
Vinokurov, G.V.: On a critical Galton-Watson branching process with emigration. Teor. Veroyatnost. I Primenen. 32(2), 378–382 (1987)
Zerner, M.P.W.: Multi-excited random walks on integers. Probab. Theory Relat. Fields 133(1), 98–122 (2005)
Acknowledgements
The author would like to thank Andrea Collevecchio, Kais Hamza and the anonymous referees for their thorough reading and their constructive suggestions which improved the quality of the manuscript.
Funding
Open Access funding enabled and organized by CAUL and its Member Institutions.
Author information
Authors and Affiliations
Corresponding author
Additional information
Communicated by Eric A. Carle.
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Nguyen, TM. Speed of Excited Random Walks with Long Backward Steps. J Stat Phys 188, 3 (2022). https://doi.org/10.1007/s10955-022-02926-2
Received:
Accepted:
Published:
DOI: https://doi.org/10.1007/s10955-022-02926-2
Keywords
- Excited random walks
- Non-nearest neighbour random walks
- Multi-type branching processes with emigration