Abstract
We study multivariate stochastic recurrence equations (SREs) with triangular matrices. If coefficient matrices of SREs have strictly positive entries, the classical Kesten result says that the stationary solution is regularly varying and the tail indices are the same in all directions. This framework, however, is too restrictive for applications. In order to widen applicability of SREs, we consider SREs with triangular matrices and we prove that their stationary solutions are regularly varying with component-wise different tail exponents. Several applications to GARCH models are suggested.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
1.1 Results and Motivation
A multivariate stochastic recurrence equation (SRE)
is studied, where \(\mathbf{A}_t\) is a random \(d\times d\) matrix with nonnegative entries, \(\mathbf{B}_t\) is a random vector in \({{\mathbb {R}}}^d\) with nonnegative components and \((\mathbf{A}_t,\mathbf{B}_t)_{t\in {{\mathbb {Z}}}}\) are i.i.d. The sequence \((\mathbf{W}_t)_{t\in {{\mathbb {N}}}}\) generated by iterations of SRE (1.1) is a Markov chain; however, it is not necessarily stationary. Under some mild contractivity and integrability conditions (see e.g., [5, 6]), \(\mathbf{W}_t\) converges in distribution to a random variable \(\mathbf{W}\) which is the unique solution to the stochastic equation:
here, \((\mathbf{A},\mathbf{B})\) denotes a generic element of the sequence \((\mathbf{A}_t,\mathbf{B}_t)_{t\in {{\mathbb {Z}}}}\), which is independent of \(\mathbf{W}\) and the equality is meant in distribution. If we put \(\mathbf{W}_0{\mathop {=}\limits ^{d}}\mathbf{W}\), then the sequence \((\mathbf{W}_t)\) of (1.1) is stationary. Moreover, under suitable conditions a strictly stationary casual solution \((\mathbf{W}_t)\) to (1.1) can be written by the formula:
and \(\mathbf{W}_t{\mathop {=}\limits ^{d}}\mathbf{W}\) for all \(t\in {{\mathbb {Z}}}\).
Stochastic iteration (1.1) has been already studied for almost half a century, and it has found numerous applications to financial models (see e.g., Section 4 of [9, 12]). Various properties have been investigated. Our particular interest here is the tail behavior of the stationary solution \(\mathbf{W}\). The topic is not only interesting on its own, but it also has applications to, for example, risk management [12, 22, Sec. 7.3].
The condition for the stationary solution \(\mathbf{W}\) having power decaying tails dates back to Kesten [19]. Since then, the Kesten theorem and its extensions have been used to characterize tails in various situations. An essential feature of Kesten-type results is that tail behavior is the same in all coordinates. We are going to call it Kesten property. However, this property is not necessarily shared by all interesting models—several empirical evidences support the fact (see e.g., [17, 23, 29] from economic data). Therefore, SREs with solutions having more flexible tails are both challenging and desirable in view of applications.
The key assumption implying Kesten property is an irreducibility condition, and it refers to the law of \(\mathbf{A}\). The simplest example when it fails is a SRE with diagonal matrices \(\mathbf{A}=diag(A_{11},\ldots ,A_{dd})\). In this case, the solution can exhibit different tail behaviors in coordinates, which is not difficult to see by the univariate Kesten result. In the present paper, we consider a particular case of triangular matrices \(\mathbf{A}\), which is much more complicated and much more applicable as well. We derive the precise tail asymptotics of solution in all coordinates. In particular, we show that it may vary in coordinates, though it is also possible that in some coordinates we have the same tail asymptotics.
More precisely, let \(\mathbf{A}\) be a nonnegative matrix matrices such that \(A_{ii}>0,\,A_{ij}=0,\,i> j\ a.s.\) Suppose that \({{\mathbb {E}}}A_{ii}^{\alpha _i}=1\) holds for \(\alpha _i>0\) in each i such that \(\alpha _i\ne \alpha _j,\,i\ne j\). We prove that when \(x\rightarrow \infty \)
for \(\widetilde{\alpha }_i>0\) depending on \(\alpha _i,\ldots ,\alpha _d\). Here and in what follows, the notation ‘\(\sim \)’ means that the quotient of the left- and right-hand sides tends to 1 as \(x\rightarrow \infty \). For \(d=2\), the result (1.3) was proved in [10] with indices \(\widetilde{\alpha }_1= \min (\alpha _1, \alpha _2)\) and \(\widetilde{\alpha }_2=\alpha _2\). The dependency of \(\widetilde{\alpha }_1\) on \(\alpha _1,\alpha _2\) comes from the SRE: \( W_{1,t} = A_{11,t}W_{1,t-1}+A_{12,t}W_{2,t-1}+B_{1,t}\) where \(W_{1,t}\) is influenced by \(W_{2,t}\).Footnote 1 The same dependency holds in any dimension since \(W_i\) depends on \(W_{i+1},\ldots ,W_d\) in a similar but more complicated manner. In order to prove (1.3), we clarify this dependency with new notions and apply an induction scheme effectively.
The structure of the paper is as follows. In Sect. 2, we state preliminary assumptions and we show existence of stationary solutions to SRE (1.1). Section 3 contains the main theorem together with its proof. The proof follows by induction, and it requires several preliminary results. They are described in Sects. 4 and 5. Applications to GARCH models are suggested in Sect. 6. Section 7 contains discussion about constants of tails and open problems related to the subject of the paper.
1.2 Kesten’s Condition and Previous Results
We state Kesten’s result and briefly review the literature. We prepare a function
where \(\Vert \cdot \Vert \) is a matrix norm. The tail behavior of \(\mathbf{W}\) is determined by h. We assume that there exists a unique \(\alpha >0\) such that \(h(\alpha )=1\). The crucial assumption of Kesten is irreducibility. It can be described in several ways, neither of which seems simple and intuitive. We are going to state a weaker property, which is much easier to understand. For any matrix \(\mathbf{Y}\), we write \(\mathbf{Y}>0\) if all entries are positive. The irreducibility assumption yields that for some \(n\in {{\mathbb {N}}}\)
Then, the tail of \(\mathbf{W}\) is essentially heavier than that of \(\Vert \mathbf{A}\Vert \): It decays polynomially even if \(\Vert \mathbf{A}\Vert \) is bounded. Indeed, there exists a function \(e_{\alpha }\) on the unit sphere \({\mathbb {S}}^{d-1}\) such that
and \(e_{\alpha }(\mathbf{y}) > 0\) for \( y\in {\mathbb {S}}^{d-1}\cap [0,\infty )^d\), where \(\mathbf{y}'\mathbf{W}=\sum _{j=1}^d y_j W_j\) denotes the inner product. If \(\alpha \not \in {{\mathbb {N}}}\), then (1.5) implies multivariate regular variation of \(\mathbf{W}\), while if \(\alpha \in {{\mathbb {N}}}\), the same holds under some additional conditions (see [9, Appendix C]). Here, we say that a d-dimensional r.v. \(\mathbf{X}\) is multivariate regularly varying with index \(\alpha \) if
where \({\mathop {\rightarrow }\limits ^{v}}\) denotes vague convergence and \({\varvec{\Theta }}\) is a random vector on the unit sphere \({\mathbb {S}}^{d-1}=\{\mathbf{x}\in {{\mathbb {R}}}^d:|\mathbf{x}| =1 \}\). This is a common tool for describing multivariate power tails; see [3, 26, 27] or [9, p.279].
We proceed to the literature after Kesten. Later on, an analogous result was proved by Alsmeyer and Mentemeier [1] for invertible matrices \(\mathbf{A}\) with some irreducibility and density conditions (see also [20]). The density assumption was removed by Guivarc’h and Le Page [16] who developed the most general approach to (1.1) with signed \(\mathbf{A}\) having possibly a singular law. Moreover, their conclusion was stronger, namely they obtained existence of a measure \(\mu \) on \({{\mathbb {R}}}^d \) being the weak limit of
which is equivalent to (1.6). As a related work, in [8] the existence of the limit (1.7) was proved under the assumption that \(\mathbf{A}\) is a similarity.Footnote 2 See [9] for the details of Kesten’s result and other related results.
For all the matrices \(\mathbf{A}\) considered above, we have the same tail behavior in all directions, one of the reasons being a certain irreducibility or homogeneity of the operations generated by the support of the law of \(\mathbf{A}\). This is not the case for triangular matrices, since (1.4) does not hold then. Research in such directions would be a next natural step.Footnote 3 In this paper, we work out the case when the indices \(\alpha _1,\ldots \alpha _d>0\) satisfying \({{\mathbb {E}}}A_{ii}^{\alpha _i}=1\) are all different. However, there remain not a few problems being settled, e.g., what happens when they are not necessarily different. This is not clear even in the case of \(2\times 2\) matrices. A particular case \(A_{11}=A_{22}>0\) and \(A_{12} \in {{\mathbb {R}}}\) was studied in [11] where the result is
where \(C>0\) is a constant which may be different line by line. Our conjecture in \(d\, (>1)\) dimensional case is that
for some slowly varying functions \(\ell _i\), and to get optimal \(\ell _i\)’s would be a real future challenge.
2 Preliminaries and Stationarity
We consider \(d\times d\) random matrices \(\mathbf{A}=[A_{ij}]_{i,j=1}^d\) and d-dimensional random vectors \(\mathbf{B}\) that satisfy the set of assumptions:
Note that (T-1) and (T-4) imply that
for \(i=1,\ldots ,d\). For further convenience, any r.v. satisfying the inequality (2.1) will be called positive. Most conditions are similar to those needed for applying Kesten–Goldie’s result (see [9]).
Let \((\mathbf{A}_t,\mathbf{B}_t)_{t\in {{\mathbb {Z}}}}\) be an i.i.d. sequence with the generic element \((\mathbf{A},\mathbf{B})\). We define the products
where \(\mathbf{I}_d\) denotes the identity \(d\times d\) matrix. In the case \(t=0\), we are going to use a simplified notation:
so that \({\varvec{\Pi }}_0=\mathbf{I}_d\). For any \(t\in {{\mathbb {Z}}}\) and \(n\in {{\mathbb {N}}}\), the products \({\varvec{\Pi }}_{t,t-n+1}\) and \({\varvec{\Pi }}_n\) have the same distribution.
Let \(\Vert \cdot \Vert \) be the operator norm of the matrix: \(\Vert \mathbf{M}\Vert =\sup _{|\mathbf{x}|=1}|\mathbf{Mx}|\), where \(|\cdot |\) is the Euclidean norm of a vector. The top Lyapunov exponent associated with \(\mathbf{A}\) is defined by
Notice that in the univariate case \(\mathbf{A}=A\in {{\mathbb {R}}}\) and \(\gamma _{\mathbf{A}}={{\mathbb {E}}}[\log |A|]\).
If \(\gamma _{\mathbf{A}}<0\), then the equation
has a unique solution \(\mathbf{W}\) which is independent of \((\mathbf{A},\mathbf{B})\). Equivalently, the stochastic recurrence equation
has a unique stationary solution and \(\mathbf{W}_t{\mathop {=}\limits ^{d}}\mathbf{W}\). Then we can write the stationary solution as a series
Indeed, it is easily checked that the process \((\mathbf{W}_t)_{t=0}^\infty \) defined by (2.4) is stationary and solves (2.3), if the series on the right-hand side of (2.4) is convergent for any \(t\in {{\mathbb {N}}}\). The convergence is ensured if the top Lyapunov exponent is negative (for the proof see [5]). Negativity of \(\gamma _\mathbf{A}\) follows from condition (T). We provide a proof in Appendix A.
From now on, \(\mathbf{W}=(W_1,\ldots ,W_d)\) will always denote the solution to (2.2) and by \(\mathbf{W}_t=(W_{1,t},\ldots ,W_{d,t})\) we mean the stationary solution to (2.3). Since they are equal in distribution, when we consider distributional properties, we sometimes switch between the two notations, but this causes no confusion.
3 The Main Result and the Proof of the Main Part
3.1 The Main Result
We are going to determine the component-wise tail indices of the solution \(\mathbf{W}\) to stochastic equation (2.2). Since \(\mathbf{A}\) is upper triangular, the tail behavior of \(W_{i}\) is affected only by \(W_{j}\) for \(j>i\), but not necessarily all of them: We allow some entries of \(\mathbf{A}\) above the diagonal to vanish a.s. In the extreme case of a diagonal matrix, the tail of any coordinate may be determined independently of each other. On the other hand, if \(\mathbf{A}\) has no zeros above the diagonal, every coordinate is affected by all the subsequent ones. In order to describe this phenomenon precisely, we define a partial order relation on the set of coordinates. It clarifies interactions between them.
Definition 3.1
For \(i,j\in \{1,\ldots ,d\}\), we say that i directly depends on j and write \(i\preccurlyeq j\) if \(A_{ij}\) is positive (in the sense of (2.1)). We further write \(i\prec j\) if \(i\preccurlyeq j\) and \(i\ne j\).
Observe that \(i\preccurlyeq j\) implies \(i\le j\) since \(\mathbf{A}\) is upper triangular, while \(i\preccurlyeq i\) follows from the positivity of diagonal entries. We extract each component of SRE (2.3) and may write
where in the latter sum all coefficients \(A_{ij,t}\) are positive. From this, we obtain the component-wise SRE in the form
where
Therefore, the tail of \(W_{i,t}\) is determined by the comparison of the autoregressive behavior, characterized by the index \(\alpha _i\), and the tail behavior of \(D_{i,t}\), which depends on the indices \(\alpha _{j},\,j\succ i\). To clarify this, we define recursively new exponents \(\widetilde{\alpha }_i\), where i decreases from d to 1:
If there is no j such that \(i\prec j\), then we set \(\widetilde{\alpha }_i=\alpha _i\). In particular, \(\widetilde{\alpha }_d=\alpha _d\). Notice that in general \(\widetilde{\alpha }_i \ne \min \{\alpha _j: i\le j\}\). Depending on zeros of \(\mathbf{A}\), two relations \(\widetilde{\alpha }_i = \min \{\alpha _j: i\le j\}\) and \(\widetilde{\alpha }_i > \min \{\alpha _j: i\le j\}\) are possible for any \(i<d\), see Example 3.4.
For further convenience, we introduce also a modified, transitive version of relation \(\preccurlyeq \).
Definition 3.2
We say that i depends on j and write \(i\trianglelefteq j\) if there exists \(m\in {{\mathbb {N}}}\) and a sequence \((i(k))_{0\le k\le m}\) such that \(i=i(0)\le \cdots \le i(m)=j\) and \(A_{i(k)i(k+1)}\) is positive for \(k=0,\ldots ,m-1\). We write \(i\vartriangleleft j\) if \(i\trianglelefteq j\) and \(i\ne j\).
Equivalently, the condition on the sequence \((i(k))_{0\le k\le m}\) can be presented in the form \(i\preccurlyeq i(1)\preccurlyeq \cdots \preccurlyeq i(m)=j\). In particular, \(i\preccurlyeq j\) implies \(i\trianglelefteq j\). Now, we can write
which is equivalent to (3.4), though has a more convenient form.
Although Definitions 3.1 and 3.2 are quite similar, there is a significant difference between them. To illustrate the difference, we introduce the following notation for the entries of \({\varvec{\Pi }}_{t,s}\):
If \(t=0\), we use the simplified form
By \(i\preccurlyeq j\), we mean that the entry \(A_{ij}\) of the matrix \(\mathbf{A}\) is positive, while \(i\trianglelefteq j\) means that for some \(n\in {{\mathbb {N}}}\) the corresponding entry \(\pi _{ij}(n)\) of the matrix \({\varvec{\Pi }}_{n}\) is positive. The former relation gives a stronger condition. On the other hand, the latter is more convenient when products of matrices are considered, especially when transitivity plays a role. Example 3.4 gives a deeper insight into the difference between the two relations. Throughout the paper, both of them are exploited.
Now, we are ready to formulate the main theorem.
Theorem 3.3
Suppose that condition (T) is satisfied for a random matrix \(\mathbf{A}\). Let \(\mathbf{W}\) be the solution to (2.2). Then, there exist strictly positive constants \(C_1,\ldots ,C_d\) such that
The following example gives some intuition of what the statement of the theorem means in practice.
Example 3.4
Let
where \({{\mathbb {P}}}(A_{ij}>0)>0\) and other components are zero a.s. Suppose that \(\alpha _4<\alpha _3<\alpha _2<\alpha _5<\alpha _1\) and \(\mathbf{A}\) satisfies the assumptions of Theorem 3.3. Then, \(\widetilde{\alpha }_1=\alpha _4\), \(\widetilde{\alpha }_2=\alpha _4\), \(\widetilde{\alpha }_3=\alpha _3\), \(\widetilde{\alpha }_4=\alpha _4\) and \(\widetilde{\alpha }_5=\alpha _5\).
We explain the example step by step. The last coordinate \(W_5\) is the solution to 1-dimensional SRE, so its tail index is \(\alpha _5\). Since there is no j such that \(4 \vartriangleleft j\), \(\alpha _4\) is the tail index of \(W_4\). For the third coordinate, the situation is different: We have \(3\vartriangleleft 5\), so the tail of \(W_3\) depends on \(W_5\). But \(\alpha _3<\alpha _5\), so the influence of \(W_5\) is negligible and we obtain the tail index \(\widetilde{\alpha }_3=\alpha _3\). Inversely, the relations \(2\vartriangleleft 4 \) and \(\alpha _2>\alpha _4\) imply \(\widetilde{\alpha }_2=\alpha _4\). The first coordinate clearly depends on the second and fifth, but recall that the second one also depends on the fourth. Hence, we have to compare \(\alpha _1\) with \(\alpha _2,\alpha _4\) and \(\alpha _5\). The smallest one is \(\alpha _4\); hence, \(\widetilde{\alpha }_1=\alpha _4\) is the tail index. Although the dependence of \(W_1\) on \(W_4\) is indirect, we see it in the relation \(1\vartriangleleft 4\).
3.2 Proof of the Main Result
Since the proof includes several preliminary results which are long and technical, they are postponed to Sects. 4 and 5. To make the argument more readable, we provide an outline of the proof here, referring to those auxiliary results. In the proof of Theorem 3.3, we fix the coordinate number k and consider two cases: \(\widetilde{\alpha }_k=\alpha _k\) and \(\widetilde{\alpha }_k<\alpha _k\).
In the first case \(\widetilde{\alpha }_k=\alpha _k\), the proof is based on Goldie’s result [15, Theorem 2.3] and it is contained in Lemmas 4.2 and 4.3.
If \(\widetilde{\alpha }_k<\alpha _k\), the proof is more complicated and requires several auxiliary results. We proceed by induction. Let \(j_0\) be the maximal coordinate among \(j \vartriangleright k\) such that the modified tail indices \(\widetilde{\alpha }_j\) and \(\widetilde{\alpha }_k\) are equal. Then clearly, \(\widetilde{\alpha }_k=\alpha _{j_0}=\widetilde{\alpha }_\ell \) for any \(\ell \) such that \(k \trianglelefteq \ell \vartriangleleft j_0\). We start with the maximal among such coordinates (in the standard order on \({{\mathbb {N}}}\)) and prove (3.6) for it. Inductively we reduce \(\ell \) using results for larger tail indices and finally we reach \(\ell =k\).
We develop a component-wise series representation (5.4) for \(\mathbf{W}_0\). In Lemma 5.1, we prove that it indeed coincides a.s. with (2.4).
The series (5.4) is decomposed into parts step by step (Lemmas 5.4, 5.9 and 5.10). Finally, we obtain the following expression:
Our goal is to prove that as \(s\rightarrow \infty \), the tail asymptotics of the first term \(\pi _{\ell j_0}(s)W_{j_0,-s}\) approaches that of (3.6), while the second term \(R_{\ell j_0}(s)\) becomes negligible.
The quantity \(R_{\ell j_0}(s)\) is defined inductively by (5.31) and (5.39) as a finite collection of negligible parts, each being estimated in a different way. In the process of finding the main dominant term, we simultaneously settle the upper bounds for the negligible parts. This is done through (5.15) in Lemmas 5.4, 5.3 and (5.29) under conditions of Lemmas 5.9 and 5.10. Then, in the end \(R_{\ell j_0}(s)\) is proved to be negligible as \(s\rightarrow \infty \).
The final step is related to Breiman’s lemma applied to \(\pi _{\ell j_0}(s)W_{j_0,-s}\). By independence between \(\pi _{\ell j_0}(s)\) and \(W_{j_0,-s}\), we obtain the equality in asymptotics:
for fixed s. Recall that we need to let \(s\rightarrow \infty \) for \(R_{\ell j_0}(s)\) to be negligible. The existence of \(\lim _{s\rightarrow \infty } {{\mathbb {E}}}[\pi _{\ell j_0}(s)^{\alpha _{j_0}} ]\) is assured in Lemma 5.7. Then, eventually we get \({{\mathbb {P}}}(W_{\ell ,0}>x)\sim C\cdot {{\mathbb {P}}}(W_{j_0,-s}>x)\) for a positive constant C, which can be explicitly computed. Now, (3.6) follows from stationarity and Lemma 4.3.
We move to the proof of Theorem 3.3.
Proof of Theorem 3.3
Fix \(k\le d\). If \(\widetilde{\alpha }_k=\alpha _k\), then the statement of the theorem directly follows from Lemma 4.3. If \(\widetilde{\alpha }_k<\alpha _k\), then there is a unique \(j_0\vartriangleright k\) such that \(\widetilde{\alpha }_k=\widetilde{\alpha }_{j_0}=\alpha _{j_0}\). Another application of Lemma 4.3 proves that
for some constant \(C_{j_0}>0\). By stationarity, we set \(t=0\) without loss of generality.
The proof follows by induction with respect to number j in backward direction, namely we start with \(\ell =j_1:=\max \{j : k \trianglelefteq j \vartriangleleft j_0\}\) and reduce \(\ell \) until \(\ell =k\). Notice that \(j_1\prec j_0\), and thus \(\ell =j_1\) satisfies conditions of Lemma 5.10. Thus, there exists \(s_0\) such that for any \(s>s_0\) we have
where \(\pi _{\ell j_0}(s)\) is independent of \(W_{j_0,-s}\) such that \(0<{{\mathbb {E}}}\pi _{\ell j_0}(s)^{\alpha _{j_0}}<\infty \), and \(R_{\ell j_0}(s)\) satisfies (5.29). We are going to estimate \({{\mathbb {P}}}(W_{\ell ,0}>x)\) from both below and above. Since \(R_{\ell j_0}(s)\ge 0\,a.s.\)
holds, and by Breiman’s lemma [9, Lemma B.5.1] for fixed \(s>s_0\)
Combining these with (3.8), we obtain the lower estimate
Now, we pass to the upper estimate. Recall (5.29) in Lemma 5.10 which implies that for any \(\delta \in (0,1)\) and \(\varepsilon >0\) there exists \(s_1\,(> s_0)\) such that for \(s> s_1\)
Then, for fixed \(s\ge s_1\), we apply Lemma B.1, which is a version of Breiman’s lemma, to (3.9) and obtain
where we also use (3.8). We may let \(\varepsilon \rightarrow 0\) and \(\delta \rightarrow 0\) together with \(s\rightarrow \infty \) here and in (3.10). The existence and positivity of the limit \(u_\ell =\lim _{s\rightarrow \infty } u_\ell (s)\) is assured by Lemma 5.7. Thus, from (3.10) and (3.11) we have
This implies that
Now, we go back to the induction process.
If \(j_1=k\), then the proof is over, and if \(j_1\ne k\), we set \(j_2=\max \{j:k\trianglelefteq j\vartriangleleft j_0, j\ne j_1\}\). Then, there are two possibilities, depending on whether \(j_2\prec j_1\) or not. If \(j_2\not \prec j_1\), then the assumptions of Lemma 5.10 are satisfied with \(\ell =j_2\) and we repeat the argument that we used for \(j_1\). If \(j_2\prec j_1\), the assumptions of Lemma 5.9 are fulfilled with \(\ell =j_2\). Since the assertion of Lemma 5.9 is the same as that of Lemma 5.10, we can again repeat the argument that we used for \(j_1\).
In general, we define \(j_{m+1}=\max (\{j: k\trianglelefteq j \vartriangleleft j_0\}\setminus \{j_1,\ldots ,j_m\})\). If \(j_{m+1}\prec j\) for some \(j\in \{j_1,\ldots ,j_m\}\), then we use Lemma 5.9 with \(\ell =j_{m+1}\). Otherwise, we use Lemma 5.10 with the same \(\ell \). Then, by induction we prove (3.12) for every \(\ell :k\trianglelefteq \ell \vartriangleleft j_0\), particularly in the end we also obtain
\(\square \)
Notice that we have two limit operations, with respect to x and s, and always the limit with respect to x precedes. We cannot exchange the limits, namely we have to let \(s\rightarrow \infty \) with s depending on x.
4 Case \(\widetilde{\alpha }_k=\alpha _k\)
The assumption \(\widetilde{\alpha }_k=\alpha _k\) means that the tail behavior of \(W_k\) is determined by its auto-regressive property, namely the tail index is the same as that of the solution V of the stochastic equation \(V{\mathop {=}\limits ^{d}}A_{kk}V+B_k\). The tails of other coordinates on which k depends are of smaller order, which is rigorously shown in Lemma 4.2. In Lemma 4.3, we obtain the tail index of \(W_k\) by applying Goldie’s result. By Lemma 4.2, we observe that the perturbation induced by random elements other than those of kth coordinate \((A_{kk},B_k)\) is negligible.
In what follows, we work on a partial sum of the stationary solution (2.4) component-wisely. To write the coordinates of the partial sum directly, the following definition is useful.
Definition 4.1
For \(m\in {{\mathbb {N}}}\) and \(1\le i,j \le d\) let \(H_m(i,j)\) be the set of all sequences of \(m+1\) indices \((h(k))_{0\le k\le m}\) such that \(h(0)=i, h(m)=j\) and \(h(k)\preccurlyeq h(k+1)\) for \(k=0,\ldots ,m-1\). For convenience, the elements of \(H_m(i,j)\) will be denoted by h.
Notice that each of such sequences is non-decreasing since \(\mathbf{A}\) is upper triangular. Moreover, \(H_m(i,j)\) is nonempty if and only if \(i\trianglelefteq j\) and m is large enough. Now, we can write
Similar expression for \(\pi _{ij}(t,s)\) can be obtained by shifting the time indices,
which will be used later. Since the sum is finite, it follows from condition (T-5) that \({{\mathbb {E}}}\pi _{ij}(s)^{\widetilde{\alpha }_i}<\infty \) for any i, j. Moreover, if \(j\trianglerighteq i\), then \({{\mathbb {P}}}(\pi _{ij}(s)>0)>0\) for s large enough (in particular \(s\ge j-i\) is sufficient). By definition, \(\pi _{ij}(s)\) is independent of \(W_{j,-s}\). Notice that when \(i=\ell \), \(j=j_0\) and \(\ell \trianglelefteq j_0\), it is the coefficient of our targeting representation (3.7).
Lemma 4.2
For any coordinate, \(i\in \{1,\ldots ,d\}\), \({{\mathbb {E}}}W_i^\alpha <\infty \) if \( 0<\alpha <\widetilde{\alpha }_i\).
Proof
For fixed i, let us approximate \(\mathbf{W}_{0}\) by partial sums of the series (2.4). We will denote \(\mathbf{S}_n=\sum _{m=0}^{n}{\varvec{\Pi }}_{m}\mathbf{B}_{-m}\) and \(\mathbf{S}_n=(S_{1,n},\ldots ,S_{d,n})\). We have then
Suppose that \(\alpha \le 1\). Then, by independence of \(A_{h(p-1)h(p),-p},\,p=1,\ldots ,m-1\) and \(B_{j,-m}\),
To estimate the right-hand side, we will need to estimate the number of elements of \(H_m(i,j)\). To see the convergence of the series (4.2) it suffices to consider \(m>2d\). Recall that the sequences in \(H_m(i,j)\) are non-decreasing; thus, for a fixed j there are at most \(j-i\) non-diagonal terms in each product on the right-hand side of (4.2). The non-diagonal terms in the product coincide with the time indices t, for which the values \(h(t-1)\) and h(t) are different. If there are exactly \(j-i\) such time indices, then the values are uniquely determined. There are \(\left( {\begin{array}{c}m\\ j-i\end{array}}\right) \) possibilities of placing the moments among m terms of the sequence and \(\left( {\begin{array}{c}m\\ j-i\end{array}}\right) <\left( {\begin{array}{c}m\\ d\end{array}}\right) \) since \(m>2d\) and \(j-i<d\).
If we have \(l<j-i\) non-diagonal elements in the product, then there are \(\left( {\begin{array}{c}m\\ l\end{array}}\right) \) possibilities of placing them among other terms and there are less than \(\left( {\begin{array}{c}j-i\\ l\end{array}}\right) \) possible sets of l values. Hence, we have at most \(\left( {\begin{array}{c}m\\ l\end{array}}\right) \left( {\begin{array}{c}j-i\\ l\end{array}}\right) <\left( {\begin{array}{c}m\\ d\end{array}}\right) d!\) sequences for a fixed l. Moreover, there are \(j-i< d\) possible values of l, and hence there are at most \(d\cdot d!\cdot \left( {\begin{array}{c}m\\ d\end{array}}\right) \) sequences in \(H_m (i,j)\). Since \(\left( {\begin{array}{c}m\\ d\end{array}}\right) <\frac{m^d}{d!}\), the number of sequences in \(H_m(i,j)\) is further bounded by \(dm^d\).
Now, recall that there is \(\rho <1\) such that \({{\mathbb {E}}}A_{jj}^\alpha \le \rho \) for each \(j\vartriangleright i\) and that there is a uniform bound M such that \({{\mathbb {E}}}A_{jl}^\alpha <M\) and \({{\mathbb {E}}}B_j^\alpha <M\) whenever \(j \trianglerighteq i\), for any l. It follows that
uniformly in n, with \(C>0\), which is bounded from above. Hence, there exists the limit \(S_i=\lim _{n\rightarrow \infty }S_{i,n}\) a.s. and \({{\mathbb {E}}}S_i^\alpha <\infty \). By (2.4), we have \(\mathbf{W}_{0}=\lim _{n\rightarrow \infty }\mathbf{S}_n\ a.s.\) and we conclude that \({{\mathbb {E}}}W_{i,0}^\alpha <\infty \).
If \(\alpha >1\), then by Minkowski’s inequality we obtain
with \(C'>0\). The same argument as above shows the uniform convergence. Thus, the conclusion follows. \(\square \)
Suppose that we have \(\widetilde{\alpha }_k=\alpha _k\). This implies that \(\alpha _k<\widetilde{\alpha }_j\) for each \(j\vartriangleright k\) and hence, by Lemma 4.2, \({{\mathbb {E}}}W_j^{\alpha _k}<\infty \). The next lemma proves the assertion of the main theorem in case \(\widetilde{\alpha }_k=\alpha _k\).
Lemma 4.3
Suppose assumptions of Theorem 3.3 are satisfied and let \(k\le d\). Provided that \(\widetilde{\alpha }_k=\alpha _k\), there exists a positive constant \(C_k\) such that
Proof
We are going to use Theorem 2.3 of Goldie [15] which asserts that if
then
where
To prove (4.5), we are going to use Lemma 9.4 from [15], which derives the equality
From (3.1), we deduce that \(W_{k,0}\ge A_{kk,0}W_{k,-1}\) a.s. Hence, the absolute value may be omitted on both sides of (4.8). We consider two cases depending on the value of \(\alpha _k\).
Case 1 \(\alpha _k<1\).
Due to (3.2) and Lemma 4.2, we obtain
Case 2 \(\alpha _k\ge 1\).
For any \(x\ge y\ge 0\) and \(\alpha \ge 1\), the following inequality holds:
Since \(W_{k,0}\ge A_{kk,0}W_{k,-1}\ge 0\) a.s., we can estimate
Since the formula
holds for any \(x,y>0\) and each \(\alpha \), by putting \(\alpha =\alpha _k-1\) we obtain
Since \({{\mathbb {E}}}D_{k,0}^{\alpha _k}<\infty \) by Lemma 4.2, it remains to prove the finiteness of the second expectation. In view of (3.2),
where we use independence of \((A_{\cdot \cdot ,0},B_{\cdot ,0})\) and \(W_{\cdot ,-1}\). Since \({{\mathbb {E}}}W_{k,-1}^{\alpha _k-1}<\infty \) by Lemma 4.2, we focus on the remaining terms. Take \(p=\frac{\alpha _k+\varepsilon }{\alpha _k+\varepsilon -1}\) and \(q=\alpha _k+\varepsilon \) with \(0<\varepsilon <\min \{\widetilde{\alpha }_j:j\vartriangleright k\}-\alpha _k\). Then, since
the Hölder’s inequality together with Lemma 4.2 yields
Similarly, \({{\mathbb {E}}}[A_{kk,0}^{\alpha _k-1} B_{k,0}]<\infty \) and \({{\mathbb {E}}}[A_{kk,0}^{\alpha _k-1}A_{kj,0}]<\infty \) hold and hence \(I<\infty \).
By [15, Theorem 2.3], (4.6) holds and it remains to prove that \(C_k>0\). Since \(W_{k,0}\ge A_{kk,0}W_{k,-1}+B_{k,0}\) holds from (3.11), \(W_{k,0}^{\alpha _k}-(A_{kk,0}W_{k,-1})^{\alpha _k}\) is strictly positive on the set \(\{B_{k,0}>0\}\) which has positive probability in view of condition (T-2). Therefore, in (4.8) we obtain \(I>0\). Condition (T-7) implies that \(0<{{\mathbb {E}}}[A_{kk}^{\alpha _k}\log A_{kk}]<\infty \). Hence, we see from (4.7) that \(C_k>0\). \(\square \)
5 Case \(\widetilde{\alpha }_k<\alpha _k\)
The situation is now quite the opposite. The auto-regressive behavior of \(W_k\) does not play any role since k depends (in terms of Definition 3.2) on coordinates which admit dominant tails. More precisely, we prove that \(W_k\) has a regularly varying tail and its tail index is smaller than \(\alpha _k\). It is equal to the tail index \(\alpha _{j_0}\) of the unique component \(W_{j_0}\) such that \(\widetilde{\alpha }_k= \alpha _{j_0}(=\widetilde{\alpha }_{j_0})\). The latter is due to the formula
and it is proved inductively for all \(\ell \) such that \(k \trianglelefteq \ell \vartriangleleft j_0\). The decomposition (5.1) is the main goal in this section. We show that \(R_{\ell j_0}(s)\) is negligible as \(s\rightarrow \infty \), so that the tail of \(W_{\ell ,0}\) comes from \(\pi _{\ell j_0}(s)W_{j_0,-s}\). Moreover, we apply Breiman’s lemma to \(\pi _{\ell j_0}(s)W_{j_0,-s}\). Then, we also need to describe the limit behavior of \({{\mathbb {E}}}\pi _{\ell j_0}(s)^{\alpha _{j_0}}\) as \(s\rightarrow \infty \).
To reach (5.1), we first prove that \(W_{\ell ,0}\) may be represented as in (5.4). Then, we divide the series (5.4) into two parts: the finite sum \(Q_F(s)\) and the tail \(Q_T(s)\) of (5.6). Next, we decompose \(Q_F(s)\) into two parts: \(Q_W(s)\) containing W-terms and \(Q_B(s)\) containing B-terms; see (5.13) and (5.14). Finally, \(Q_W(s)\) is split into \(Q_W'(s)\) and \(Q_W''(s)\), where the former contains the components with dominating tails, while the latter gathers those with lower-order tails. This decomposition suffices to settle the induction basis in Lemma 5.10, since (5.28) is satisfied if we set \(R_{\ell j_0}(s)=Q''_W(s)+Q_B(s)+Q_T(s)\). The three parts \(Q_T(s),Q_B(s)\) and \(Q_W''(s)\) are estimated in Lemmas 5.3 and 5.4.
For the induction step in Lemma 5.9, we find \(s_1,s_2\in {{\mathbb {N}}}\) such that \(s=s_1+s_2\) and extract another term \(Q_W^*(s_1,s_2)\) from \(Q'_W(s_1)\), so that \(Q'_W(s_1)=\pi _{\ell j_0}(s)W_{j_0,-s}+Q_W^*(s_1,s_2)\). Then, we set
Notice that the two definitions of \(R_{\ell j_0}(s)\) above coincide, since \(Q_W^*(s,0)=0\) in the framework of Lemma 5.10. The term \(Q_W^*(s_1,s_2)\) is estimated in the induction step in Lemma 5.9, where both \(s_1\) and \(s_2\) are required to be sufficiently large. The limit behavior of \({{\mathbb {E}}}\pi _{\ell j_0}(s)^{\alpha _{j_0}}\) as \(s\rightarrow \infty \) is given in Lemma 5.7.
5.1 Decomposition of the Stationary Solution
We need to introduce some notation. For the products of the diagonal entries of the matrix \(\mathbf{A}\), we write
with a simplified form for \(t=0\)
We also define the subset \(H'_m(i,j)\subset H_m(i,j)\) as
namely \(H_m'(i,j)\) is the subset of \(H_m(i,j)\) such that the first two terms of its elements are not equal: \(i=h(0)\ne h(1)\). The latter naturally yields another definition:
which looks similar to \(\pi _{ij}(t,t-s)\) and can be interpreted as an entry of the product \(\mathbf{A}^0_t{\varvec{\Pi }}_{t-1,t-s}\), where \(\mathbf{A}^0\) stands for a matrix that has the same entries as \(\mathbf{A}\) outside the diagonal and zeros on the diagonal.
Now, we are going to formulate and prove the auxiliary results of this section. We start with representing \(W_{\ell ,t}\) as a series.
Lemma 5.1
Let \(\mathbf{W}_t=(W_{1,t},\ldots ,W_{d,t})\) be the stationary solution of (2.3). Then
where \(D_{i,t-n}\) is as in (3.3).
Remark 5.2
Representation (5.4) allows us to exploit the fast decay of \(\Pi ^{(i)}_{t,t-n+1}\) as \(n\rightarrow \infty \).
Proof
First, we show that for a fixed \(\ell \) the series on the right-hand side of (5.4) converges. For this, we evaluate the moment of some order \(\alpha \) from the range \(0<\alpha <\widetilde{\alpha }_i\). Without loss of generality, we let \(\alpha <1\) and \(t=0\). By Fubini’s theorem,
where \(L=\max \{{{\mathbb {E}}}W_j^\alpha :j\succ i\}\) and \(M={{\mathbb {E}}}B_i^\alpha \vee \max \{{{\mathbb {E}}}A_{ij}^\alpha :j\succ i\}\) are finite constants, and \(\rho ={{\mathbb {E}}}A_{ii}^\alpha < 1\). Hence, the series defined in (5.4) converges.
Denote \(Y_{i,t}:=\sum _{n=0}^\infty \Pi _{t,t-n+1}^{(i)}D_{i,t-n}\). We are going to show that the vector \(\mathbf{Y}_t=(Y_{1,t},\ldots ,Y_{d,t})\) is a stationary solution to (2.3). Since stationarity is obvious, without loss of generality set \(t=0\). For \(i=1,\ldots ,d\), we have
The d equations of the form (5.5) (one for each i) can be written together as a matrix equation:
which is the special case of (2.3). It remains to prove that this implies that \(Y_{i,0}=W_{i,0}\) a.s.
The series representation (2.4) allows to write \(D_{i,t}\) as a measurable function of \(\{(\mathbf{A}_s,\mathbf{B}_s):s<t\}\). Since the sequence \((\mathbf{A}_t,\mathbf{B}_t)_{t\in {{\mathbb {Z}}}}\) is i.i.d. and hence ergodic, it follows from Proposition 4.3 of [21] that \((D_{i,t})_{t\in {{\mathbb {Z}}}}\) is also ergodic. Brandt [6] proved that then \(Y_{i,t}\) defined in the paragraph above (5.5) is the only proper solution to (5.4) [6, Theorem 1]. Therefore, \(Y_{i,t}=W_{i,t}\) a.s. \(\square \)
Now, we divide the series (5.4) into the two parts: the finite sum \(Q_F(s)\) of the s first elements and the corresponding tail \(Q_T(s)\):
For notational simplicity, we abbreviate coordinate number \(\ell \) in \(Q_F(s)\) and \(Q_T(s)\). In the paper, the appropriate coordinate is always denoted by \(\ell \). The same applies for other parts \(Q_W(s)\), \(Q_B(s)\), \(Q'_W(s)\), \(Q''_W(s)\) and \(Q_W^*(s_1,s_2)\) of \(W_{\ell ,0}\) which will be defined in the sequel.
Both \(Q_F(s)\) and \(Q_T(s)\) have the same tail order for any s. However, if s is large enough, one can prove that \(Q_T(s)\) becomes negligible in the sense that the ratio of \(Q_T(s)\) and \(Q_F(s)\) tends to zero. The following lemma describes precisely that phenomenon. Recall that \(j_0\) is the unique index with the property \(\widetilde{\alpha }_\ell =\alpha _{j_0}\) and so \(\widetilde{\alpha }_j=\alpha _{j_0}\) for all j such that \(\ell \vartriangleleft j \trianglelefteq j_0 .\)
Lemma 5.3
Assume that for any j satisfying \(j_0\trianglerighteq j\vartriangleright \ell \),
Then, for every \(\varepsilon >0\) there exists \(s_0=s_0(\ell ,\varepsilon )\) such that
for any \(s\ge s_0\).
Proof
For \(\rho :={{\mathbb {E}}}A_{\ell \ell }^{\alpha _{j_0}}<1\), we choose a constant \(\gamma \in (0,1)\) such that \(\rho \gamma ^{-\alpha _{j_0}}=:\delta <1\). Notice that \( {{\mathbb {E}}}\big (\Pi ^{(\ell )}_{n})^{\alpha _{j_0}}\) decays to 0 exponentially fast, which is crucial in the following argument. We have
where we divided \(D_{\ell ,-n-s}=B_{\ell ,-n-s}+\sum _{j\succ \ell } A_{\ell j,-n-s}W_{j,-n-s-1}\) into two parts:
The first part \(D'_{\ell ,-n-s}\) contains those components of \(\mathbf{W}_{-n-s-1}\) which are dominant, while \(D''_{\ell ,-n-s}\) gathers the negligible parts: the components with lower-order tails and the B-term. For the second sum \(\mathrm {II}\), we have \({{\mathbb {E}}}(D''_{\ell ,-n-s})^{\alpha _{j_0}}<\infty \) by Lemma 4.2 and condition (T), so that Markov’s inequality yields
For the first sum \(\mathrm {I}\), we use conditioning in the following way:
where \(G_n=\Pi ^{(\ell )}_{n+s}A_{\ell j,-n-s}(1-\gamma )^{-1}\gamma ^{-n}\cdot 2d\). Notice that \(G_n\) and \(W_{j,-n-s-1}\) are independent and
where \(c_\ell = \max _{j}\{{{\mathbb {E}}}A_{\ell j}^{\alpha _{j_0}}: \ell \prec j \trianglelefteq j_0 \}\). Recall that by Assumption (5.7), there is a constant \(c_j\) such that for every \(x>0\)
Therefore, recalling \(\rho \gamma ^{-\alpha _{j_0}}=:\delta \), we further obtain
with \(c'=d\cdot c_\ell \cdot \max \{c_j:\ell \prec j \trianglelefteq j_0 \}\cdot (1-\gamma )^{-\alpha _{j_0}}(2d)^{\alpha _{j_0}}\). Now, in view of (5.10) and (5.11), since \(\rho <1\), there is \(s_0\) such that
for \(s>s_0\). \(\square \)
The dominating term \(Q_F(s)\) of (5.6) can be further decomposed.
Lemma 5.4
For any \(\ell \le d\) and \(s\in {{\mathbb {N}}}\), the sum \(Q_F(s)\) admits the decomposition
where
Moreover,
Remark 5.5
Each ingredient \(D_{\ell , -n}=\sum _{j:j\succ \ell }A_{\ell j,-n}W_{j,-n-1}+B_{\ell ,-n}\) of \(Q_F(s)\) in (5.6) contains several W-terms with index \(-n-1\) and a B-term with time index \(-n\) (see (3.3)). The idea is to change \(W_{\cdot ,-n-1}\) terms into \(W_{\cdot ,-s}\) by the iterative use of the recursion (2.3) component-wise. Meanwhile, some additional B-terms, with time indices between \(-n\) and \(-s+1\), are produced. When all W-terms have the same time index \(-s\), we gather all ingredients containing a W-term in \(Q_W(s)\), while the remaining part \(Q_B(s)\) consists of all ingredients containing a B-term. The quantity \(Q_W(s)\) is easily treatable, and \(Q_B(s)\) is proved to be negligible in the tail (comparing to \(Q_W(s)\), for which the lower bound is shown in the proof of Theorem 3.3).
Proof
Step 1. Decomposition of \(Q_F(s)\).
In view of (5.6) and (3.2), we have
We will analyze each ingredient \(\xi _n\) of the last sum. We divide it into two parts, \(\eta _n\) and \(\zeta _n\), starting from \(n=s-1\).
Next, for \(n=s-2\) applying component-wise SRE (3.1), we obtain
In this way, for \(n<s\) we define \(\eta _n\) which consists of terms including \((W_{j,-s})\) and \(\zeta _n\) which contains terms \((B_{j,-i})\). Observe that in most \(\eta \)’s and \(\zeta \)’s, a multiple summation appears, which is not convenient. To write them in simpler forms, we are going to use the notation (5.2). This yields
For each \(\eta _n\), we obtain an expression of a simple form similar to (5.17). To confirm this, let us return to the decomposition of \(\xi \) and see one more step of the iteration for \(\eta \). For \(n=s-2\), we have
so that similarly to the case \(n=s-2\), we obtain the expression
Similarly for any \(0\le n\le s-1\), we obtain inductively
The expression for \(\zeta _n\) is similar but slightly different. Let us write it for \(n=s-3\). From (5.18), we infer
and the general formula for \(\zeta _n\) is:
Therefore,
where
Finally to obtain (5.13), we use (5.3) and the identity:
for \(n<s\), \(h'\in H'_{s-n}(\ell ,j)\) and \(h\in H_{s}(\ell ,j)\) such that \(h(0)=h(1)=\ldots =h(n)=\ell \) and \(h(n+p)=h'(p)\) for \(p=0,\ldots ,s-1-n\). Each element \(h\in H_{s}(\ell ,j)\) is uniquely represented in this way.
Step 2 Estimation of \(Q_B(s)\).
Recall that \(\widetilde{\alpha }_\ell \le \alpha _j\) for \(j \vartriangleright \ell \). Hence, \({{\mathbb {E}}}A_{ij}^{\widetilde{\alpha }_\ell }<\infty \) and \({{\mathbb {E}}}B_i^{\widetilde{\alpha }_\ell }<\infty \) for any \(i,j \trianglerighteq \ell \). We have
Due to Minkowski’s inequality for \(\widetilde{\alpha }_\ell >1\) and Jensen’s inequality for \(\widetilde{\alpha }_\ell \le 1\), (5.23) is bounded by finite combinations of the quantities
for \(0\le n \le s-1\), \(j:j\trianglerighteq \ell \) and \(1\le m\le s-1-n\). Here, we recall that \(\pi '_{\ell j}(-n,-n-m+1)\) is a polynomial of \(A_{ij},\,i,j \trianglerighteq \ell \) (see (5.3)). Thus, \({{\mathbb {E}}}Q_B(s)^{\widetilde{\alpha }_\ell }<\infty \). \(\square \)
Example 5.6
In order to grasp the intuition of decomposition (5.12), we consider the case \(d=2\) and let
Then, applying the recursions \(W_{2,r}=A_{22,r}W_{2,r-1}+B_{2,r},\ -s \le r \le 1\) to the quantity \(Q_F(s)\) of (5.16), we obtain
where we recall that we use the convention \(\Pi ^{(\cdot )}_{-n-1,-n}=1\).
Let us focus on the first sum in the last expression, which is equal to \(Q_W(s)\). Each term of this sum contains a product of \(s+1\) factors of the form \(A_{11,\cdot },\,A_{12,\cdot }\) or \(A_{22,\cdot }\). Each product is completely characterized by the non-decreasing sequence \((h(i))_{i=0}^{s}\) of natural numbers with \(h(0)=1\) and \(h(s)=2\). If \(h(i)=1\) for \(i\le q\) and \(h(i)=2\) for \(i>q\), then there are q factors of the form \(A_{11,\cdot }\) in front of \(A_{h(q)h(q+1),-q}=A_{12,-q}\), and \(s-1-q\) factors of the form \(A_{22,\cdot }\) behind. All such sequences constitute \(H_{s}(1 ,2)\) of Definition 4.1. Thus, we can write
The second sum, which corresponds to \(Q_B(s)\), has another sum of the products in the nth term. All terms in these secondary sums of m are starting with \(A_{12,0}\) and then have a product of \(A_{22,\cdot }\) until we reach \(B_{\cdot ,\cdot }\). Each of the products is again characterized by a non-decreasing sequence \((h(i))_{i=0}^m\) such that \(h(0)=1,\,h(m)=2\) and \(h(1)\ne 1\), because there is just one such sequence for each m. Thus, we can use \(H_m'(1,2)\) of Definition 4.1 and write
5.2 The Dominant Term
Lemmas 5.4 and 5.3 imply that the tail of \(W_{\ell ,0}\) is determined by
in (5.13). In the subsequent lemmas (Lemmas 5.9 and 5.10), we will apply the recurrence to \(W_{j,-s},\ j\vartriangleleft j_0\) until they reach \(W_{j_0,t}\) for some \(t<s\). Those \(W_{j,-s}\) could survive as the dominant terms, and terms \(W_{j,-s},\ j\not \trianglelefteq j_0\) are proved to be negligible. In these steps, the behaviors of coefficients for all \(W_{j,-s}\) are inevitable.
The following property is crucial in subsequent steps, particularly in the main proof.
Lemma 5.7
Let \(\alpha _i>\widetilde{\alpha }_i=\alpha _{j_0}\) for some \(j_0\vartriangleright i\) and \(u_i (s):={{\mathbb {E}}}\pi _{i j_0}(s)^{\alpha _{j_0}}\). Then, the limit \(u_i=\lim _{s\rightarrow \infty } u_i (s)\) exists, and it is finite and strictly positive.
Proof
First notice that \(u_i(s)>0\) for s large enough since \(\pi _{ij}(s)\) is positive whenever \(j\trianglerighteq i\) and \(s\ge j-i\). We are going to prove that the sequence \(u_i(s)\) is non-decreasing w.r.t. s. Observe that
and therefore, by independence,
Since \(u_i(s)\) is non-decreasing in s, if \((u_i(s))_{s\in {{\mathbb {N}}}}\) is bounded uniformly in s, the limit exists. For the upper bound on \(u_i(s)\), notice that each product in \(\pi _{ij}(s)\) (see (4.1)) can be divided into two parts:
where m denotes the length of the product of \(A_{j_0j_0}\) terms. In the first part, all \(h(\cdot )\) but the last one \(h(s-m)=j_0\) are strictly smaller than \(j_0\). Since \({{\mathbb {E}}}A_{j_0j_0}^{\alpha _{j_0}}=1\), clearly
Now, let \(H''_{n}(j,j_0)\) denote the set of all sequences \(h\in H_{n}(j,j_0)\) which have only one \(j_0\) at the end. Suppose first that \(\alpha _{j_0}<1\). Then, by Jensen’s inequality we obtain
Here, we put
The term \(d(s-m)^d\) in (5.24) is an upper bound on the number of elements of \(H''_{s-m}(i,j)\). Another term \(M^{j_0-1}\) bounds the contribution of non-diagonal elements in the product, since any sequence \(h\in H_{s-m}(i,j_0)\) generates at most \(j_0-i\) such elements. The last term \(\rho ^{s-m-(j_0-i)}\) in (5.24) is an estimate of contribution of the diagonal elements since there are at least \(s-m-(j_0-i)\) of them.
We obtain (5.25) from (5.24) by substituting \(l=s-m\) and extending the finite sum to the infinite series. Since the series converges and its sum does not depend on s, the expectation \({{\mathbb {E}}}\pi _{ij_0}(s)^{\alpha _{j_0}}\) is bounded from the above uniformly in s.
Similarly, if \(\alpha _{j_0}\ge 1\), then by Minkowski’s inequality we obtain
which is again bounded from above, uniformly in s, by the same argument. \(\square \)
The number \(u_i\) depends only on i, since the index \(j_0=j_0(i)\) is uniquely determined for each i. We are going to use \(u_i\) as an upper bound for \({{\mathbb {E}}}\pi _{ij}(s)^{\alpha _{j_0}}\) with any j such that \(j_0 \trianglerighteq j\trianglerighteq i\). This is justified by the following lemma.
Lemma 5.8
For any j such that \(j_0 \vartriangleright j \trianglerighteq i \) there is \({\hat{s}}_j\) such that
Proof
The argument is similar to that in the proof of Lemma 4.2. Indeed, one finds \(\pi _{ij}(s)\) in \(S_{i,n}\) of (4.2) by setting \(s=m\). We briefly recall the argument for \(\alpha _{j_0}\le 1\). The case \(\alpha _{j_0}>1\) is similar. Without loss of generality, we set \(s\ge 2d-1\). The number of sequences in \(H_{s}(i,j)\) is less than \(ds^d\). Taking \(\rho \) and M as in (5.26), from (4.1) and Jensen’s inequality we infer
which implies (5.27). \(\square \)
We have already done all the preliminaries, and we are ready for the goal of this section, i.e., to establish the expression (3.7):
and prove the negligibility of the term \(R_{\ell j}(s)\) in the tail when \(s\rightarrow \infty \).
Lemma 5.9
Suppose that \(\widetilde{\alpha }_\ell <\alpha _\ell \) and \(j_0\vartriangleright \ell \) is the unique number with the property \(\widetilde{\alpha }_\ell =\alpha _{j_0}\). Assume that
whenever \(j_0\trianglerighteq j\vartriangleright \ell \). Then, for any \(\varepsilon >0\) there exists \(s_\ell =s_\ell (\varepsilon )\) such that if \(s>s_\ell \), \(W_{\ell ,0}\) has a representation
where \(R_{\ell j_0}(s)\) satisfies
The proof is given by induction, and first we consider the case when indeed \(\ell \prec j_0\), not only \(j_0\vartriangleright \ell \). More precisely, we prove the following lemma which serves as a basic tool in each inductive step.
Lemma 5.10
Suppose that \(\alpha _\ell >\widetilde{\alpha }_\ell =\alpha _{j_0}\) for some \(j_0\vartriangleright \ell \) and \(\widetilde{\alpha }_j > \alpha _{j_0}\) for every \(j\vartriangleright \ell ,\,j\ne j_0\). Then, for any \(\varepsilon >0\) there exists \(s_\ell =s_\ell (\varepsilon )\) such that for any \(s>s_\ell \), \(W_{\ell ,0}\) has a representation (5.28) which satisfies (5.29).
The condition in the lemma says that \(j_0\) is the unique coordinate which has the heaviest tail among \(j:j\vartriangleright \ell \), all the other coordinates that determine \(\ell \) do not depend on \(j_0\) and they have lighter tails.
Notice that as long as we only represent \(W_{\ell ,0}\) by \(\pi _{\ell j_0}(s)W_{j_0,-s}\) plus some r.v., we need not take a large s. Indeed, \(s=0\) is enough to obtain \((\pi _{\ell j_0}(0)=A_{\ell j_0,0})\) if \(\ell \prec j_0\). Thus, the number \(s_\ell \) is specific for (5.29).
Proof
In view of (5.13), we may write
where \({{\mathbb {E}}}\hat{Q}_W(s)^{\alpha _{j_0}}<\infty \) by the two facts;
(1) \({{\mathbb {E}}}\pi _{\ell j}(s)^{\alpha _{j_0}}<\infty \) for any j (see the explanation (4.1)),
(2) \({{\mathbb {E}}}W_j^{\alpha _{j_0}}<\infty \) for \(j\vartriangleright \ell ,\,j\ne j_0,\) which is due to Lemma 4.2.
Thus, by Lemma 5.4 we have
where \(\hat{Q}_W(s)\) and \(Q_B(s)\) have finite moment of order \(\alpha _{j_0}\) and \(Q_T(s)\) satisfies (5.8). Now putting
and setting \(s_\ell (\varepsilon )=s_0(\ell ,\varepsilon )\) as in Lemma 5.3, we obtain the result. \(\square \)
Proof of Lemma 5.9
Here, we may allow the existence of \(j:j_0 \vartriangleright j \vartriangleright \ell \), so that there exist sequences \((j_i)_{0\le i\le n}\) such that \(j_0\vartriangleright j_1\vartriangleright \cdots \vartriangleright j_n=\ell \). Since these sequences are strictly decreasing, their lengths are at most \(j_0-\ell +1\), i.e. possibly smaller than \(j_0-\ell +1\). Let \(n_0=n_0(\ell )\) denote the maximal length of sequence such that \((j_i)_{1\le i\le n_0}\) satisfies \(j_0\vartriangleright j_1\vartriangleright \cdots \vartriangleright j_{n_0}=\ell \). Then clearly \(j_0\succ j_1\succ \cdots \succ j_{n_0}=\ell \). In the same way, we define \(n_0(j)\) for any j in the range \(\ell \vartriangleleft j\trianglelefteq j_0\). We sometimes abbreviate to \(n_0\) when we mean \(n_0(\ell )\). We use induction with respect to this maximal number \(n_0\) to prove (5.28) and (5.29). First, we directly prove these two properties in the cases \(n_0=0,1\) which serve as the induction basis. For \(n_0>1\), the proof relies on the assumption that analogues of (5.28) and (5.29) hold for any j with \(n_0(j)<n_0(\ell )\). The proof is divided into four steps.
Step 1 Scheme of the induction and the basis.
Consider arbitrary coordinate \(j_0\) satisfying \(\alpha _{j_0}=\widetilde{\alpha }_{j_0}\). We prove that for any \(\ell \trianglelefteq j_0\) satisfying \(\widetilde{\alpha }_\ell =\alpha _{j_0}\) the properties (5.28) and (5.29) hold. The proof follows by induction with respect to \(n_0(\ell )\). Namely, we assume that for any \(j \trianglelefteq j_0\) satisfying \(\widetilde{\alpha }_j=\alpha _{j_0}\) and \(n_0(j)<n_0(\ell )\) and for any \(\varepsilon >0\) there is \(s_j=s_j(\varepsilon )\) such that for any \(s>s_j\) we have
where \(R_{jj_0}(s)\) satisfies
Then, we prove (5.28) and (5.29).
First, we settle the induction basis. Here, we consider the cases \(n_0=0\) and \(n_0=1\). The former case is equivalent to \(\ell =j_0\). In that setting, Theorem 3.3 was already proved in Lemma 4.3; nevertheless, (5.28) and (5.29) need to be shown separately. The iteration of (3.2) yields then that
where we recall that \(\Pi ^{(j_0)}_{0}=1\) and \(\Pi ^{(j_0)}_{1}=A_{j_0j_0,0}\). From the definition of \(j_0\) and Lemma 4.2, it follows that \({{\mathbb {E}}}D_{j_0}^{\alpha _{j_0}}<\infty \). Since \(R_{j_0j_0}(s)\) is constituted by a finite sum of ingredients which have the \(\alpha _{j_0}\)th moment, we conclude that \({{\mathbb {E}}}R_{j_0j_0}(s)^{\alpha _{j_0}}<\infty \) and (5.29) holds for any s. Thus, we may let \(s_\ell (\varepsilon )=1\) for any \(\varepsilon >0\). The case \(n_0=1\) is precisely the setting of Lemma 5.10, in which we have already proved (5.28) and (5.29).
If \(n_0>1\), then there is at least one coordinate j satisfying \(\ell \vartriangleleft j\trianglelefteq j_0\). For any such j, it follows that \(n_0(j)<n_0(\ell )\); hence, we are allowed to use the induction assumptions (5.32) and (5.33). In the next step, we prove that this range of j is essential, while for any other \(j\vartriangleright \ell \) the induction is not necessary.
Step 2 Decomposition of \(Q_W(s)\) and estimation of the negligible term.
The first term in (5.28) comes from the part \(Q_W(s)\) of \(W_{\ell ,0}\) in Lemma 5.4 and we further write
Recall that the relation ‘\(\not \trianglerighteq \)’ describes dependence between the components of the solution \(\mathbf{W}_t\) after a finite number of iterations of (1.1). Therefore, the range of summation \(j_0\not \trianglerighteq j\vartriangleright \ell \) in \(Q''_W(s)\) means that \(\ell \,(\ne j)\) depends on both j and \(j_0\) (by definition of \(j_0\)), but j does not depend on \(j_0\). The relation \(j\vartriangleright \ell \) implies that \(\widetilde{\alpha }_j\ge \widetilde{\alpha }_\ell \), while \(j_0\not \trianglerighteq j\) yields that \(\widetilde{\alpha }_j\ne \alpha _{j_0}\). Recalling that \(\widetilde{\alpha }_\ell =\alpha _{j_0}\), we can say that in \(Q'_W(s)\) we gather all coordinates \(j:j\vartriangleright \ell \) such that \(\widetilde{\alpha }_j=\alpha _{j_0}\) and \(Q''_W(s)\) consists of \(j:j\vartriangleright \ell \) such that \(\widetilde{\alpha }_j>\alpha _{j_0}\). Hence, by Lemma 4.2 each \(W_j\) appearing in \(Q''_W(s)\) has a tail of lower order than the tail of \(W_{j_0}\).
Notice that \(\hat{Q}_W(s)\) defined in (5.30) is the form that \(Q''_W(s)\) takes under the assumptions of Lemma 5.10. In this special case, we also have \(Q'_W(s)=\pi _{\ell j_0}(s)W_{j_0,-s}\). We are going to study this expression in the more general setting of Lemma 5.9.
Step 3 The induction step: decomposition of \(Q'_W(s)\).
To investigate \(Q_W'(s)\) in more detail, we will introduce a shifted version of \(R_{ij}(s)\). First, recall that the r.v.’s \(\pi _{ij}(t,s)\) and \(\pi _{ij}(t-s+1)\) have the same distribution and \(\pi _{ij}(t,s)\) can be understood as a result of applying t times a shift to all time indices in \(\pi _{ij}(t-s-1)\). In the same way, we define \(R_{ij}(t,s)\) as a result of applying t times a shift to all time indices in \(R_{ij}(t-s-1)\). In particular, we have \(R_{ij}(0,-s+1)=R_{ij}(s)\). By the shift invariance of the stationary solution, (5.32) and (5.33) are equivalent to their time-shifted versions:
and
respectively. Fix arbitrary numbers \(s_1,s_2\in {{\mathbb {N}}}\). Letting \(t=-s_1\) in (5.36), we obtain
where \(\pi _{\ell j_0}(s_1+s_2)\) consists of all combinations \(\pi _{\ell j}(s_1)\pi _{jj_0}(-s_1,-s_1-s_2+1)\) on \(j:j_0 \trianglerighteq j \trianglerighteq \ell \). This is clear when we recall that \(\pi _{\ell j}(s_1), \pi _{jj_0}(-s_1,-s_1-s_2+1)\) and \(\pi _{\ell j_0}(s_1+s_2)\) are appropriate entries of the matrices \({\varvec{\Pi }}_{s_1},{\varvec{\Pi }}_{-s_1,-s_1-s_2+1}\) and \({\varvec{\Pi }}_{s_1+s_2}={\varvec{\Pi }}_{s_1}{\varvec{\Pi }}_{-s_1,-s_1-s_2+1}\), respectively.
Now, a combination of (5.12), (5.35) and (5.38) yields
Our goal will be achieved by putting \(s=s_1+s_2\) in (5.39) and showing (5.29) for \(R_{\ell j_0}(s)\) with \(s=s_1+s_2\).
Step 4 The induction step: estimation of the negligible terms.
To obtain (5.29), we evaluate the four ingredients of \(R_{\ell j_0}(s_1+s_2)\) of (5.39), where the second hypothesis (5.33) of induction is used. Three of them, \(Q''_W(s_1),Q_B(s_1)\) and \(Q_T(s_1)\), are nonnegative; hence, it is sufficient to establish an upper bound for each of them. The fourth term, \(Q^*_W(s_1,s_2)\), may attain both positive and negative values; thus, we are going to establish an upper bound for its absolute value.
First, since the terms with \(j:j_0 \not \trianglerighteq j \vartriangleright \ell \) in \(Q_W''(s_1)\) of (5.35) satisfy the same condition as those of the sum \(\hat{Q}_W(s)\) of the previous lemma:
holds. Secondly,
holds in view of Lemma 5.4. Moreover, by Lemma 5.3 for any \(\varepsilon '>0\) there is \(s_0=s_0(\ell ,\varepsilon ')\) such that for \(s_1 > s_0\)
For the evaluation of \(Q^*_W(s_1,s_2)\), we will use (5.27) and therefore we have to assume that \(s_1>\max \{{\hat{s}}_j; j_0 \trianglerighteq j \vartriangleright \ell \}\) where \({\hat{s}}_j\) are defined in (5.27). Furthermore, we fix arbitrary \(\varepsilon ''>0\) and assume that \(s_2>\max \{s_j(\varepsilon ''):j_0\trianglerighteq j\vartriangleright \ell \}\) where \(s_j(\varepsilon '')\) are defined right before (5.32). Recall that \(\pi _{\ell j}(s_1),\,j:j_0\vartriangleright j \vartriangleright \ell \) is independent of \(R_{jj_0}(-s_1,-s_1-s_2+1)\) and has finite moment of order \(\alpha _{j_0}+\delta \) with some \(\delta >0\). We use (5.27), (5.37) and Lemma B.1 to obtain
The situation is different for \(j=j_0\). We cannot use Lemma B.1, because we do not know whether \({{\mathbb {E}}}\pi _{\ell j_0}(s_1)^{\alpha _{j_0}+\delta }<\infty \) for some \(\delta >0\). Indeed, it is possible that \({{\mathbb {E}}}A_{j_0j_0}^{\alpha _{j_0}+\delta }=\infty \) for all \(\delta >0\) and then clearly \(\pi _{\ell j_0}(s_1)\) also does not have any moment of order greater than \(\alpha _{j_0}\). However, \({{\mathbb {E}}}\pi _{\ell j_0}(s_1)^{\alpha _{j_0}}<\infty \) holds for any \(s_1\) and this is enough to obtain the desired bound. Since the term \(R_{j_0j_0}(s_2)\) is nonnegative (see (5.34)) and it was already proved to have finite moment of order \(\alpha _{j_0}\), we obtain for any \(s_1,s_2\)
and thus
Hence, setting \(N=\# \{j:j_0 \trianglerighteq j\vartriangleright \ell \}\), we obtain that
For the last inequality, we used Lemma B.1 and the fact that \(\pi _{\ell \ell }(s_1)=\Pi ^{(\ell )}_{0,-s_1+1}\). Since \(\rho ={{\mathbb {E}}}A_{\ell \ell }^{\alpha _{j_0}}<1\), there is \(s'=s'(\varepsilon '')\) such that \(\rho ^{s_1}u_\ell \cdot C_{j_0}<(N+1)^{\alpha _{j_0}}u_\ell \cdot \varepsilon ''\) for all \(s_1>s'\). Then, recalling that the sum in the last expression contains at most \(N-1\) nonzero terms, the final estimate is
Now, we are going to evaluate \(R_{\ell j_0}(s)\) of (5.29). The desired estimate can be obtained only if s is chosen properly.
For convenience, we briefly recall the conditions on \(s_1\) and \(s_2\) that were necessary to obtain the estimates (E.1–E.4). The inequalities (E.1) and (E.2) do not rely on any assumption on \(s_1\) or \(s_2\). The other relations are the following. Firstly, to obtain the inequality (E.3) we need to assume that \(s_1>s_0(\ell ,\varepsilon ')\). Secondly, the estimates \(s_1>\max \{\hat{s}_j:j_0\trianglerighteq j\vartriangleright \ell \}\) and \(s_2>\max \{s_j(\varepsilon ''):j_0\trianglerighteq j\vartriangleright \ell \}\) are used to prove (5.40). Passing from (5.40) to (E.4) relies on the condition \(s_1>s'\).
Now, let
where \(\cdot \vee \cdot =\max \{\cdot ,\cdot \}\). Then, there are \(s_1> s_0(\ell ,\varepsilon ')\vee \max \{\hat{s}_j:j_0 \trianglerighteq j \vartriangleright \ell \}\vee s'\) and \(s_2> \max \{s_j(\varepsilon ''):j_0\trianglerighteq j\vartriangleright \ell \}\) such that \(s=s_1+s_2\). Then,
The numbers \(s_1\) and \(s_2\) were chosen in such a way that (E.3) and (E.4) hold. The terms \(Q''_W(s_1)\) and \(Q_B(s_1)\) are negligible in the asymptotics. Therefore, it follows that
Since \(\varepsilon '\) and \(\varepsilon ''\) are arbitrary, we obtain (5.29). \(\square \)
6 Applications
Although there must be several applications, we focus on the multivariate GARCH(1, 1) processes, which is our main motivation. In particular, we consider the constant conditional correlations model by [4] and [18], which is the most fundamental multivariate GARCH process. Related results are the following. The tail of multivariate GARCH(p, q) has been investigated in [13] but with the setting of Goldie’s condition. A bivariate GARCH(1, 1) series with a triangular setting has been studied in [23] and [10]. Particularly in [10], detailed analysis was presented including exact tail behaviors of both price and volatility processes. Since the detail of application is an analogue of the bivariate GARCH(1, 1), we only see how the upper triangular SREs are constructed from multivariate GARCH processes.
Let \(\varvec{\alpha }_0\) be a d-dimensional vector with positive elements, and let \(\varvec{\alpha }\) and \(\varvec{\beta }\) be \(d\times d\) upper triangular matrices such that nonzero elements are strictly positive. For a vector \(\mathbf{x}=(x_1,\ldots ,x_d)\), write \(\mathbf{x}^\gamma =(x_1^\gamma ,\ldots ,x_d^\gamma )\) for \(\gamma >0\). Then, we say that d-dimensional series \(\mathbf{X}_t=(X_{1,t},\ldots ,X_{d,t})',\,t\in {{\mathbb {Z}}}\) has GARCH(1, 1) structure if it satisfies
where \(\mathbf{Z}_t=(Z_{1,t},\ldots ,Z_{d,t})'\) constitute an i.i.d. d-variable random vectors and the matrix \({\varvec{\Sigma }}_t\) is
Moreover, the system of volatility vector \((\sigma _{1,t},\ldots ,\sigma _{d,t})'\) is given by that of squared process \(\mathbf{W}_t=(\sigma _{1,t}^2, \ldots , \sigma _{d,t}^2)'\). Observe that \(\mathbf{X}_t={\varvec{\Sigma }}_t \mathbf{Z}_t=diag(\mathbf{Z}_t) \mathbf{W}_t^{1/2}\), so that \(\mathbf{X}_t^2 =diag(\mathbf{Z}_t^2) \mathbf{W}_t\). Then, \(\mathbf{W}_t\) is given by the following auto-regressive model.
Now, putting \(\mathbf{B}_t:=\varvec{\alpha }_0\) and \(\mathbf{A}_t:=(\varvec{\alpha }diag(\mathbf{Z}_t^2)+\varvec{\beta })\), we obtain the SRE: \(\mathbf{W}_t=\mathbf{A}_t \mathbf{W}_{t-1}+\mathbf{B}_t\) with \(\mathbf{A}_t\) the upper triangular with probability one. Each component of \(\mathbf{A}_t\) is written as
Thus, we could apply our main theorem to the squared volatility process \(\mathbf{W}_t\) and obtain the tail indices for \(\mathbf{W}_t\). From this, we could derive tail behavior of \(\mathbf{X}_t\) as done in [10].
Note that we have more applications in GARCH-type models. Indeed, we are considering applications in BEKK-ARCH models, of which tail behavior has been investigated with the diagonal setting (see [25]). At there, we should widen our results into the case where the corresponding SRE takes values on whole real line. The extension is possible if we assume certain restrictions and consider positive and negative extremes separately. Since the BEKK-ARCH model is another basic model in financial econometrics, the analysis with the triangular setting would provide more flexible tools for empirical analysis.
7 Conclusions and Further Comments
7.1 Constants
In the bivariate case, we can obtain the exact form of constants for regularly varying tails (see [10]). The natural question is whether we can obtain the form of constants even in the d-dimensional case. The answer is positive. We provide an example which illustrates the method of finding these constants when \(d=4\). Let
and suppose that \(\alpha _3<\alpha _4<\alpha _2<\alpha _1\).
For coordinate \(k=3,4\), we have
where \(W_k\) is independent of \(A_{kk}\) and \(B_k\). This is the Kesten–Goldie constant (see [15, Theorem 4.1]). Indeed, since \(W_4\) is a solution to the univariate SRE, we immediately obtain the constant. Since the tail index of \(W_3\) is equal to \(\widetilde{\alpha }_3=\alpha _3\), the constant follows by (4.6) in Lemma 4.3. For the second coordinate, we have an equation \(W_2{\mathop {=}\limits ^{d}}A_{22}W_2+A_{23}W_3+A_{24}W_4\). Since \(\widetilde{\alpha }_2=\alpha _3\) and \(\widetilde{\alpha }_4>\alpha _3\), the term \(A_{23}W_3\) dominates all others in the asymptotics. In view of (3.12), we obtain
where the quantity \(u_2\) is given in Lemma 5.7.
The situation seems more complicated for the first coordinate, because we have the condition \(\widetilde{\alpha }_1=\widetilde{\alpha }_2=\alpha _3\) on the SRE: \(W_1{\mathop {=}\limits ^{d}}A_{11}W_1+A_{12}W_2+A_{13}W_3+A_{14}W_4\). This means that the tail of \(W_1\) comes from \(W_2\) and \(W_3\) both of which have dominating tails, and we could not single out the dominant term. However, by Lemmas 5.7 and 5.9 again we obtain a simple formula:
We can write the general recursive formula for constants in any dimension:
Finally, we notice that these \(u_k\) have only closed form including infinite sums. The exact values of \(u_k\) seem to be impossible, and the only method to calculate them would be numerical approximations. The situation is similar to the Kesten–Goldie constant (see [24]).
7.2 Open Questions
In order to obtain the tail asymptotics of SRE such as (1.1), the Kesten’s theorem has been the key tool (see [9]). However, when the coefficients of SRE are upper triangular matrices as in our case, the assumptions of the theorem are not satisfied, so that we could not rely on the theorem. Fortunately in our setting, we can obtain the exact tail asymptotic of each coordinate, which is \({{\mathbb {P}}}(W_k>x) \sim C_k x^{-\widetilde{\alpha }_k}\). However, in general setting, one does not necessarily obtain such asymptotic even in the upper triangular case.
The example is given in [11], which we briefly see. Let \(\mathbf{A}\) be an upper triangular matrix with \(A_{11}=A_{22}\) having the index \(\alpha >0\). Then, depending on additional assumptions, it can be either \({{\mathbb {P}}}(W_1>x)\sim Cx^{-\alpha }(\log x)^{\alpha /2}\), or \({{\mathbb {P}}}(W_1>x)\sim C'x^{-\alpha }(\log x)^{\alpha }\) for some constant \(C,\,C'>0\).
There are many natural further questions to ask. What happens to the solution if some indices \(\alpha _i\) of different coordinates are equal? How we could find the tail asymptotics when the coefficient matrix is neither in the Kesten’s framework nor upper triangular? Moreover, if \(\mathbf{A}\) includes negative entries, could we derive the tail asymptotics? They are all open questions.
References
Alsmeyer, G., Mentemeier, S.: Tail behavior of stationary solutions of random difference equations: the case of regular matrices. J. Differ. Equ. Appl. 18, 1305–1332 (2012)
Belitskii, G.R., Lyubich, YuI: Matrix Norms and Their Applications. Birkhäuser, Basel (1988)
Bingham, N.H., Goldie, C.M., Teugels, J.L.: Regular Variation. Cambridge University Press, Cambridge (UK) (1987)
Bollerslev, T.: Modelling the coherence in short-run nominal exchange rates: a multivariate generalised ARCH model. Rev. Econ. Stat. 72, 498–505 (1990)
Bougerol, P., Picard, N.: Strict stationarity of generalized autoregressive processes. Ann. Probab. 20, 1714–1730 (1992)
Brandt, A.: The stochastic equation \(Y_{n+1}=A_n\, Y_n+B_n\) with stationary coefficients. Adv. Appl. Probab. 18, 211–220 (1986)
Buraczewski, D., Damek, E.: Regular behavior at infinity of stationary measures of stochastic recursion on NA groups. Colloq. Math. 118, 499–523 (2010)
Buraczewski, D., Damek, E., Guivarc’h, Y., Hulanicki, A., Urban, R.: Tail-homogeneity of stationary measures for some multidimensional stochastic recursions. Probab. Theory Relat. Fields 145, 385–420 (2009)
Buraczewski, D., Damek, E., Mikosch, T.: Stochastic Models with Power-Law Tails. The Equation \(X=AX+B\). Springer Int. Pub., Switzerland (2016)
Damek, E., Matsui, M., Świątkowski, W.: Componentwise different tail solutions for bivariate stochastic recurrence equations with application to GARCH(1,1) processes. Colloq. Math. 155, 227–254 (2019)
Damek, E., Zienkiewicz, J.: Affine stochastic equation with triangular matrices. J. Differ. Equ. Appl. 24, 520–542 (2018)
Embrechts, P., Klüppelberg, C., Mikosch, T.: Modelling Extremal Events for Insurance and Finance. Springer, Berlin (1997)
Fernández, B., Muriel, N.: Regular variation and related results for the multivariate GARCH(p, q) model with constant conditional correlations. J. Multivariate Anal. 100, 1538–1550 (2009)
Gerencsér, L., Michaletzky, G., Orlovits, Z.: Stability of block-triangular stationary random matrices. Syst. Control Lett. 57, 620–625 (2008)
Goldie, C.M.: Implicit renewal theory and tails of solutions of random equations. Ann. Appl. Probab. 1, 126–166 (1991)
Guivarc’h, Y., Le Page, É.: Spectral gap properties for linear random walks and Pareto’s asymptotics for affine stochastic recursions. Ann. Inst. H. Poincaré Probab. Statist. 52, 503–574 (2016)
Horváth, R., Boril, S̆.: GARCH models, tail indexes and error distributions: an empirical investigation. N. Am. J. Econ. Finance 37, 1–15 (2016)
Jeantheau, T.: Strong consistency of estimators for multivariate ARCH models. Econ. Theory 14, 70–86 (1998)
Kesten, H.: Random difference equations and renewal theory for products of random matrices. Acta Math. 131, 207–248 (1973)
Klüppelberg, C., Pergamenchtchikov, S.: The tail of the stationary distribution of a random coefficient AR(\(q\)) model. Ann. Appl. Probab. 14, 971–1005 (2004)
Krengel, U.: Ergodic Theorems. De Gruyter, Berlin (1985)
McNeil, A.J., Frey, R., Embrechts, P.: Quantitative Risk Management: Concepts, Techniques and Tools (Princeton Series in Finance). Princeton Univ. Pr, Princeton (2015)
Matsui, M., Mikosch, T.: The extremogram and the cross-extremogram for a bivariate GARCH(1,1) process. Adv. Appl. Probab. 48A, 217–233 (2016)
Mikosch, T., Samorodnitsky, G., Tafakori, L.: Fractional moments of solutions to stochastic recurrence equations. J. Appl. Probab. 50, 969–982 (2013)
Pedersen, R.S., Wintenberger, O.: On the tail behavior of a class of multivariate conditionally heteroskedastic processes. Extremes 21, 261–284 (2018)
Resnick, S.I.: Extreme Values, Regular Variation, and Point Processes. Springer, New York (1987)
Resnick, S.I.: Heavy-Tail Phenomena: Probabilistic and Statistical Modeling. Springer, New York (2007)
Straumann, D.: Estimation in conditionally heteroscedastic time series models. Lecture Notes in Statistics 181, Springer-Verlag, Berlin (2005)
Sun, P., Zhou, C.: Diagnosing the distribution of GARCH innovations. J. Empir. Finance 29, 287–303 (2014)
Acknowledgements
The authors are grateful to Ewa Damek and Dariusz Buraczewski for valuable discussion on the subject of the paper.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Muneya Matsui is partly supported by the JSPS Grant-in-Aid for Young Scientists B (16k16023) and for Scientific Research C (19K11868) and Nanzan University Pache Research Subsidy I-A-2 for the 2020 academic year. Witold Świątkowski is partly supported by the NCN Grant UMO-2014/15/B/ST1/00060.
Appendices
Appendix A. Negativity of Top Lyapunov Exponent
We provide the proof for negativity of \(\gamma _\mathbf{A}\) in Sect. 2. It can also be deduced from more general results of Straumann [28] or Grencsér, Michaletzky and Orlovitz [14] on the top Lyapunov exponent for block triangular matrices. However, in our case there is a direct elementary approach based on equivalence of norms and Gelfand’s formula.
Since all matrix norms are equivalent, we can use the norm
which is submultiplicative: \(\Vert \mathbf{A}\mathbf{B}\Vert _1 \le \Vert \mathbf{A}\Vert _1\,\Vert \mathbf{B}\Vert _1\). Then, since \(\mathbf{A}\) has nonnegative entries, we have \({{\mathbb {E}}}\Vert \mathbf{A}\Vert _1=\Vert {{\mathbb {E}}}\mathbf{A}\Vert _1\). Moreover, for any \(\varepsilon \in (0,1)\), \(\Vert \mathbf{A}\Vert _1^\varepsilon \le \Vert \mathbf{A}^\varepsilon \Vert _1\), where \(\mathbf{A}^\varepsilon \) denotes the matrix \(\mathbf{A}\) with each entry raised to the power of \(\varepsilon \). We are going to apply these to the form of top Lyapunov exponent \(\gamma _{\mathbf{A}}\). For any \(\varepsilon >0\), by Jensen’s inequality we have
Then, from properties of \(\Vert \cdot \Vert _1\) above, we infer that
where \({\varvec{\Pi }}_{n}^{(\varepsilon )}=\mathbf{A}_0^\varepsilon \cdots \mathbf{A}_{-n+1}^\varepsilon \). The last inequality follows from the superadditivity of the function \(f(x)=x^\varepsilon \). Since the matrices \(\mathbf{A}_i\) are i.i.d., we have \({{\mathbb {E}}}{\varvec{\Pi }}_{n}^{(\varepsilon )}=({{\mathbb {E}}}\mathbf{A}^\varepsilon )^n\). Here, we take the nth power in terms of matrix multiplication. Hence,
From Gelfand’s formula (e.g., [2, (1.3.3)]), for any matrix norm \(\Vert \cdot \Vert \) we can write
Taking the norm \(\Vert \cdot \Vert _1\), we obtain
Hence, it suffices to show that \(\rho ({{\mathbb {E}}}\mathbf{A}^{\varepsilon })<1\). If \(0<\varepsilon <\min \{\alpha _1,\ldots ,\alpha _d\}\), then from condition (T-4)
Since the spectral radius \(\rho ({{\mathbb {E}}}\mathbf{A}^{\varepsilon })\) is the maximal eigenvalue of \({{\mathbb {E}}}\mathbf{A}^{\varepsilon }\), stationarity is implied by (A.2).
Remark A.1
(i) By the equivalence of matrix norms, the argument above works for any norm. In order to observe this, take a certain norm \(\Vert \cdot \Vert \) and apply the inequality \(\Vert \mathbf{A}\Vert \le c \Vert \mathbf{A}\Vert _1\). Then, by (A.1) we obtain \({{\mathbb {E}}}\Vert {\varvec{\Pi }}_{n}\Vert ^\varepsilon \le c^\varepsilon \Vert {{\mathbb {E}}}{\varvec{\Pi }}_{n}^{(\varepsilon )}\Vert _1\). Since \(\lim _{n\rightarrow \infty }c^{\varepsilon /n}=1\) for any constant \(c>0\), the whole argument holds.
(ii) By the submultiplicativity of \(\Vert \cdot \Vert _1\), it is immediate to see that
However, we do not have any control on the norm \(\Vert {{\mathbb {E}}}\mathbf{A}^\varepsilon \Vert _1\); in particular, it can be greater than 1 for any \(\varepsilon \). It is essential in our situation that \(\rho ({{\mathbb {E}}}\mathbf{A}^\varepsilon ) \le \Vert {{\mathbb {E}}}\mathbf{A}^\varepsilon \Vert _1\) and involving Gelfand’s formula is necessary to obtain the desired bound.
Appendix B. Version of Breiman’s Lemma
We provide a slightly modified version of the classical Breiman’s lemma (e.g., [9, Lemma B.5.1]), since it is needed in the proof for (5.29) of Lemma 5.9. In the Breiman’s lemma, we usually assume regular variation for the dominant r.v.’s of the two, which we could not apply in our situation. Instead, we require only an upper estimate of the tail. The price of weakening assumptions is also a weaker result: On behalf of the exact asymptotics of a product, we obtain just an estimate from above. The generalization is rather standard, but we include it for completeness.
Lemma B.1
Assume that X and Y are independent r.v.’s and for some \(\alpha >0\), the following conditions hold:
Then,
Proof
The idea is the same as that in the original proof of Breiman’s lemma; see e.g., [9, Lemma B.5.1]. Let \(P_X\) denote the law of X. Then, for any fixed \(m>0\) we can write
By (B.1), there is \(x_0\) such that \(x^\alpha {{\mathbb {P}}}(Y>x)\le M\) uniformly in \(x\ge x_0\). For the first integral since \((x/z)\ge x_0\) for \(x\ge mx_0\) and \(z\in (0,m]\), by Fatou’s lemma
Since \(x/z >x_0\), the same argument as above is applicable to the second integral:
The assumption (B.2) allows us to use Markov’s inequality to estimate the last integral:
and therefore, the last integral is negligible as \(x\rightarrow \infty \) regardless of m. Now, in view of (B.3)-(B.6), letting m to infinity, we obtain the result. \(\square \)
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Matsui, M., Świątkowski, W. Tail indices for \(\mathbf{A}\mathbf{X}+\mathbf{B}\) Recursion with Triangular Matrices. J Theor Probab 34, 1831–1869 (2021). https://doi.org/10.1007/s10959-020-01019-8
Received:
Revised:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10959-020-01019-8
Keywords
- Stochastic recurrence equation
- Kesten’s theorem
- Regular variation
- Multivariate GARCH(1, 1) processes
- Triangular matrices