Abstract
In this paper, we establish a functional central limit theorem on high dimensional random fields in the context of model-based survey analysis. For strongly-mixing non-stationary random fields, we provide an upper bound for the fourth moment of the finite population total. This inequality is the generalization of a key tool for proving functional central limit theorems in Rio (Asymptotic theory of weakly dependent random processes, Springer, Berlin, 2017). Under the nested sampling strategy, we introduce assumptions on strongly-mixing coefficients and quantile functions to show that a functional stochastic process asymptotically approaches to a Gaussian process.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
In survey samplings, the model-based approach models finite populations as outputs from some random process, see for example Brewer (1963), Royall (1970), and the descriptions in Chambers and Clark (2012) and Valliant et al. (2000). A natural model for the case of a spatially based survey is that of a random field, not necessarily stationary, e.g. the study on non-stationary Gaussian random fields with the varying anisotropy property by Fuglstad et al. (2015) and the research on simulations of non-stationary random fields by Emery and Arroyo (2018). Generally, to have an estimation of asymptotically based confidence intervals on such random fields, we require a suitable central limit theorem, or if a function of means or empirical distribution functions are to be considered, a functional central limit theorem.
There is an extensive literature on functional central limit theorems for dependent random processes and fields. Most of this work concentrates on stationary processes, using a variety of conditions to control the weak dependence of the process. One strand of development does this using uniform mixing e.g. Billingsley (1968), Peligrad (1985), Deo (1975) and Chen (1991) for random fields. Another strand of work assumes positive or negative dependence Newman (1984). A group of papers works with the weaker concept of strongly-mixing, e.g. Doukhan et al. (1994), Doukhan (1995), Rio (2017), Merlevède and Peligrad (2000). Peligrad (1998) uses the maximal coefficient of correlations together with strongly-mixing to prove the invariance principle for stationary random processes. Some of these results have been extended to non-stationary processes, e.g. Matuła and Rychlik (1990) for positively dependent random processes, and Utev and Peligrad (2003) using strongly-mixing and conditions on the maximal coefficient of correlations. However, we have been unable to find any results for general non-stationary random fields with strongly-mixing.
In this paper, by using the approach in Doukhan et al. (1994), we prove a functional central limit theorem under conditions on the strongly-mixing coefficients and the quantile functions of the variables, where the finite population is modelled by a non-stationary strongly-mixing random field. Let \(X_{i\in {\mathbb {Z}}^d}\) stand for a random field, \(X_i\) be a random variable, \({\mathbb {Z}}^d\) be a high dimensional grid, and \({\mathbb {Z}}\) be the set of integers. We use the subset \(D_n\) of \({\mathbb {Z}}^d\) to index the finite population on the random field \(X_{i\in {\mathbb {Z}}^d}\), and \(|D_n|\), the cardinal number of the set \(D_n\), is used to describe the population size, i.e. \(X_{i\in D_n}\) stands for the finite population with the size \(|D_n|\). The subscript n is used to express the change of subsets. We write
and \(\sigma _n^2={\text {Var}}(S_n)\). Then, under certain assumptions we will show that
where \(t\in [0,1]\), G(t) is a Gaussian process on [0, 1], and the symbol \(\xrightarrow {\mathcal {D}}\) is interpreted as weak convergence in the Skorohod space D[0, 1].
After introducing some preliminaries in Sect. 2, we introduce the Skorohod space, our main result, and key tools used in Sect. 3. All proofs are put into Sect. 4.
2 Preliminaries
2.1 Mixing coefficients and a central limit theorem
Let \(X_{i\in {\mathbb {Z}}^d}\) be a zero-mean non-stationary random field satisfying the strongly-mixing condition: \(\alpha _{k,\ell }(m)\rightarrow 0\) as \(m\rightarrow \infty \), where \(\alpha _{k,\ell }(m)\) is the strongly-mixing coefficient of \(X_{i\in {\mathbb {Z}}^d}\),
In the above definition, \(\Lambda _1\) and \(\Lambda _2\) are subsets of \({\mathbb {Z}}^2\), \(|\Lambda _1|=k\), \(|\Lambda _2|=\ell \), \({\text {dist}}(\Lambda _1, \Lambda _2)\ge m\), and \(\sigma (\Lambda )\) denotes the sigma-field generated by the random variables \(X_i, i\in \Lambda \). We also use \(\alpha _{k,\ell }^X(m)\) to clarify this coefficient belongs to the random field \(X_{i\in {\mathbb {Z}}^d}\).
For strongly-mixing coefficients, we introduce two applications, which are used in our proofs. One of the applications is used to bound the covariance between variables on random fields, see the following lemma:
Lemma 2.1
[Theorem 3 of Doukhan (1994)] Let \(X_{i\in {\mathbb {Z}}^d}\) be a random field, \(\Lambda _1, \Lambda _2\subseteq {\mathbb {Z}}^d\), \(|\Lambda _1|=k\), \(|\Lambda _2|=\ell \). Then for each \(\delta >0\), we have
The other application of strongly-mixing coefficient is to provide the bound for the information of correlations as in the following lemma:
Lemma 2.2
Let \(X_{t\in {\mathbb {Z}}^d}\) be a random field,
Then we have
In our proofs, we also require the following central limit theorem for non-stationary random fields:
Theorem 2.1
[Theorem 3.3.1 in Guyon (1995)] Let \(X_{i\in {\mathbb {Z}}^d}\) be a zero mean random field, not necessarily stationary, and \(\{D_n\}\) an increasing sequence of domains. Put \(S_n=\sum _{i \in {\mathbb {Z}}^d}X_i\) and \(\sigma ^2_n = {\text {Var}} (S_n)\). If the random field satisfies
![figure a](http://media.springernature.com/lw685/springer-static/image/art%3A10.1007%2Fs11203-022-09273-9/MediaObjects/11203_2022_9273_Figa_HTML.png)
then
If we assume additionally that
![figure b](http://media.springernature.com/lw685/springer-static/image/art%3A10.1007%2Fs11203-022-09273-9/MediaObjects/11203_2022_9273_Figb_HTML.png)
then we have
2.2 Quantile functions
Following Rio (2017), the quantile function of a random variable X is denoted by \(Q_{X}\). It is the inverse of the non-increasing and left-continuous tail function of |X|. Specifically,
Some properties of the quantile function are established in Rio (2017), where they are used to connect the covariance function of the random field to the mixing coefficients. We expand the strongly-mixing coefficient for two random variables X and Y:
where \(\sigma (X)\) and \(\sigma (Y)\) are the sigma fields generated by X and Y respectively. Then we have the following lemma:
Lemma 2.3
[Theorem 1.1 (a) in Rio (2017)] Let X and Y be integrable real-valued random variables, and \(\alpha =\alpha (X, Y)\). Then
If X and Y are complex-valued, the constants 2 and 4 before the integrals are replaced by 8 and 16.
Lemma 2.4
[Lemma 2.1 (b) in Rio (2017)] Let \(X_1,\cdots , X_p\) be random variables. Then
Practically, Lemma 2.4 remains true if the upper limit in the integrals is replaced by any \(\alpha \in [0,1]\):
Lemma 2.5
Let \(X_1,\cdots , X_p\) be random variables. Then, for any \(\alpha \in [0,\frac{1}{4}]\),
3 Skorohod space and main results
3.1 The space D[0, 1]
Let D[0, 1] be the Skorohod space of càdlàg functions, i.e. \(x:[0,1]\mapsto {\mathbb {R}}\), where \(x\in D[0,1]\) is a right-continuous function, and the left-limits exist for any points of x. Let x and y be two elements in the space D[0, 1]. Define two metrics
and
where \(\lambda (t)\) is in the class of strictly increasing, continuous mappings from [0, 1] onto itself,
and \(\lambda (t)\) satisfies \(\Vert \lambda \Vert <\infty \). The Skorohod topology is defined by the metric \(d(\cdot ,\cdot )\), but the space D[0, 1] with metric d is not a complete metric space. However, the metric \(d_0\) generates the same topology as d, and \((D[0,1],d_0)\) is a complete metric space, so we can apply the standard theory of weak convergence in complete metric spaces. For details, see Billingsley (1968), where we can also see a convenient set of sufficient conditions for weak convergence in D[0, 1] is contained in the following theorem:
Theorem 3.1
[Theorem 15.6 of Billingsley (1968)] Let \(W_n(t)\) and W(t) be elements in space D[0, 1], and F(t) a continuous non-decreasing function on [0, 1]. Suppose that
-
(1)
\((W_n(t_1), W_n(t_2), \cdots , W_n(t_k)) \xrightarrow {\mathcal {D}} (W(t_1), W(t_2), \cdots , W(t_k))\) whenever \(t_1, t_2, \cdots , t_k\in [0,1]\);
-
(2)
\({\mathbb {P}}(W(1)\ne W(1-))=0\);
-
(3)
$$\begin{aligned} {\mathbb {P}}(|W_n(t)-W_n(t_1)|\ge \lambda , |W_n(t)-W_n(t_2)|\ge \lambda ) \le \frac{1}{\lambda ^{2\gamma }}[F(t_2)-F(t_1)]^{2\alpha }, \end{aligned}$$(3)
for \(t_1\le t\le t_2\) and \(n\ge 1\), where \(\gamma \ge 0\), \(\alpha >\frac{1}{2}\). Then we have
$$\begin{aligned} W_n(t)\xrightarrow {\mathcal {D}}W(t). \end{aligned}$$
3.2 A moment inequality
Along the line in Rio (2017), to generalize Theorem 2.1 therein, for the strongly-mixing coefficient, \(\alpha _{k, \ell }(m)\), we define the inverse function of \(\alpha _{k, \ell }(m)\) by
We note that
Therefore, for any constant \(K>0\), we have
By applying the approach in Rio (2017), we prove the following inequality on non-stationary high-dimensional random fields. To simplify the expression, we combine \(\alpha _{2,2}^{-1}(u)\), \(\alpha _{1,3}^{-1}(u)\) and \(\alpha _{3,1}^{-1}(u)\) by defining
Theorem 3.2
Let \(X_{i\in {\mathbb {Z}}^d}\) be a zero-mean random field with strongly-mixing coefficients \(\alpha _{k,\ell }(m)\) and let \(\alpha _4^{-1}(u)\) be defined as in (6), and \(D\subseteq {\mathbb {Z}}^d\). Then there exists a positive constant K, which depends only on d, such that
It is obvious that if (H1) is satisfied and
then there exists a constant K such that for each \(D\subseteq {\mathbb {Z}}^d\),
3.3 Nested structure and a functional CLT
To state our functional central limit theorem, we need an assumption on the increasing sequence of domains \(\{D_n\}\), used in the definition of \(S_n\). Specifically, we require that there exist constants \(K_0>0\) and \(K_1>0\), such that, for any \(n_1\le n_2\), the following inequality is satisfied:
where \(D_n^c\) is the complement of \(D_n\). In this case, we say that \(\{D_n\}\) has a nested spatial structure. For example, in Fig. 1, we use black dots to stand for a random field, the closed solid lines are boundaries of domains, and the dots inside a boundary are used to describe the domain introduced above. In this case, \(K_0=1\), \(K_1=2\), and \({\text {dist}}(D_{n_1},D_{n_2}^c) = 2(n_2-n_1)\), i.e. as it is shown in Fig. 1, \(D_1\subset D_2\) and \({\text {dist}}(D_{1},D_{2}^c) = 2\).
Regarding to the strongly-mixing coefficient in a structured random field, the coefficient between \(D_{n_1}\) and \(D_{n_2}^c\) can be specified by
and satisfies
due to, for every domain \(D_*\subseteq {\mathbb {Z}}^d\), \(D_{*}\) is just an element of \(\{D\subseteq {\mathbb {Z}}^d:|D|=|D_{*}|\}\). If we introduce a random field as shown in Fig. 1, and define \(D_1\) as a 3-by-3 grid, \(D_2\) a 6-by-6 grid, \(\cdots \), \(D_n\) a \((2+n^2)\)-by-\((2+n^2)\) grid, \(\cdots \), then we have \(|D_{n}|=(2+n^2)^2\), and \({\text {dist}}(D_{n-1},D_{n}^c)\ge n\), for all \(n\ge 1\). Furthermore, if we assume \(|D_n^c|=\infty \), the definition of strongly-mixing coefficient implies
In this case, \(|D_{n-1}|\rightarrow \infty \) together with \({\text {dist}}(D_{n-1},D_{n}^c)\rightarrow \infty \), as \(n\rightarrow \infty \).
Practically, a random field with nested structure can be applied in many scenarios, such as the soil chemical concentration in earth science Christakos (1992), the colorectal cancer in health science Jha et al. (2021), the mobile robotic sensors of artificial intelligent products Nguyen et al. (2021), and so on. However, the nested structure assumes all objects in the random field can be observed at any time, i.e. if an object is observable in a smaller domain, it has to be observable for any wider domains. This assumption could be hard to suit some short lifespan objects such as mayflies researched in Sroka et al. (2021).
Theorem 3.3
Let the zero-mean random field \(X_{i\in {\mathbb {Z}}^d}\) satisfy (H1), (H2) and (H3) in Theorem 2.1. We also assume
-
(1)
The sequence of domains \(D_n\subset {\mathbb {Z}}^d\) has a nested spatial structure, and for any \(n_1<n_2\), we have \( \alpha _{D_{n_1}, D_{n_2}^c}({\text {dist}}(D_{n_1},D_{n_2}^c))\rightarrow 0 \) as \({\text {dist}}(D_{n_1},D_{n_2}^c)\rightarrow \infty \);
-
(2)
There exists a strictly increasing Lipschitz function F(t), \(t\in [0,1]\), such that
$$\begin{aligned} \lim _{n\rightarrow \infty } \frac{|D_{[nt]}|}{|D_{n}|} = F(t), \hbox { for each fixed}\ t\in [0,1] ; \end{aligned}$$ -
(3)
The limit \(\displaystyle \lim _{n\rightarrow \infty }\frac{\sigma _n^2}{|D_n|}\) exists, and is positive;
-
(4)
\(\displaystyle \sup _{i\in {\mathbb {Z}}^d}\int _0^{\frac{1}{4}}\bigg ([\alpha ^{X}_{4}]^{-1} (u) \bigg )^{3d} Q_{X_i}^4(u)du<\infty , \) where \([\alpha ^{X}_{4}]^{-1}(u)\) is defined as in (6).
If we set \(G_n(t)=S_{[nt]}/\sigma _n\), then, there exists a one-dimensional Gaussian process, G(t), with the covariance function \({\text {C}}(s, t) = \min \{F(s), F(t)\}\), such that \(G_n(t)\xrightarrow {\mathcal {D}}G(t)\) in Space D[0, 1].
In Theorem 3.3, Assumption (1) uses the relationship between \({\text {dist}}(D_{n_1},D_{n_2}^c)\) and \(D_{n_1}\). Although it is possible that \(|D_{n_1}|\rightarrow \infty \) as \({\text {dist}}(D_{n_1},D_{n_2}^c)\rightarrow \infty \), it is different to the case \(\alpha _{\infty , \infty }(m)\) considered in Bradley (1993), where the distance m is not related to the size of domains. The nested domains endow a connection between \({\text {dist}}(D_{n-1},D_{n}^c)\) and \(|D_{n-1}|\). It will benefit the checking of the dependency asymptotically. Therefore, in Theorem 3.3, we introduce Assumption (1) rather than assume \(\alpha _{\infty , \infty }(m)\rightarrow 0\) as \(m\rightarrow \infty \).
Assumption (2) of Theorem 3.3 reflects that when the cardinal number can be expressed in the way \(|D_n|={\text {O}}(n^d)\), the limitation function is just \(F(t)=t^d\). Therefore, in the proof of Theorem 3.2, it will directly contribute to verifying the third assumption of Theorem 3.1.
For Assumption (3) of Theorem 3.3, it is a stronger version of condition (H4) of Theorem 2.1. A sufficient condition in the cases \(d=1\) is that the random field be stationary, and the mixing coefficients decay sufficiently fast, in the sense that if \(||X_i||_{2+\delta }<\infty \) then
For more details we refer to Lemma 1.1 in Rio (2017). The result of Theorem 18.2.3 in Ibragimov (1971) is also related to the sufficient condition of Assumption (3). A similar result for strongly-mixing coefficients is Lemma 2.2 in Merlevède and Peligrad (2000). However, when \(d\ge 2\), we have been unable to find such sufficient conditions for Assumption (3) so one is provided in Sect. 4, where the domains are assumed in the form \(D_n = ([0,n]\times [0,n])\cap {\mathbb {Z}}^2\), and the result therein can be extended to a higher dimension random field.
Inspired by DMR condition used in Corollary 1.2 and Theorem 4.3 in Rio (2017), we introduce Assumption (4) for Theorem 3.3. Because the random field in this paper has no stationary property, we cannot get a unique integral as that in DMR condition. Therefore, we use the supremum to unify the expression. Additionally, since the dimension of random fields in this paper is higher, we have to introduce higher orders for the inverse function and the quantile function in Assumption (4). Reasoning for why a higher order is used can be found in the proof of Theorem 3.2.
Comparing with (H1) and (H3), Assumption (4) uses the information of distances and quantiles of random variables, whereas (H1) and (H3) use the distances and the strongly-mixing coefficients. Assumption (4) looks more directly putting restriction on the moments of random variables. However, although the fourth moment is bounded, when \(\delta =2\) in (H3), Assumption (4) may not be satisfied, due to the distance, \([\alpha ^{X}_{4}]^{-1}(u)\), could be unbounded. On the other hand, the bounded product in Assumption (4) does not imply any statements in (H1) or (H3). Therefore, in this present theorem, we use those relative weaker assumptions rather than integrate them into stronger conditions.
4 Proofs
4.1 Proof of Lemma 2.2
We use the induction method. For \(r=1\), the result (2) holds. We suppose it holds for \(r=p\), i.e. let
then we suppose
Following this, we need to prove it is still true for \(r=p+1\).
Now let
then, by using the monotonicity properties of strongly-mixing coefficients, we have
which implies
and
Therefore, we have
This completes the proof.
4.2 Proof of Lemma 2.5
For any \(\alpha \), \(\alpha \in [0,\frac{1}{4}]\), we define a random variable,
Then the tail function of Z is
Now, for any random variables, \(X_1\) and \(X_2\), we have
4.3 Proof of Theorem 3.2
Let \(\{S_1, S_2\}\) be a partition of distinct \(\{i,j,k,\ell \}\), \(i,j,k,\ell \in {\mathbb {Z}}^d\). Then \(S_1\) and \(S_2\) are two non-empty sets. We define
where \(\mathcal {P}_{ijk\ell }\) stands for the set of all partitions of \(\{i,j,k,\ell \}\), i.e.
Lemma 4.1
Let \(M_{ijk\ell }\) be defined as in (8), but with \(\{i,j,k,\ell \}\) not necessary distinct. Then \(j,k,\ell \) must be within \(3M_{ijk\ell }\) of i, i.e.
Proof
We suppose the contrary of (9). Then, it fall into two cases (a) and (b).
(a) Firstly, we suppose \(\{i,j,k,\ell \}\) are distinct, then there are three cases to consider.
(a.1) All of \(j,k,\ell \) are more than \(3M_{ijk\ell }\) from i. But this is impossible as the split \(S_1=\{i\}\) and \(S_2=\{j,k,\ell \}\) would have a separation greater than \(M_{ijk\ell }\).
(a.2) Two of \(j,k,\ell \) are more than \(3M_{ijk\ell }\) from i, say \(k,\ell \). Then the separation between i and \(k,\ell \) is more than \(M_{ijk\ell }\). Therefore we must have \({\text {dist}}(i,j)\le M_{ijk\ell }\), or else the separation \(S_1=\{i\}\) and \(S_2=\{j,k,\ell \}\) would be more than \(M_{ijk\ell }\). Hence
i.e. \(d(j,k)\ge 2M_{ijk\ell }\). By the same argument, we have \(d(j,\ell )\ge 2M_{ijk\ell }\). Thus, the seperation between \(S_1=\{i,j\}\) and \(S_2=\{k,\ell \}\) is more than \(M_{ijk\ell }\). This contradicts the definition of \(M_{ijk\ell }\).
(a.3) One of \(j,k,\ell \) is more than \(3M_{ijk\ell }\) from i, say \(\ell \). For example, if \(d(i,\ell )>3M_{ijk\ell }\), \(d(i,j)<M_{ijk\ell }\) and \(d(i,k)<M_{ijk\ell }\), then by the same argument, we have \(d(i,\ell )>2M_{ijk\ell }\) and \(d(k,\ell )>2M_{ijk\ell }\). This contradicts to the definition of \(M_{ijk\ell }\).
Thus all these alternatives are impossible, so that \(j,k,\ell \) are within \(3M_{ijk\ell }\) of i.
(b) Secondly, we suppose \(\{i,j,k,\ell \}\) are not distinct. We might also have the following cases:
(b.1) If only three of \(i,j,k,\ell \) are distinct, say \(i=j\), then in (8), we can interpret \(S_1\) and \(S_2\) as the partition of \(\{i,k,\ell \}\). In this case, with a similar argument, we must have the distance between i and the other two distinct points is less than or equal to \(2M_{ijk\ell }\).
(b.2) If they are equal in pairs, then we must have the distance between the two distinct points equal to \(M_{ijk\ell }\).
(b.3) If they are all equal, then it becomes trivial.
So, in every case, we must have the conclusion of Lemma 4.1. \(\square \)
Returning to the proof of Theorem 3.2, we write
Consider partitioning \(\{i,j,k,\ell \}\) into two non-empty subsets, \(S_1\) and \(S_2\). Let \(M_{ijk\ell }\) be defined as in (8).
Case 1: Suppose the maximum distance \(M_{ijk\ell }\) occurs at the partition \(\{S_1, S_2\}\), where \(S_1\) contains two elements, and \(S_2\) contains the other two. Without loss of generality, we set \(S_1=\{i,j\}\) and \(S_2=\{k,\ell \}\). Then, by using the inequality \(abcd\le \frac{1}{4}(a^4+b^4+c^4+d^4)\), where \(a,b,c,d\in {\mathbb {R}}\), Lemmas 2.3 and 2.5 imply
Case 2: If \(M_{ijk\ell }\) occurs at the partition, where \(S_1\) contains one element, and \(S_2\) contains the other three. Without losing of generality, we set \(S_1=\{i\}\) and \(S_2=\{j,k,\ell \}\). Then, similarly, we have
The above two estimations of \({\mathbb {E}}(X_iX_jX_kX_\ell )\) also work for the cases where \(i,j,k,\ell \) are not distinct.
We note that
Lemma 4.1 and (8) imply \(|\{(j,k,\ell ):M_{ijk\ell }\le m\}|\le K(m+1)^{3d}\), where K is a constant. Then by using (4) and (5),
Similarly, we have
Therefore the above discussion implies the upper bound of this fourth moment is not greater than the sum of these two possible cases. It leads to
4.4 Proof of Theorem 3.3
Let’s introduce a lemma as the preparation of the proof.
Lemma 4.2
Let \(X_n\) and \(Y_n\) be two random sequences, with \(X_n-Y_n\) and \(Y_n\) being asymptotically independent. Also assume that \(X_n\xrightarrow {d}X\), \(Y_n\xrightarrow {d}Y\), \(X-Y\) is independent of Y, and Y has a non-zero characteristic function. Then \(X_n- Y_n\xrightarrow {d}X- Y\).
Proof
Let \(\varphi _Z\) be the characteristic function of random variable Z. Now the asymptotic independence implies \(\varphi _{X_n-Y_n}(t)\varphi _{Y_n}(t)-\varphi _{X_n}(t)\rightarrow 0\). Since \(\varphi _{X_n-Y_n}(t)\varphi _{Y_n}(t)-\varphi _{X_n}(t)\rightarrow 0\), due to \(X_n\xrightarrow {d}X\), \(Y_n\xrightarrow {d}Y\), and the bounded \(\varphi _{X_n-Y_n}(t)\), the independence between \(X-Y\) and Y implies \(\varphi _{X_n-Y_n}(t)\varphi _{Y}(t)-\varphi _{X-Y}(t)\varphi _{Y}(t)\rightarrow 0\). Therefore, the fact that \(\varphi _Y(t)\ne 0\) completes the proof.
which converges to zero. Since the characteristic function of Y is non-zero, this proves the lemma. \(\square \)
Continuing with the proof of Theorem 3.3, we need to show the existence of G(t), and check the three conditions in Theorem 3.1. See the following two parts.
Part one: showing the existence of G(t)
To show G(t) has independent increments, we split the interval [0, 1] in the following way:
Then we have \(\delta >0\), \( [ns_i]-[nt_{i-1}]+1\ge [n\delta ]\) and
for all n and \(i=2,\cdots ,r\). We set
Then Assumption (1) implies that, for all \(i=1,\cdots , r-1\),
Because
if we set
where \(H_i\in {\mathscr {B}}^d\), d-dimensional Borel sets, we have \(E_i\in {\mathscr {M}}_i\). Lemma 2.2 guarantees the following estimation,
Since the strongly-mixing condition implies \(\alpha _{\ell ,\ell }^X(m)\rightarrow 0\) as \(m\rightarrow \infty \), the right-hand side of the above inequality goes zero as \(n\rightarrow \infty \). Hence \(G_n(t)\) has asymptotically independent increments. By using Theorem 4.5.3 in Chung (2001), the fact of \({\mathbb {E}}(|G_n(t)|)\le M<\infty \), where M is a constant, and the fact of that, for all \(\epsilon >0\), there exists \(\lambda (\epsilon , M)\) such that for any \(E\in \sigma (G_n(t))\), if \({\mathbb {P}}(E)<\lambda (\epsilon , M)\), then \(\int _E|G_n(t)|d{\mathbb {P}}<\epsilon \), imply \(G_n(t)\) is uniformly integrable. Then, referring to the arguments in the proof of Theorem 19.2 in Billingsley (1968) and that in the proof of Theorem 4.3 in Rio (2017), Theorem 8.2 in Billingsley (1968) implies \(G_n(t)\) is tight. Therefore G(t) has independent increments. Its variance is F(t), which is proved by (13) in Part two.
Furhtermore, the covariance matricx of G(t) is positive definite, due to the convariance function of G(t) is that of the Gaussian process \((B_{F(t)})_{0\le t\le 1}\), where \((B_{t})_{0\le t\le 1}\) is a standard Brownian motion. Therefore, the covariance function of G(t) is \({\text {C}}(s,t)=\min \{F(s), F(t)\}\).
Part two: checking conditions in Theorem 3.1
We first check the three conditions in Theorem 3.1. For a finite positive integer k, reals \(0=t_0<t_1<\cdots<t_{k-1}<t_k=1\), and for sufficiently large n, we set
To prove convergence of the finite dimensional distributions, it is enough to show that (a) \(G_n(t) \rightarrow N(0, F(t))\) and that (b)
For (a), by Theorem 2.1 and assumptions (2) and (3), we have
where
For (b), as in the proof of Theorem 4.3 in Rio (2017), we consider characteristic functions. Let \(s=(s_1,s_2,\cdots ,s_k)\in {\mathbb {R}}^k\), and
Since G(t) has independent increments, \((G(t_i)-G(t_{i-1}))_{1\le i\le k}\) are independent. Therefore, to prove (12), it is sufficient to prove
We write
For term (IV), by recalling that \(\epsilon = n^{-1/2}\), and using the continuity property of characteristic functions, we have
For term (I), using the inequality \(|e^{ix}-e^{iy}|\le |x-y|\). we have
Thus
Because (H4) implies \(\sigma _n^2\ge K|D_n|\), where K is a constant, we get
so that by (H3) and assumption 2, there is a constant C such that
Therefore, for all s, we have \(|\varphi _n(s)-\varphi _{n,\epsilon }(s)|\rightarrow 0\) as \(n\rightarrow \infty \).
For (II), set
Then
The random variable \(Y_j\) is \(\sigma (D_{[nt_j-\epsilon ]}-D_{[nt_{j-1}}])\)-measurable, \(Y_{j+1}\cdots Y_k\) is \(\sigma (D_{[nt_j]}^c)\)-measurable, and
We set \(K_3 [ n\epsilon ]^{K_0}=d_n\). Then we have \(d_n\rightarrow \infty \) as \(n\rightarrow \infty \), i.e. \(K_3 [ n\epsilon ]^{K_0}\rightarrow \infty \) as \(n\rightarrow \infty \). Furthermore, since \(|Y_j| = |Y_1\cdots Y_k|=1\) and \(Q_{|1|}(u) = 1\), Lemma 2.3 implies that (14) is less than
where \(D_{1,n}=D_{[nt_j]}^c\) and \(D_{2,n}=D_{[n(t_j-\epsilon )]}-D_{[nt_{j-1}]}\). Thus Assumption 1) ensures \((II)\rightarrow 0\).
For term (III), it is sufficient to prove that
for \(t_2>t_1\). Since Theorem 2.1 implies \(G_n(t)\rightarrow G(t)\) for any fixed \(t\in [0,1]\), by Lemma 4.2 we only need to show that \(G_n(t_2)-G_n(t_1)\) and \(G_n(t_1)\) are asymptotically independent, i.e. that
where \(\varphi _1(s)\), \(\varphi _2(s)\) and \(\varphi _3(s)\) are characteristic functions of \(G_n(t_2)\), \(G_n(t_2)-G_n(t_1)\) and \(G_n(t_1)\) respectively. Let \(\varphi _*(s)\) be the characteristic function of \(G_n(t_2)-G_n(t_1)+G_n(t_1-\epsilon )\), and \(\varphi _{\epsilon }(s)\) be the characteristic function of \(G_n(t_1-\epsilon )\). Then we have
For terms (i) and (iii), we use the same arguments as in the proof of (I). We have
and
For (ii), arguing as above yields
Hence, the first condition of Theorem 3.1 is satisfied.
To check the second condition of Theorem 3.1, we will show that G(t) has continuous sample paths with probability one. For \(t>s\), we have a constant K such that
where the last inequality follows from the fact that \(\log x <\sqrt{x} \) for \(x>1\). Hence by using Theorem 3.4.1 in Adler (2010), G(t) has continuous sample paths almost surely.
Finally, we check the third condition of Theorem 3.1. We set \(\gamma =2\), \(\alpha =1\), then for \(t_1<t<t_2\), Theorem 3.2 and Assumption 4) imply
Therefore the third condition of Theorem 3.1 is satisfied. This completes the proof.
4.5 A sufficient condition for Assumption (3) in Theorem 3.3
Let \(D_n=[0, n]^2\cap {\mathbb {Z}}^2\). We put \(v_n={\text {Var}}(S_n)-{\text {Var}}(S_{n-1})\). By using Kronecker’s Lemma, Lemma 2 on page 390 in Shiryaev (1996), we only need to show \(\frac{v_n}{n}\rightarrow v\) under the assumption that
We write
and split \(D_n-D_{n-1}\) up into 3 parts, i.e.
and
Then
and
Since
we consider the first term, the second is similar. For the first term, the term
so contribute to \(\frac{v_n}{n}\) is
But assumption (15) implies the series \(\sum _{j=0}^{n-1}\sum _{\ell =0}^{n-1}{\text {Cov}}(X_{lj}, X_{0,n})\) converges absolutely, since, by using Lemma 2.1, there exists a constant K such that
Therefore, (18) converges to a limit.
Let’s consider a term like the following:
The contribution to \(\frac{v_n}{n}\) is
The same arguments imply the absolute convergence of the above last two terms, and apply to those terms in (16) and (17), so that \(\frac{v_n}{n}\) converges to a value v. Put \(w_n=\frac{v_n}{n}\). Now
which also converges to v.
References
Adler RJ (2010) The geometry of random fields. SIAM
Billingsley P (1968) Convergence of probability measures. Wiley, New York
Bradley RC (1993) Some examples of mixing random fields. The Rocky Mt J Math 23(2):495–519
Brewer K (1963) Ratio estimation and finite populations: some results deducible from the assumption of an underlying stochastic process. Aust J Stat 5(3):93–105
Chambers R, Clark R (2012) An introduction to model-based survey sampling with applications, vol 37. Oxford University Press, New York
Chen D (1991) A uniform central limit theorem for nonuniform \(phi\)-mixing random fields. Ann Probab 19(2):636–649
Christakos G (1992) Random field models in earth sciences. Academic Press, San Diego
Chung KL (2001) A course in probability theory. Academic Press, Cambridge
Deo CM et al (1975) A functional central limit theorem for stationary random fields. Ann Probab 3(4):708–715
Doukhan P (1994) Mixing: properties and examples, vol 85. Lecture Notes in Statistics. Springer, New York
Doukhan P (1995) Mixing: properties and examples, volume Lecture Notes in Statistics 85. New York, Springer
Doukhan P, Massart P, Rio E (1994) The functional central limit theorem for strongly mixing processes. Ann de l’IHP Prob et Stat 30(1):63–82
Emery X, Arroyo D (2018) On a continuous spectral algorithm for simulating non-stationary gaussian random fields. Stoch Env Res Risk Assess 32(4):905–919
Fuglstad G-A, Lindgren F, Simpson D, Rue H (2015) Exploring a new class of non-stationary spatial gaussian random fields with varying local anisotropy. Statistica Sinica, 115–133
Guyon X (1995) Random fields on a network: modeling, statistics, and applications. Springer, New York
Ibragimov I (1971) Independent and stationary sequences of random variables. Wolters-Noordhoff Publishing Groningen, The Netherlands
Jha D, Smedsrud PH, Johansen D, de Lange T, Johansen HD, Halvorsen P, Riegler MA (2021) A comprehensive study on colorectal polyp segmentation with resunet++, conditional random field and test-time augmentation. IEEE J Biomed Health Inform 25(6):2029–2040
Matuła P, Rychlik Z (1990) The invariance principle for nonstationary sequences of associated random variables. Ann de l’IHP Prob et Stat 26(3):387–397
Merlevède F, Peligrad M (2000) The functional central limit theorem under the strong mixing condition. Ann Probab 28(3):1336–1352
Newman CM (1984) Asymptotic independence and limit theorems for positively and negatively dependent random variables. Lecture Notes-Monograph Series 5:127–140
Nguyen L, Kodagoda S, Ranasinghe R, Dissanayake G (2021) Mobile robotic sensors for environmental monitoring using gaussian markov random field. Robotica 39(5):862–884
Peligrad M (1985) An invariance principle for \(phi\)-mixing sequences. Ann Probab 13(4):1304–1313
Peligrad M (1998) Maximum of partial sums and an invariance principle for a class of weak dependent random variables. Proc Am Math Soc 126(4):1181–1189
Rio E (2017) Asymptotic theory of weakly dependent random processes, vol 80. Springer, Berlin
Royall RM (1970) On finite population sampling theory under certain linear regression models. Biometrika 57(2):377–387
Shiryaev AN (1996) Probability, volume Graduate Texts in Mathematics 95. New York, Springer
Sroka P, Godunko RJ, Sinitshenkova ND, Prokop J (2021) Life history, systematics and flight ability of the early permian stem-mayflies in the genus misthodotes sellards, 1909 (insecta, ephemerida, permoplectoptera). BMC Ecol Evolut 21(1):1–18
Utev S, Peligrad M (2003) Maximal inequalities and an invariance principle for a class of weakly dependent random variables. J Theor Probab 16(1):101–115
Valliant R, Dorfman A, Royall R (2000) Finite population sampling and inference: a prediction approach. Wiley, New York
Acknowledgements
The authors are grateful to two anonymous reviewers for their constructive comments, helpful suggestions on the revise, and recommended references. This research was partially supported by the grant from the Marsden Fund, New Zealand, the NSERC Strategic Partnership Grant, Canada, and the grant from Pacific Blue Cross Health Informatics Laboratory, Canada.
Funding
Open Access funding enabled and organized by CAUL and its Member Institutions.
Author information
Authors and Affiliations
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
In memory of Alastair Scott, 1939–2017.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Xu, L., Lee, A. & Lumley, T. A functional central limit theorem on non-stationary random fields with nested spatial structure. Stat Inference Stoch Process 26, 215–234 (2023). https://doi.org/10.1007/s11203-022-09273-9
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11203-022-09273-9