Abstract
We establish explicit bounds on the convex distance between the distribution of a vector of smooth functionals of a Gaussian field and that of a normal vector with a positive-definite covariance matrix. Our bounds are commensurate to the ones obtained by Nourdin et al. (Ann Inst Henri Poincaré Probab Stat 46(1):45–58, 2010) for the (smoother) 1-Wasserstein distance, and do not involve any additional logarithmic factor. One of the main tools exploited in our work is a recursive estimate on the convex distance recently obtained by Schulte and Yukich (Electron J Probab 24(130):1–42, 2019). We illustrate our abstract results in two different situations: (i) we prove a quantitative multivariate fourth moment theorem for vectors of multiple Wiener–Itô integrals, and (ii) we characterize the rate of convergence for the finite-dimensional distributions in the functional Breuer–Major theorem.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
Fix \(m\ge 1\) and consider random vectors \(\mathbf{F}\) and \(\mathbf{G}\) with values in \({\mathbb {R}}^m\). The convex distance between the distributions of \(\mathbf{F}\) and \(\mathbf{G}\) is defined as
where the supremum runs over the class \({\mathcal {I}}_m\) of indicator functions of the measurable convex subsets of \({\mathbb {R}}^m\). For \(m\ge 2\), the distance \(d_c\) represents a natural counterpart to the well-known Kolmogorov distance on the class of probability distributions on the real line and enjoys a number of desirable invariance properties that make it well-adapted to applications.Footnote 1
The aim of the present note is to establish explicit bounds on the quantity \(d_{\mathrm{c}}(\mathbf{F }, \mathbf{G })\), in the special case where \(\mathbf{F}\) is a vector of smooth functionals of an infinite-dimensional Gaussian field, and \(\mathbf{G} = N_\Sigma \) is a m-dimensional centered Gaussian vector with covariance \(\Sigma >0\). Our main tool is the so-called Malliavin–Stein method for probabilistic approximations [17], that we will combine with some powerful recursive estimates on \(d_c\), recently derived in [29] in the context of multidimensional second-order Poincaré inequalities on the Poisson space—see Lemma 2.1.
Multidimensional normal approximations in the convex distance have been the object of an intense study since several decades, mostly in connection with multivariate central limit theorems (CLTs) for sums of independent random vectors—see, e.g., [3, 9, 13, 14] for some classical references, as well as [28] for recent advances and for a discussion of further relevant literature. The specific challenge we are setting ourselves in the present work is to establish bounds on the quantity \(d_{\mathrm{c}}(\mathbf{F}, N_{\Sigma })\) that coincide (up to an absolute multiplicative constant) with the bounds deduced in [19] on the 1-Wasserstein distance
where \(\mathrm{Lip}(1)\) denotes the class of \(C^1\) mappings \(h:{\mathbb {R}}^m\rightarrow {\mathbb {R}}\) with Lipschitz constant not exceeding 1. We will see that our estimates systematically improve the bounds that one can infer from the general inequality
where K is an absolute constant uniquely depending on \(\Sigma \). For the sake of completeness, a full proof of (3) is presented in “Appendix A”, where one can also find more details on the constant K.
Remark 1.1
In order for the quantity \(d_W(\mathbf{F}, N_{\Sigma })\) to be well-defined, one needs that \({\mathbb {E}}\Vert \mathbf{F}\Vert _{{\mathbb {R}}^m} < \infty \). In “Appendix A” we will also implicitly use the well-known representation
where the infimum runs over all couplings \(({ \mathbf{U}},\mathbf{V})\) of \(\mathbf{F}\) and \(N_\Sigma \). See [30, Ch. I-6] for a proof of this fact and for further relevant properties of Wasserstein distances.
The main contributions of our paper are described in full detail in Sects. 1.4 and 1.5. Section 1.1 contains some elements of Malliavin calculus that are necessary in order to state our findings. Section 1.2 discusses some estimates on the smooth distance \(d_2\) (to be defined therein) that can be obtained by interpolation techniques, whereas Sect. 1.3 provides an overview of the main results of [19].
Remark on notation. From now on, every random element is assumed to be defined on a common probability space \((\Omega , \mathscr {F}, {\mathbb {P}})\), with \({\mathbb {E}}\) denoting expectation with respect to \({\mathbb {P}}\). For \(p\ge 1\), we write \(L^p(\Omega ) := L^p (\Omega , \mathscr {F}, {\mathbb {P}})\).
1.1 Elements of Malliavin Calculus
The reader is referred, for example, to the monographs [17, 23, 24] for a detailed discussion of the concepts and results presented in this subsection.
Let \(\mathfrak {H}\) be a real separable Hilbert space, and write \(\langle \cdot , \cdot \rangle _\mathfrak {H}\) for the corresponding inner product. In what follows, we will denote by \(X=\{X(h) : h\in \mathfrak {H}\}\) an isonormal Gaussian process over \(\mathfrak {H}\), that is, X is a centered Gaussian family indexed by the elements of \(\mathfrak {H}\) and such that \({\mathbb {E}}[X(h)X(g)] = \langle h,g\rangle _\mathfrak {H}\) for every \(h,g \in \mathfrak {H}\). For the rest of the paper, we will assume without loss of generality that \(\mathscr {F}\) coincides with the \(\sigma \)-field generated by X.
Every \(F\in L^2(\Omega )\) admits a Wiener–Itô chaos expansion of the form
where \(f_q\) belongs to the symmetric qth tensor product \(\mathfrak {H}^{\odot q}\) (and is uniquely determined by F), and \(I_q(f_q)\) is the q-th multiple Wiener–Itô integral of \(f_q\) with respect to X. One writes \(F\in {\mathbb {D}}^{1,2}(\Omega )\) if
For \(F\in {\mathbb {D}}^{1,2}(\Omega )\), we denote by DF the Malliavin derivative of F, and recall that DF is by definition a random element with values in \(\mathfrak {H}\). The operator D satisfies a fundamental chain rule: if \(\varphi :{\mathbb {R}}^m\rightarrow {\mathbb {R}}\) is \(C^1\) and has bounded derivatives and if \(F_1,\dots ,F_m\in {\mathbb {D}}^{1,2}(\Omega )\), then \(\varphi (F_1,\ldots ,F_m)\in {\mathbb {D}}^{1,2}(\Omega )\) and
For general \(p>2\), one writes \(F\in {\mathbb {D}}^{1,p}(\Omega )\) if \(F\in L^p(\Omega )\cap {\mathbb {D}}^{1,2}(\Omega ) \) and \({\mathbb {E}}[\left\Vert {DF}\right\Vert _\mathfrak {H}^p]<\infty \).
The adjoint of D, customarily called the divergence operator or the Skorohod integral, is denoted by \(\delta \) and satisfies the duality formula,
for all \(F\in {\mathbb {D}}^{1,2}(\Omega )\), whenever \(u: \Omega \rightarrow \mathfrak {H}\) is in the domain \(\mathrm{Dom}(\delta )\) of \(\delta \).
The generator of the Ornstein–Uhlenbeck semigroup, written L, is defined by the relation \(L F= - \sum _{q\ge 1} q I_q(f_q)\) for every F such that \(\sum _{q\ge 1} q^2 q! \left\Vert {f_q}\right\Vert _{\mathfrak {H}^{\otimes q}}^2<\infty \). The pseudo-inverse of L, denoted by \(L^{-1}\), is the operator defined for any \(F\in L^2(\Omega )\) as \(L^{-1} F = - \sum _{q\ge 1} \frac{1}{q} I_q(f_q). \) The crucial relation that links the objects introduced above is the identity
which is valid for any \(F\in L^2(\Omega )\) (in particular, one has that, for every \(F\in L^2(\Omega )\), \(DL^{-1}F \in \mathrm{Dom} (\delta )\)).
1.2 Bounds on the Smooth Distance \(d_2\)
Fix \(m\ge 1\) and assume that \(\mathbf{F }=(F_1,...,F_m)\) is a centered random vector in \({\mathbb {R}}^m\) whose components belong to \({\mathbb {D}}^{1,2}(\Omega )\). Without loss of generality, we can assume that each \(F_i\) is of the form \(F_i=\delta (u_i)\) for some \(u_i\in \mathrm{Dom}(\delta )\); indeed, by virtue of (6) one can always set \(u_i=-DL^{-1}F_i\) (although this is by no means the only possible choice). Let also \(N_\Sigma =(N_1,\ldots ,N_m)\) be a centered Gaussian vector with invertible \(m\times m\) covariance matrix \(\Sigma = \{\Sigma (i,j) \} = \{\Sigma (i,j) : i,j = 1,...,m \}\). Finally, consider the so-called \(d_2\) distance (between the distributions of \(\mathbf{F}\) and \(N_\Sigma \)) defined by
where the supremum is taken over all \(C^2\) functions \(h:{\mathbb {R}}^m\rightarrow {\mathbb {R}}\) that are 1-Lipschitz and such that \(\sup _{\mathbf{x}\in {\mathbb {R}}^m}\Vert (\mathrm{Hess}\,h)(\mathbf{x})\Vert _{\mathrm{H.S}.}\le 1\); here, \(\mathrm{Hess}\, h\) stands for the Hessian matrix of h, whereas \(\Vert \cdot \Vert _{\mathrm{H.S}.}\) (resp. \(\langle \cdot ,\cdot \rangle _{\mathrm{H.S}.}\)) denotes the Hilbert–Schmidt norm (resp. scalar product), that is, \(\langle A,B\rangle _{\mathrm{H.S}.}=\mathrm{Tr(AB^T)}=\sum _{1\le i,j\le m}A(i,j)B(i,j)\) and \(\Vert A\Vert ^2_{\mathrm{H.S}.}=\langle A,A\rangle _{\mathrm{H.S}.}\) for any \(m\times m\) matrices \(A=\{A(i,j)\}\) and \(B=\{B(i,j)\}\).
For a given \(h:{\mathbb {R}}^m\rightarrow {\mathbb {R}}\in C^2\) with bounded partial derivatives, let us introduce its mollification at level \(\sqrt{t}\), defined by
One has
Supposing in addition (and without loss of generality) that \({\mathbf{F }}\) and \(N_\Sigma \) are independent, we can write, using the Gaussian integration by parts
and, combining the duality formula (5) with the chain rule (4),
Putting everything together leads to
where \(M_F\) is the random \(m\times m\) matrix given by
It is then immediate that
Inequalities in the spirit of (9) were derived, for example, in [18] (in the context of limit theorems for homogeneous sums) and [27] (in the framework of multivariate normal approximations on the Poisson space)—see also [29] and the references therein.
1.3 Bounds on the 1-Wasserstein Distance
For random vectors \(\mathbf{F}\) and \(N_\Sigma \) as in the previous section, we will now discuss a suitable method for assessing the quantity \(d_W(\mathbf{F },N_\Sigma )\) defined in (2), that is, for uniformly bounding the absolute difference \(|{\mathbb {E}}h(\mathbf{F }) - {\mathbb {E}}h(N_\Sigma ) | \) over all 1-Lipschitz functions h of class \(C^1\).
Since we do not assume h to be twice differentiable, the method presented in Sect. 1.2 no longer works. A preferable approach is consequently the so-called Malliavin–Stein method, introduced in [16] in dimension 1, and later extended to the multivariate setting in [19]. Let us briefly recall how this works (see [17, Chapter 4 and Chapter 6] for a full discussion, and [12] for a constantly updated list of references).
Start by considering the following Stein’s equation, with \(h:{\mathbb {R}}^m\rightarrow {\mathbb {R}}\) given and \(f:{\mathbb {R}}^m\rightarrow {\mathbb {R}}\) unknown:
When \(h\in C^1\) has bounded partial derivatives, it turns out that (10) admits a solution \(f=f_h\) of class \(C^2\) and whose second partial derivatives are bounded—see, e.g., [17, Proposition 4.3.2] for a precise statement. Taking expectation with respect to the distribution of \(\mathbf{F }\) in (10) gives
We can apply again the duality formula (5) together with the chain rule (4), to deduce that
where \(M_F\) is defined in (8). Taking the supremum over the set of all 1-Lipschitz functions \(h:{\mathbb {R}}^m\rightarrow {\mathbb {R}}\) of class \(C^1\), we infer
with
and \(\left\Vert {\cdot }\right\Vert _{\mathrm{op}}\) is the operator norm for \(m\times m\) matrices. Estimate (12) is the main result of [19] (see also [17, Theorem 6.1.1]), whereas a self-contained proof of (13) can be found in [17, Proposition 4.3.2].
1.4 Main Results: Bounds on the Convex Distance
The principal aim of the present paper is to address the following natural question: can one obtain a bound similar to (12) for distances based on non-smooth test functions \(h:{\mathbb {R}}^m\rightarrow {\mathbb {R}}\), such as, for example, indicator functions of measurable convex subsets of \({\mathbb {R}}^m\)?
If h is such an indicator function, then we recall, for example, from [29, Lemma 2.2] that, for all \(t\in (0,1)\),
where \(h_t\) stands for the mollification at level \(\sqrt{t}\) of h, as defined in (7). Let \(f_t=f_{h_t}\) be the solution of the Stein’s equation (10) associated with \(h=h_t\). In [5] (see also [29]), it is shown that
with \(c_2=c_2(m,\Sigma )\) a constant depending only on m and \(\Sigma \). Combining such an estimate with (11) yields the existence of a constant \(c_3=c_3(m,\Sigma )>0\) such that
From (14), it is straightforward to deduce the existence of \(c_4=c_4(m,\Sigma )>0\) such that
Comparing (15) with (9) and (12) shows that such a strategy yields a bound on \(d_{\mathrm{c}}(\mathbf{F },N_\Sigma )\) differing from those deduced above for the distances \(d_2\) and \(d_W\) by an additional logarithmic factor. See also [10, 20] for more inequalities analogous to (15)—that is, displaying a multiplicative logarithmic factor—related, respectively, to the (multivariate) Kolmogorov and total variation distances.
In this paper, we will show that one can actually remove the redundant logarithmic factor on the right-hand side of (15), thus yielding a bound on \(d_{\mathrm{c}}(\mathbf{F}, N_\Sigma )\) that is commensurate to (9) and (12) (with moreover an explicit multiplicative constant). Our main result is the following:
Theorem 1.2
Let \(\mathbf{F }=(F_1,...,F_m)=(\delta (u_1),\ldots ,\delta (u_m))\) be a vector in \({\mathbb {R}}^m\) of centered random variables such that \(u_i\in \mathrm{Dom}(\delta )\), for \(i=1,...,m\). Let also \(N_\Sigma =(N_1,\ldots ,N_m)\) be a centered Gaussian vector with invertible \(m\times m\) covariance matrix \(\Sigma =\{\Sigma (i,j)\}\). Then
with \(M_F\) defined in (8).
As anticipated, to prove Theorem 1.2, we shall combine the somewhat classical smoothing estimate (14) with a remarkable bound by Schulte and Yukich [29].
1.5 Applications
We illustrate the use of Theorem 1.2 by developing two examples in full detail.
Quantitative fourth moment theorems. A fourth moment theorem (FMT) is a mathematical statement, implying that a given sequence of centered and normalized random variables converges in distribution to a Gaussian limit, as soon as the corresponding sequence of fourth moments converges to 3 (that is, to the fourth moment of the standard Gaussian distribution). Distinguished examples of FMTs are, for example, de Jong’s theorem for degenerate U-statistics (see [7, 8]) as well as the CLTs for multiple Wiener–Itô integrals proved in [25, 26]; the reader is referred to the webpage [12] for a list (composed of several hundreds of papers) of applications and extensions of such results, as well as to the lecture notes [31] for a modern discussion of their relevance in mathematical physics. Our first application of Theorem 1.2 is a quantitative multivariate fourth moment theorem for a vector of multiple Wiener–Itô integrals, considerably extending the qualitative multivariate results proved in [26]. Note that such a result was already obtained by Nourdin and Rosiński [22, Theorem 4.3] for the 1-Wasserstein distance \(d_W\). Thanks to Theorem 1.2, it is not difficult to generalize their result to the \(d_{\mathrm{c}}\) metric.
Corollary 1.3
Fix \(m\ge 1\) as well as \(q_1,\ldots ,q_m\ge 1\). Let \(\mathbf{F } = (F_1,...,F_m)\) where \(F_i = I_{q_{i}}(f_{q_i})\) with \(f_{q_i}\in {\mathfrak {H}}^{\odot q_i}\). Let \(N_\Sigma \) be a centered Gaussian vector with covariance matrix \(\Sigma = ({\mathbb {E}}F_iF_j)_{i,j\in [m]}\) supposed to be invertible. Then
In particular, for a vector \(\mathbf{F}\) of multiple Wiener–Itô integrals to be close in the convex distance to a centered Gaussian vector \(N_\Sigma \) with matching covariance matrix, it is enough that \({\mathbb {E}}\Vert \mathbf{F}\Vert ^4\approx {\mathbb {E}}\left\Vert {N_\Sigma }\right\Vert ^4\).
The multivariate Breuer–Major theorem. The second example concerns the convergence toward a Brownian motion occurring in the Breuer–Major theorem proved in [4]. Let us briefly recall this fundamental result (see [17, Chapter 7] for an introduction to the subject, as well as [6, 15] for recent advances in a functional setting). Let \(\{G_k : k\in {\mathbb {Z}}\}\) be a centered Gaussian stationary sequence with \(\rho (j-k)={\mathbb {E}}[G_jG_k]\) and \(\rho (0)=1\); in particular, \(G_k\sim N(0,1)\) for all k. Let \(\varphi \in L^2({\mathbb {R}},\gamma )\) where \(\gamma (dx)=(2\pi )^{-1/2}e^{-x^2/2}dx\) denotes the standard Gaussian measure on \({\mathbb {R}}\). Since the Hermite polynomials \(\{H_k : k\ge 0\}\) form an orthonormal basis of \(L^2({\mathbb {R}},\gamma )\), one has
with \(d\in {\mathbb {N}}\) and \(a_d\ne 0\). The index d is known as the Hermite rank of \(\varphi \in L^2({\mathbb {R}},\gamma )\). Suppose in addition that \(\int _{\mathbb {R}}\varphi d\gamma ={\mathbb {E}}[\varphi (G_0)]=0\), that is, suppose \(d\ge 1\). The Breuer–Major theorem [4] states the following: if \(\sum _{k\in {\mathbb {Z}}}|\rho (k)|^d<\infty \), then
where W is a standard Brownian motion, \(\overset{f.d.d.}{\longrightarrow }\) indicates convergence in the sense of finite-dimensional distributions, and
(That \(\sigma ^2\) is a well-defined positive real number is part of the conclusion.) We refer to our note [21] and references therein for results on the rate of convergence in the total variation distance for one-dimensional marginal distributions (that is, in dimension 1). We intend to apply Theorem 1.2 to address the rate of convergence for the following multivariate CLT implied by (16): for every \(0=t_0<t_1<...<t_m=T<\infty \),
where \(\overset{d}{\longrightarrow }\) indicates converges in distribution, and \(N(0,\Sigma (t_1, ..., t_m))\) is a m-dimensional centered Gaussian vector with covariance \(\Sigma (t_1, ..., t_m)\) having entries \(\sigma ^2 t_i\wedge t_j\), \(i,j=1,...,m\). Notice that for any \(m\times m\) invertible matrix A,
Therefore, choosing A appropriately, it suffices to consider the vector \(\mathbf{F }_n =(F_{n,1},...,F_{n,m})\) with
and obtain the rate of convergence for
The following result provides a quantitative version of this CLT with respect to the distance \(d_{\mathrm{c}}\). Recall from [21] that the minimal regularity assumption over \(\varphi \) for obtaining rates of convergence via the Malliavin–Stein method is that \(\varphi \in {\mathbb {D}}^{1,4}({\mathbb {R}},\gamma )\), meaning that \(\varphi \) is absolutely continuous and both \(\varphi \) and its derivative \(\varphi '\) belong to \(L^4({\mathbb {R}},\gamma )\). We say that \(\varphi \) is 2-sparse if its expansion in Hermite polynomials does not have consecutive nonzero coefficients. In particular, even functions are 2-sparse.
Corollary 1.4
Let \(\mathbf{F }_n\) and \(N_\Sigma \) be given in (17). Suppose that \(\varphi \in {\mathbb {D}}^{1,4}({\mathbb {R}},\gamma )\) with Hermite rank \(d\ge 1\). Then,
-
(i)
There exists a constant C depending only on \(\varphi , m, \Sigma \) such that for each \(n\in {\mathbb {N}}\),
$$\begin{aligned} d_{\mathrm{c}}(\mathbf{F }_n, N_\Sigma ) \le C\sum _{i,j=1}^m |\Sigma (i,j) -{\mathbb {E}}[F_iF_j]|+ Cn^{-\frac{1}{2}} \left( \sum _{|k|< n} |\rho (k)|\right) ^\frac{3}{2}. \end{aligned}$$ -
(ii)
If \(d=2\), \(\varphi \) is 2-sparse and \(b\in [1,2]\), then there exists a constant C depending only on \(\varphi , m, \Sigma \) such that for each \(n\in {\mathbb {N}}\),
$$\begin{aligned} d_{\mathrm{c}}(\mathbf{F }_n, N_\Sigma )&\le C\sum _{i,j=1}^m |\Sigma (i,j) - {\mathbb {E}}[F_iF_j]|\\&\quad + C n^{-(\frac{1}{b}-\frac{1}{2})} \left( \sum _{|k|<n}|\rho (k)|^2\right) ^{\frac{1}{2}} \left( \sum _{|k|< n}|\rho (k)|^b\right) ^{\frac{1}{b}} . \end{aligned}$$ -
(iii)
If \(d=2\), \(\varphi \) is 2-sparse, and \(\sum _{k\in {\mathbb {Z}}}|\rho (k)|^2<\infty \), then as \(n\rightarrow \infty \),
$$\begin{aligned} d_{\mathrm{c}}(\mathbf{F }_n, N_\Sigma )\rightarrow 0. \end{aligned}$$
The rest of the note is organized as follows: The proof of Theorem 1.2 is given in Sect. 2.1, Corollary 1.3 in Sect. 2.2, Corollary 1.4 in Sect. 2.3. We use C to denote a generic constant whose value may change from line to line.
2 Proofs
2.1 Proof of Theorem 1.2
We divide the proof into several steps.
Step 1 (smoothing). For any bounded and measurable h and \(t\in (0,1)\), recall its mollification at level \(\sqrt{t}\) from (7). Then it is plain that \(h_t\) is \(C^\infty \) with bounded derivatives of all orders and the solution to (10) with \(h=h_t\) is given by
see [29, p.12]. Finally, recall from, for example, [29, Lemma 2.2] that, for any \(t\in (0,1)\),
Step 2 (integration by parts). An integration by parts by (5) and (4) (see [17, Chapter 4] for more details), together with Cauchy–Schwarz’s inequality, implies,
The following remarkable estimate is due to M. Schulte and J. Yukich.
Lemma 2.1
(Proposition 2.3 in [29]) Let \(\mathbf{Y} \) be an \({\mathbb {R}}^m\)-valued random vector and \(\Sigma \) be an invertible \(m\times m\) covariance matrix. Then,
where the left-hand side depends on h through the function \(f_t\) solving Stein’s equation with test function \(h_t\) given by (7).
Remark 2.2
Lemma 2.1 improves upon the uniform bound (see [5] or [29])
when some a priori estimate on \(d_{\mathrm{c}}(\mathbf{Y },N_\Sigma )\) is available.
As a consequence,
Letting
we have thus established
Step 3 (exploiting the recursive inequality). If \(\gamma \ge 1/e\), then the bound we intend to prove holds trivially (observe that \(d_{\mathrm{c}}(\mathbf{F },N_\Sigma )\le 1\) by definition). Without loss of generality, we can and will therefore assume that \(\gamma \le 1/e\). Let \(t=\gamma ^2\). Using the fact that \(\kappa \le 1\) for the \(\kappa \) on the right-hand side of the (19), one has
Therefore,
Since \(\sup _{x\in (0, 1/e]} x^{1/2}| \log x |^{3/2}\le 4\), one has
Hence, putting the estimate back into (19) with \(t=\gamma ^2\) gives
The proof is complete.
2.2 Proof of Corollary 1.3
We will obtain the desired conclusion as a direct application of Theorem 1.2 with \(u_i=-DL^{-1}F_i\), see (6). Indeed, recall that by Step 2 of [22, Proof of Theorem 4.3], for any \(i,j\in [m]\),
On the other hand, Step 3 of [22, Proof of Theorem 4.3] shows that
Plugging these estimates into Theorem 1.2 gives the result.
2.3 Proof of Corollary 1.4
We follow closely the arguments of [21] and assume without loss of generality that \(T=1\). First, one can embed the Gaussian sequence in the statement in an isonormal Gaussian process \(\{X(h) : h\in \mathfrak {H} \}\), in such a way that
for some appropriate family \(\{e_k\}\subset \mathfrak {H}\) verifying \(\langle e_j , e_k \rangle _{\mathfrak {H}} = \rho (k-j)\) for all j, k. For \(\varphi =\sum _{\ell \ge d}a_\ell H_\ell \in L^2({\mathbb {R}},\gamma )\), we define the shift mapping \(\varphi _1 := \sum _{\ell \ge 1}a_\ell H_{\ell -1}\) and set
Then, standard computations using (6) lead to
Applying Theorem 1.2 and the triangle inequality implies that
Note that, by the chain rule and the relation \(D(G_k) =e_k\),
where \(k\sim t_i\) means that the sum is taken over \(k\in \{ \lfloor nt_{i-1}\rfloor +1,..., \lfloor nt_i\rfloor \}\) and similarly for the symbol \(\ell \sim t_j\). Hence,
The variance is bounded because of the assumption that \(\varphi \in {\mathbb {D}}^{1,4}\). Once (21) is in place, one can apply Gebelein’s inequality as in [21]. In particular, one infers that (see [21, Proposition 3.4])
If, in addition, \(\varphi \) is 2-sparse, then
Items i)-ii) now follow from these inequalities, as shown in [21]; we include a proof for completeness. Applying twice Young’s inequality for convolutions, one has, with \(\rho _n(k)=|\rho (k)|{1}_{|k|<n}\),
yielding Item i). Rewrite the sum of products as a sum of the product of convolutions by introducing the function \(1_n(k):= 1_{|k|<n}\). We have
For \(b\in [1,2]\), we have
yielding Item ii). Now we move to the proof of Item iii). Notice that taking \(b=2\) for the right-hand side of (22), together with an application of Young’s inequality, yields that
Thus,
To proceed, we handle the convolution involving \(1_n\) a bit differently. Set
so that \(\rho _n=\widetilde{\rho }_n + \widehat{\rho }_n\). One has
from which Item iii) follows. The proof is complete.
References
Azmoodeh, E., Peccati, G., Poly, G.: The law of iterated logarithm for subordinated Gaussian sequences: uniform Wasserstein bounds. ALEA 13, 659–686 (2016)
Ball, K.: The reverse isoperimetric problem for the Gaussian measure. Discrete Comput. Geom. 10(4), 4111–420 (1993)
Bentkus, V.: On the dependence of the Berry-Esseen bound on dimension. J. Stat. Plan. Inference 113, 385–402 (2003)
Breuer, P., Major, P.: Central limit theorems for nonlinear functionals of Gaussian fields. J. Multivar. Anal. 13(3), 425–441 (1983)
Chen, L.H.Y., Goldstein, L., Shao, Q.-M.: Normal Approximation by Stein’s Method. Probability and Its Applications (New York). Springer, Heidelberg (2011)
Campese, S., Nourdin, I., Nualart, D.: Continuous Breuer–Major theorem: tightness and non-stationarity. Ann. Probab. 48(1), 147–177 (2020)
de Jong, P.: A central limit theorem for generalized multilinear forms. J. Multivar. Anal. 34(2), 275–289 (1990)
Döbler, Ch., Peccati, G.: Quantitative de Jong theorems in any dimension. Electron. J. Probab. 22(2), 1–35 (2017)
Götze, F.: On the rate of convergence in the multivariate CLT. Ann. Probab. 19(2), 724–739 (1991)
Kim, Y.T., Park, H.S.: Kolmogorov distance for multivariate normal approximation. Korean J. Math. 23(1), 1–10 (2015)
Koike, Y.: High-dimensional central limit theorems for homogeneous sums. Preprint (2019)
Malliavin–Stein approach: a webpage maintained by Ivan Nourdin. https://sites.google.com/site/malliavinstein/home
Nagaev, S.V.: An estimate of the remainder term in the multidimensional central limit theorem. In: Proceedings of the Third Japan—USSR Symposium on Probability Theory (Tashkent, 1975). Lecture Notes in Math., vol. 550, pp. 419–438. Springer, Berlin (1976)
Nazarov, F.: On the maximal perimeter of a convex set in {\mathbb{R}}^n with respect to a Gaussian measure. In: Milman, V.D., Schechtman, G. (eds.) Geometric Aspects of Functional Analysis. Lecture Notes in Mathematics, vol. 1807, pp. 169–187. Springer, Berlin (2003)
Nourdin, I., Nualart, D.: The functional Breuer–Major theorem. Probab. Theory Relat. Fields 176, 203–218 (2020)
Nourdin, I., Peccati, G.: Stein’s method on Wiener chaos. Probab. Theory Relat. Fields 145, 75–118 (2009)
Nourdin, I., Peccati, G.: Normal Approximations with Malliavin Calculus. From Stein’s Method to Universality. Cambridge Tracts in Mathematics, vol. 192. Cambridge University Press, Cambridge (2012)
Nourdin, I., Peccati, G., Reinert, G.: Invariance principles for homogeneous sums: universality of Gaussian Wiener chaos. Ann. Probab. 38(5), 1947–1985 (2010)
Nourdin, I., Peccati, G., Réveillac, A.: Multivariate normal approximation using Stein’s method and Malliavin calculus. Ann. Inst. Henri Poincaré Probab. Stat. 46(1), 45–58 (2010)
Nourdin, I., Peccati, G., Swan, Y.: Entropy and the fourth moment phenomenon. J. Funct. Anal. 266(5), 3170–3207 (2014)
Nourdin, I., Peccati, G., Yang, X.: Berry-Esseen bounds in the Breuer–Major CLT and Gebelein’s inequality. Electron. Commun. Probab. 24(34), 12 (2019)
Nourdin, I., Rosiński, J.: Asymptotic independence of multiple Wiener-Itô integrals and the resulting limit laws. Ann. Probab. 42(2), 497–526 (2014)
Nualart, D.: The Malliavin Calculus and Related Topics, 2nd edn. Springer, Berlin (2006)
Nualart, D., Nualart, E.: Introduction fo Malliavin Calculus. Cambridge University Press, Cambridge (2018)
Nualart, D., Peccati, G.: Central limit theorems for sequences of multiple stochastic integrals. Ann. Probab. 33(1), 177–193 (2005)
Peccati, G., Tudor, C.A.: Gaussian limits for vector-valued multiple stochastic integrals. In: Séminaire de Probabilités XXXVIII, pp. 247–262 (2004)
Peccati, G., Zheng, C.: Multi-dimensional Gaussian fluctuations on the Poisson space. Electron. J. Probab. 15, 1487–1527 (2010)
Raic, M.: A multivariate Berry-Esseen theorem with explicit constants. Bernoulli 25(4A), 2824–2853 (2019)
Schulte, M., Yukich, J.E.: Multivariate second order Poincaré inequalities for Poisson functionals. Electron. J. Probab. 24(130), 1–42 (2019)
Villani, C.: Optimal Transport. Old and New. Grundlehren der mathematischen Wissenschaften, vol. 338. Springer, Berlin (2009)
Zygouras, N.: Discrete stochastic analysis. Lecture notes available on the webpage. https://warwick.ac.uk/fac/sci/statistics/staff/academic-research/zygouras/ (2019)
Acknowledgements
We thank Simon Campese and Nicola Turchi for pointing out an error in an earlier version. I. Nourdin was supported by the FNR grant APOGee (R-AGR-3585-10) at Luxembourg University; G. Peccati is supported by the FNR grant FoRGES (R-AGR-3376-10) at Luxembourg University; X. Yang was supported by the FNR Grant MISSILe (R-AGR-3410-12-Z) at Luxembourg and Singapore Universities.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
A Proof and Discussion of Relation (3)
A Proof and Discussion of Relation (3)
Inequality (3) is a direct consequence of the following statement, whose proof exploits a strategy already adopted in [1, Proof of Theorem 3.1].
Proposition A.1
Fix \(m\ge 1\), and let \(N_\Sigma \) denote a m-dimensional centered Gaussian vector with invertible covariance matrix \(\Sigma \). Then, for any m-dimensional random vector \(\mathbf{F}\) one has that
where \(\Gamma (\Sigma )\) is the isoperimetric constant defined by
where Q ranges over all Borel measurable convex subsets of \({\mathbb {R}}^m\), and \(Q^{\epsilon }\) indicates the set of all elements of \({\mathbb {R}}^m\) whose Euclidean distance from Q does not exceed \(\epsilon \).
Remark A.2
In [14] it is proved that, for some absolute constants \(0<c<C<\infty \),
where \(\Vert \cdot \Vert _{\mathrm{H.S}.}\) stands as above for the Hilbert–Schmidt norm. When \(\Sigma = I_m\) (identity matrix), one has also the well-known estimate \(\Gamma (I_m) \le 4m^{1/4}\) (see [2]), as well as Nazarov’s upper and lower bounds
see [14, p. 170]. In [28, Theorem 1.2], it is proved that Nazarov’s upper bound can be reduced from 0.64 to 0.59; see also [3] for related computations in the framework of the multivariate CLT.
Proof of Proposition A.1
We can assume that \(\mathbf{F}\) and \(N_{\Sigma }\) are defined on a common probability space and that \({\mathbb {E}} \Vert \mathbf{F} - N_{\Sigma }\Vert _{{\mathbb {R}}^m} = d_W(\mathbf{F}, N_{\Sigma })\). Fix a convex set Q, as well as \(\epsilon >0\). We have that
On the other hand, defining \(Q^{-\epsilon }\) as the set of those \(y\in Q\) such that the closed ball with radius \(\epsilon \) centered at y is contained in Q,
where we have used the inequality
The conclusion follows from a standard optimization in \(\epsilon \). \(\square \)
Remark A.3
Fix \(m\ge 1\), and let \(\mathscr {R}_m\) be the collection of all hyper-rectangles of the type \(R = (-\infty , t_1]\times \cdots \times (-\infty , t_m]\). In [1, Theorem 3.1] it is proved that, if N is a m-dimensional centered Gaussian vector with identity covariance matrix and \(\mathbf{F}\) is any m-dimensional random vector, then
The left-hand side of the previous inequality is usually referred to as the Kolmogorov distance between the distributions of \(\mathbf{F}\) and N. The presence of the factor \((\log m)^{1/4}\) is consistent with the fact that, for the standard Gaussian measure on \({\mathbb {R}}^m\), the isoperimetric constant associated with all hyper-rectangles of \({\mathbb {R}}^m\) is bounded from above by \(\sqrt{\log m}\), see [2, 14]. An estimate analogous to (24) is established by different methods in [11, Corollary 3.1].
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Nourdin, I., Peccati, G. & Yang, X. Multivariate Normal Approximation on the Wiener Space: New Bounds in the Convex Distance. J Theor Probab 35, 2020–2037 (2022). https://doi.org/10.1007/s10959-021-01112-6
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10959-021-01112-6
Keywords
- Breuer–Major Theorem
- Convex distance
- Fourth moment theorems
- Gaussian fields
- Malliavin–Stein method
- Multidimensional normal approximations