Abstract
We consider a class of sparse random matrices which includes the adjacency matrix of the Erdős–Rényi graph \({{\mathcal {G}}}(N,p)\). We show that if \(N^{\varepsilon } \leqslant Np \leqslant N^{1/3-\varepsilon }\) then all nontrivial eigenvalues away from 0 have asymptotically Gaussian fluctuations. These fluctuations are governed by a single random variable, which has the interpretation of the total degree of the graph. This extends the result (Huang et al. in Ann Prob 48:916–962, 2020) on the fluctuations of the extreme eigenvalues from \(Np \geqslant N^{2/9 + \varepsilon }\) down to the optimal scale \(Np \geqslant N^{\varepsilon }\). The main technical achievement of our proof is a rigidity bound of accuracy \(N^{-1/2-\varepsilon } (Np)^{-1/2}\) for the extreme eigenvalues, which avoids the \((Np)^{-1}\)-expansions from Erdős et al. (Ann Prob 41:2279–2375, 2013), Huang et al. (2020) and Lee and Schnelli (Prob Theor Rel Fields 171:543–616, 2018). Our result is the last missing piece, added to Erdős et al. (Commun Math Phys 314:587–640, 2012), He (Bulk eigenvalue fluctuations of sparse random matrices. arXiv:1904.07140), Huang et al. (2020) and Lee and Schnelli (2018), of a complete description of the eigenvalue fluctuations of sparse random matrices for \(Np \geqslant N^{\varepsilon }\).
We’re sorry, something doesn't seem to be working properly.
Please try refreshing the page. If that doesn't work, please contact support so we can address the problem.
Avoid common mistakes on your manuscript.
1 Introduction and main results
Let \({{\mathcal {A}}}\) be the adjacency matrix of the Erdős–Rényi graph \({{\mathcal {G}}}(N,p)\). Explicitly, \({{\mathcal {A}}} = ({\mathcal {A}}_{ij})_{i,j = 1}^N\) is a symmetric \(N\times N\) matrix with independent upper triangular entries \(({\mathcal {A}}_{ij} :i \leqslant j)\) satisfying
We introduce the normalized adjacency matrix
where the normalization is chosen so that the eigenvalues of A are typically of order one.
The goal of this paper is to obtain the asymptotic distribution of the extreme eigenvalues of A. The extreme eigenvalues of graphs are of fundamental importance in spectral graph theory and have attracted much attention in the past 30 years; see for instance [1, 4, 17] for reviews. The Erdős–Rényi graph is the simplest model of a random graph and its adjacency matrix is the canonical example of a sparse random matrix.
Each row and column of A has typically Np nonzero entries, and hence A is sparse whenever \(p \rightarrow 0\) as \(N \rightarrow \infty \). In the complementary dense regime, where p is of order one, A is a Wigner matrix (up to a centring of the entries). The edge statistics of Wigner matrices have been fully understood in [8, 10, 23, 26,27,28], where it was shown that the distribution of the largest eigenvalue is asymptotically given by the GOE Tracy–Widom distribution [29, 30].
To discuss the edge statistics of A in the sparse regime, we introduce the following conventions. Unless stated otherwise, all quantities depend on the fundamental parameter N, and we omit this dependence from our notation. We write \(X \ll Y\) to mean \(X = O_\varepsilon (N^{-\varepsilon } Y)\) for some fixed \(\varepsilon > 0\). We write \(X \asymp Y\) to mean \(X = O(Y)\) and \(Y = O(X)\). We denote the eigenvalues of A by \(\lambda _1 \leqslant \cdots \leqslant \lambda _N\). The largest eigenvalue \(\lambda _N\) of A is its Perron–Frobenius eigenvalue. For \(Np \gg 1\), it is typically of order \(\sqrt{Np}\), while the other eigenvalues \(\lambda _1, \lambda _2, \ldots , \lambda _{N-1}\) are typically of order one.
The edge statistics of sparse matrices were first studied in [8, 9], where it was proved that when \(Np \gg N^{2/3}\) the second largest eigenvalue of A exhibits GOE Tracy–Widom fluctuations, i.e.
where \(F_1(s)\) is the distribution function of the GOE Tracy–Widom distribution. In [24], this result was extended to \(Np \gg N^{1/3}\), which it turns out is optimal. Indeed, in [19] it was shown that when \(N^{2/9}\ll Np \ll N^{1/3}\) the Tracy–Widom distribution for \(\lambda _{N - 1}\) no longer holds, and the extreme eigenvalues have asymptotically Gaussian fluctuations. More precisely, in [19] it was shown that if \(N^{2/9}\ll Np \ll N^{1/3}\) then
In this paper we show (1.2) for the whole range \(1 \ll Np \ll N^{1/3}\). In fact, we show this for a general class of sparse random matrices introduced in [8, 9]. It is easy to check that the normalized adjacency matrix A (1.1) of \({\mathcal {G}}(N,p)\) satisfies the following definition with the choice
Definition 1.1
(Sparse matrix). Let \(1 \leqslant q \leqslant \sqrt{N}\). A sparse matrix is a real symmetric \(N\times N\) matrix \(H=H^* \in {\mathbb {R}}^{N \times N}\) whose entries \(H_{ij}\) satisfy the following conditions.
-
(i)
The upper-triangular entries (\(H_{ij}:1 \leqslant i \leqslant j\leqslant N\)) are independent.
-
(ii)
We have \({\mathbb {E}} H_{ij}=0\), \( {\mathbb {E}} H_{ij}^2=(1+O(\delta _{ij}))/N\), and \({\mathbb {E}} H_{ij}^4\asymp 1/(Nq^2)\) for all i, j.
-
(iii)
For any \(k\geqslant 3\), we have \({\mathbb {E}}|H_{ij}|^k \leqslant C_k/ (Nq^{k-2})\) for all i, j.
We define the random matrix
where \({{\mathbf {e}}} :=N^{-1/2}(1,1,\ldots ,1)^*\), and \(f \geqslant 0\).
For simplicity of presentation, in this paper we focus only on real matrices, although our results and proofs extend to matrices with complex entries with minor modifications which we omit; see also Remark 8.2 below.
To describe the fluctuations of the eigenvalues of A, we define the random variable
Defining
one easily finds
We denote by \(\gamma _{\mathrm {sc},i}\) be the ith N-quantile of the semicircle distribution, which is the limiting empirical eigenvalue measure of A for \(Np \gg 1\). Explicitly, \(\int _{-2}^{\gamma _{\mathrm {sc},i}}\frac{1}{2\pi }\sqrt{4-x^2} \, \mathrm {d}x =\frac{i}{N}\,\).
Throughout the following we fix an exponent \(\beta \in (0,1/2]\) and set
If A is the normalized adjacency matrix (1.1) of \({\mathcal {G}}(N,p)\) then from (1.3) and (1.6) we find that the condition \(1 \ll Np \ll N^{1/3}\) reads \(1 \ll q \ll N^{1/6}\), i.e. \(\beta \in (0,1/6)\). We may now state our main result.
Theorem 1.2
Fix \(\beta \in (0,1/6)\) and set
Let H be as in Definition 1.1 with q given by (1.6). Fix \(\varepsilon > 0\) and \(D > 0\). Then for large enough N we have with probability at least \(1 - N^{-D}\)
for all \(1\leqslant i \leqslant N-1\).
Theorem 1.2 implies, for all \(i\in \{1,2,\ldots ,N-1\}\) such that \(\gamma _{\mathrm {sc},i}\) is away from 0, that the fluctuations of \(\lambda _i\) are simultaneously governed by those of \({\mathcal {Z}}\). In fact, by the rigidity result of [9, Theorem 2.13] and a simple moment estimate of \({\mathcal {Z}}\) [see (2.5) below], we deduce from (1.5) and Theorem 1.2 that under its conditions, with probability at least \(1 - N^{-D}\) we have
for all \(i = 1, \ldots , N - 1\). Thus, for \(1 \ll q \ll N^{1/6}\), the fluctuation of all eigenvalues away from 0 is given by a global random scaling by the factor \(1 + {\mathcal {Z}}/2\).
Remark 1.3
If \(f = 0\) in Definition 1.1, i.e. \(A = H\) is centred, then the conclusion of Theorem 1.2 holds for all eigenvalues \(\lambda _1, \ldots , \lambda _N\). Indeed, if \(f = 0\) then A and \(-A\) both satisfy Definition 1.1, and \(\lambda _N(A) = - \lambda _1(-A)\).
Our main result is a rigidity estimate for the eigenvalues of A with accuracy
In contrast, the corresponding rigidity results of [9, 19, 24] have accuracy up to a fixed power of \(q^{-1}\): up to \(q^{-2}\) in [9], \(q^{-4}\) in [24], and \(q^{-6}\) in [19]. For arbitrarily small polynomial values of q, the rigidity provided by an expansion up to a fixed power of \(q^{-1}\) is not sufficient to analyse the fluctuations of the extreme eigenvalues. Thus, the main technical achievement of our paper is the avoidance of \(q^{-1}\)-expansions in the error bounds.
Remark 1.4
The variable \({\mathcal {Z}}\) was introduced in [19], where its importance for the edge fluctuations of sparse random matrices was first recognized. Using it, the authors proved (1.8) for \(\beta \in (1/9,1/6)\).
Remark 1.5
Let A be the rescaled adjacency matrix (1.1) of \({\mathcal {G}}(N,p)\). The fluctuations of the eigenvalues of A have a particularly transparent interpretation in terms of the fluctuation of the average degree of \({\mathcal {G}}(N,p)\), or, equivalently, its total number of edges. To that end, denote by \({\mathcal {D}} :=\frac{1}{N} \sum _{i,j} {\mathcal {A}}_{ij}\) the average degree of \({\mathcal {G}}(N,p)\) and by \(d :={\mathbb {E}}{\mathcal {D}} = Np\) its expectation. Defining the randomly rescaled adjacency matrix
we claim that under the assumptions of Theorem 1.2 we have
with probability at least \(1 - N^{-D}\). Indeed, a short calculation yields \({\mathcal {D}} = d \bigl ({1 + (1 - p) {\mathcal {Z}} + O (p)}\bigr )\), from which (1.11) follows using (1.9) and the bounds \(p = O(N^{-\delta } \Sigma )\) and \(|{\mathcal {Z}} |^2 = O(N^{-\delta } \Sigma )\) with probability at least \(1 - N^{-D}\) [by (2.5) below].
In (1.11), the Gaussian fluctuations (1.9) present for \(\lambda _i(A)\) are absent for \(\lambda _i({\widehat{A}})\). Hence, the fluctuations of the eigenvalues of A can be all simultaneously eliminated to leading order by an appropriate random rescaling. Note that we can write \(A = d^{-1/2} {\mathcal {A}} (1 + O(N^{-\delta } \Sigma ))\), in analogy to (1.10). Thus, (1.11) states that if one replaces the deterministic normalization \(d^{-1/2}\) with the random normalization \({\mathcal {D}}^{-1/2}\) the fluctuations vanish to leading order. In fact, although it is not formulated that way, our proof can essentially be regarded as a rigidity result for the matrix \({\widehat{A}}\).
Remark 1.5 is consistent with the fact that for more rigid graph models where the average degree is fixed, \({\mathcal {Z}}\) does not appear: for a random d-regular graph, the second largest eigenvalue of the adjacency matrix has Tracy–Widom fluctuations for \(N^{2/9}\ll d \ll N^{1/3}\) [2]. Moreover, in [19] it was proved that the second largest eigenvalue of \({\widehat{A}}\) has Tracy–Widom fluctuations for \(q \gg N^{1/9}\).
Theorem 1.2 trivially implies the following result.
Corollary 1.6
We adopt the conditions in Theorem 1.2. Fix \(\varepsilon >0\). Define
for all \(i \in \{1,2,\ldots , \lfloor {(\frac{1}{2}-\varepsilon ) N} \rfloor , \lfloor {(\frac{1}{2}+\varepsilon ) N} \rfloor ,\ldots ,N-1\}=:{\mathcal {I}}\). We have
for all fixed k and \(i_1,\ldots ,i_k \in {\mathcal {I}}\). Here \({\mathcal {J}}\in {{\mathbb {R}}}^{k\times k}\) is the matrix of ones, i.e. \({\mathcal {J}}_{ij}= 1\) for all \(i,j\in \{1,2,\ldots ,k\}\).
Next, we remark on the fluctuations of single eigenvalues inside the bulk. This problem was first addressed in [11] for GUE, extended to GOE in [25], and recently extended to general Wigner matrices in [3, 21]. In these works, it was proved that the bulk eigenvalues of Wigner matrices fluctuate on the scale \(\sqrt{\log N}/N\). More precisely,
for all bulk eigenvalues \(\mu _i\), \(\varepsilon N \leqslant i \leqslant (1 - \varepsilon ) N\), of a real Wigner matrix. The bulk eigenvalue fluctuation of sparse matrices was studied in [12], where it was shown that for fixed \(\beta \in (0,1/2)\), there exists \(c\equiv c(\beta )>0\) such that with probability at least \(1 - N^{-D}\)
for all bulk eigenvalues \(\lambda _i\), \(\varepsilon N \leqslant i \leqslant (1 - \varepsilon ) N\), of A.
In summary, we have the following general picture of fluctuations of eigenvalues for sparse random matrices. The fluctuations of any single eigenvalue consists of two components: a random matrix component and a sparseness component. The random matrix component is independent of the sparseness and coincides with the corresponding fluctuations of GOE. It has order \(N^{-2/3}\) at the edge and order \(\sqrt{\log N} / N\) in the bulk. The sparseness component is captured by the random variable \({\mathcal {Z}}\) and has order \(1/(\sqrt{N} q)\) throughout the spectrum except near the origin. Thus, the sparseness component dominates in the bulk as soon as \(q \ll \sqrt{N}\) and at the edge as soon as \(q \ll N^{1/6}\). In fact, our proof suggests that \({\mathcal {Z}}\) is only the leading order such Gaussian contribution arising from the sparseness, and that there is an infinite hierarchy of strongly correlated and asymptotically Gaussian random variables of which \({\mathcal {Z}}\) is the largest and whose magnitudes decrease in powers of \(q^{-2}\). In order to obtain random matrix Tracy–Widom statistics near the edge, one would have to subtract all of such contributions up order \(N^{-2/3}\). For \(q = N^{\beta }\) with \(\beta \) arbitrarily small, the number of such terms becomes arbitrarily large.
For completeness, we mention that the bulk eigenvalue statistics have also been analysed in terms of their correlation functions and eigenvalue spacings, which have a very different behaviour from the single eigenvalue fluctuations described above. It was proved in [8, 9, 18, 22] that the asymptotics of the local eigenvalue correlation functions in the bulk coincide with those of GOE for any \(q \gg 1\). Thus, the sparseness has no impact on the asymptotic behaviour of the correlation functions and the eigenvalue spacings.
We conclude this section with a few words about the proof. The fluctuations of the extreme eigenvalues are considerably harder to analyse than those of the bulk eigenvalues, and in particular the method of [12] breaks down at the edge because the self-consistent equations on which it relies become unstable. The key difficulty near the edge is to obtain strong rigidity estimates on the locations of the extreme eigenvalues, while no such estimates are needed in the bulk. Indeed, the central step of the proof is Proposition 4.1 below, which provides an upper bound for the fluctuations of the largest eigenvalue of H. This is obtained by showing, for suitable E outside the bulk of the spectrum and \(\eta >0\), that the imaginary part of the Green’s function \(G(E+\mathrm {i}\eta ):=(H-E-\mathrm {i}\eta )^{-1}\) satisfies \( {{\,\mathrm{Im}\,}}{{\,\mathrm{Tr}\,}}G(E+\mathrm {i}\eta ) \ll 1/\eta \). Our basic approach is the self-consistent polynomial method for sparse matrices developed in [19, 24]. Thus, we first obtain a highly precise bound on the self-consistent polynomial P of the Green’s function, which provides a good estimate of \({{\,\mathrm{Tr}\,}}G\) outside the bulk. The key observation in this part is that the cancellation built into P persists also in the derivative of P. Armed with the good estimate of \({{\,\mathrm{Tr}\,}}G\), our second key idea is to estimate the imaginary part of P, which turns out to be much smaller than P itself; from this we deduce strong enough bounds on the imaginary part of G. These two estimates together conclude the proof. We refer to Sect. 3 below for more details of the proof strategy.
The rest of the paper is organized as follows. In Sect. 2 we introduce the notations and previous results that we use in this paper. In Sect. 3 we explain the strategy of the proof. In Sect. 4 we prove Theorem 1.2, assuming key rigidity estimates at the edge (Proposition 4.1) and inside the bulk (Lemma 4.2). In Sect. 5 we give a careful construction of the self-consistent polynomial P of the Green’s function. In Sects. 6–8, we prove Proposition 4.1, by assuming several improved estimates for large classes of polynomials of Green’s functions. In Sect. 9 we prove Lemma 4.2. Finally in Sect. 10 we prove the estimates that we used in Sects. 6–8.
2 Preliminaries
In this section we collect notations and tools that will be used. For the rest of this paper we fix \(\beta \in (0,1/6)\) and define \(\delta \) as in (1.7).
Let M be an \(N \times N\) matrix. We denote \(M^{*n}:=(M^{*})^n\), \(M^{*}_{ij}:=(M^{*})_{ij} = {{\overline{M}} \,}_{ji}\), \(M^n_{ij}:=(M_{ij})^n\), and the normalized trace of M by \({{\underline{M}} \,} :=\frac{1}{N} {{\,\mathrm{Tr}\,}}M\). We denote the Green’s function of H by
Convention
Throughout the paper, the argument of G and of any Stieltjes transform is always denoted by \(z \in {\mathbb {C}}{\setminus } {\mathbb {R}}\), and we often omit it from our notation.
The Stieltjes transform of the eigenvalue density at z is denoted by \({\underline{G}} \,(z)\). For deterministic z we have the differential rule
If h is a real-valued random variable with finite moments of all order, we denote by \({\mathcal {C}}_k(h)\) the kth cumulant of h, i.e.
We state the cumulant expansion formula, whose proof is given in e.g. [16, Appendix A].
Lemma 2.1
(Cumulant expansion). Let \(f:{\mathbb {R}}\rightarrow {\mathbb {C}}\) be a smooth function, and denote by \(f^{(k)}\) its kth derivative. Then, for every fixed \(\ell \in {\mathbb {N}}\), we have
assuming that all expectations in (2.3) exist, where \({\mathcal {R}}_{\ell +1}\) is a remainder term (depending on f and h), such that for any \(t>0\),
The following result gives bounds on the cumulants of the entries of H, whose proof follows from Definition 1.1 and the homogeneity of the cumulants.
Lemma 2.2
For every \(k \in {\mathbb {N}}\) we have
uniformly for all i, j.
We use the following convenient notion of high-probability bound from [7].
Definition 2.3
(Stochastic domination). Let
be two families of random variables, where \(Y^{(N)}(u)\) are nonnegative and \(U^{(N)}\) is a possibly N-dependent parameter set. We say that X is stochastically dominated by Y, uniformly in u, if for all (small) \(\varepsilon >0\) and (large) \(D>0\) we have
for large enough \(N \geqslant N_0(\varepsilon ,D)\). If X is stochastically dominated by Y, uniformly in u, we use the notation \(X \prec Y\), or, equivalently \(X=O_{\prec }(Y)\).
Note that for deterministic X and Y, \(X =O_\prec (Y)\) means \(X= O_{\varepsilon }(N^{\varepsilon }Y)\) for any \(\varepsilon > 0\). Sometimes we say that an event \(\Xi \equiv \Xi ^{(N)}\) holds with very high probability if for all \(D > 0\) we have \({\mathbb {P}}(\Xi ) \geqslant 1 - N^{-D}\) for \(N \geqslant N_0(D)\).
By estimating the moments of \({\mathcal {Z}}\) defined in (1.4) and invoking Chebyshev’s inequality, we find
We have the following elementary result about stochastic domination.
Lemma 2.4
-
(i)
If \(X_1 \prec Y_1\) and \(X_2 \prec Y_2\) then \(X_1 X_2 \prec Y_1 Y_2\).
-
(ii)
Suppose that X is a nonnegative random variable satisfying \(X \leqslant N^C\) and \(X \prec \Phi \) for some deterministic \(\Phi \geqslant N^{-C}\). Then \({\mathbb {E}}X \prec \Phi \).
Fix (a small) \(c>0\) and define the spectral domains
We recall the local semicircle law for Erdős–Rényi graphs from [9].
Proposition 2.5
(Theorem 2.8, [9]). Let H be a sparse matrix defined as in Definition 1.1, and \(m_{\mathrm {sc}}\) be the Stieltjes transform of the semicircle distribution. We have
uniformly in \(z = E+\mathrm {i}\eta \in {{\mathbf {S}}}\).
As a standard consequence of the local law, we have the complete delocalization of eigenvectors.
Lemma 2.6
Let \({{\mathbf {u}}}_1,\ldots ,{{\mathbf {u}}}_N\) be the (\(L^2\)-normalized) eigenvectors of H. We have
uniformly for all \(i,k \in \{1,2,\ldots ,N\}\).
Remark 2.7
Proposition 2.5 was proved in [9] under the additional assumption \({\mathbb {E}} H^2_{ii}=1/N\) for all i. However, the proof is insensitive to the variance of the diagonal entries, and one can easily repeat the steps in [9] under the general assumption \({\mathbb {E}} H_{ii}^2=C_i/N\). A weak local law for H with general variances on the diagonal can also be found in [15].
Lemma 2.8
(Ward identity). We have
for all \(z=E+\mathrm {i}\eta \in {{\mathbf {S}}}\).
The following Lemmas 2.9–2.12 characterize the asymptotic eigenvalues density of H. The proof of the following result is postponed to Sect. 5.
Lemma 2.9
There exists a deterministic polynomial
of degree \(2\lceil {\beta ^{-1}} \rceil \) such that
uniformly for all deterministic \(z \in {{\mathbf {S}}}\). Here \(a_2, a_3, \ldots \) are real, deterministic, and bounded. They depend on the law of H.
Lemma 2.9 states that when x is replaced with \({{\underline{G}} \,}(z)\), the expectation of \(P_0(z,x)\) is very small. This is because of a cancellation built into P, which however holds only in expectation and not with high probability. The following two results are essentially proved in [19, Propositions 2.5–2.6], and we state them without proof. We denote by \({\mathbb {C}}_+\) the complex upper half-plane.
Lemma 2.10
There exists a deterministic algebraic function \(m_0:{\mathbb {C}}_+ \rightarrow {\mathbb {C}}_+\) satisfying \(P_0(z,m_0(z))=0\), such that \(m_0\) is the Stieltjes transform of a deterministic symmetric probability measure \(\varrho _0\). We have \({{\,\mathrm{supp}\,}}\varrho _0=[-L_0,L_0]\), where
Moreover,
and
for all \(z \in {\widetilde{{{{\mathbf {S}}}}}}\), where \(\tau _0 \equiv \tau _0(z):=|E^2-L_0^2|\).
Next, define \(P(z,x):=P_0(z,x)+{\mathcal {Z}} x^2\).
Lemma 2.11
There exists a random algebraic function \(m:{\mathbb {C}}_+ \rightarrow {\mathbb {C}}_+\) satisfying \(P(z,m(z))=0\), such that m is the Stieltjes transform of a random symmetric probability measure \(\varrho \). We have \({{\,\mathrm{supp}\,}}\varrho =[-L,L]\), where
Moreover,
and
for all \(z \in {\widetilde{{{{\mathbf {S}}}}}}\), where \(\tau \equiv \tau (z) :=|E^2-L^2|\).
Let \(\gamma _i\) denote the ith N-quantile of \(\varrho \), i.e.
Similarly, let \(\gamma _{0,i}\) and \(\gamma _{\mathrm {sc},i}\) denote the ith N-quantile of \(\varrho _0\) and the semicircle distribution respectively. We have the following result, whose proof is given in “Appendix A” below.
Lemma 2.12
We have
uniformly for \(i \in \{1,2,\ldots ,N\}\). Here \(\gamma _{0,i}\) is deterministic and satisfies \(\gamma _{0,i}=\gamma _{\mathrm {sc},i}+O(q^{-2})\).
3 Outline of the proof
In this section we describe the strategy of the proof. The foundation of the proof is the method of recursive self-consistent estimates for high moments using the cumulant expansion introduced in [13], building on the previous works [5, 6, 20]. It was first used to study sparse matrices in [24], which also introduced the important idea of estimating moments of a self-consistent polynomial in the trace of the Green’s function. There, the authors derived a precise local law near the edge and obtained the extreme eigenvalue fluctuations for \(p \gg N^{-2/3}\). Subsequently, in [19], by developing the key insight that for \(N^{-7/9} \ll p \ll N^{-2/3}\) the leading fluctuations are fully captured by the random variable \({\mathcal {Z}}\) from (1.4), the authors obtained the extreme eigenvalue fluctuations for \(N^{-7/9}\ll p \ll N^{-2/3}\). In this paper we use the same basic strategy as [19, 24]. As in most results on the extreme eigenvalue statistics, the main difficulty is to establish rigidity bounds for the extreme eigenvalues.
The proof of Theorem 1.2 consists of essentially two separate results: an upper bound on the largest eigenvalue of H (Proposition 4.1 below) and a rigidity estimate in the bulk (Lemma 4.2 below). The latter is a modification of [19, Proposition 2.9], and our main task is to show the former.
We use the random spectral parameter \(z=L_0+{\mathcal {Z}} +w\) introduced in [19], where \(w=\kappa +\mathrm {i}\eta \) is deterministic. In order to obtain the estimate of Proposition 4.1 for the largest eigenvalue of H using the Green’s function, one has to preclude the existence of an eigenvalue near \({{\,\mathrm{Re}\,}}z\) for a suitable z, which follows provided one can show
[see (6.5) and the discussions afterwards for more details]. The proof of (3.1) is the main work of our proof. It relies on the following key new ideas.
-
1.
In the previous works [19, 24], following the work [10] on Wigner matrices, (3.1) is always proved using
$$\begin{aligned} {{\,\mathrm{Im}\,}}{\underline{G}} \,\leqslant {{\,\mathrm{Im}\,}}m + |{\underline{G}} \,-m| \end{aligned}$$and estimating the two terms on right-hand side separately. There, the term \(|{\underline{G}} \,-m|\) is estimated by obtaining an estimate on \(|P(z, {{\underline{G}} \,}) |\) from which an estimate on \(|{\underline{G}} \,-m|\) follows by inverting a self-consistent equation associated with the polynomial P. In our current setting, \(|{{\underline{G}} \,} - m |\) turns out to be much larger than \({{\,\mathrm{Im}\,}}{\underline{G}} \,\) and hence this approach does not work. Thus, we have to estimate \(|{{\,\mathrm{Im}\,}}({{\underline{G}} \,} - m)|\) instead of \(|{\underline{G}} \,-m|\) and take advantage of the fact that it is much smaller than \(|{\underline{G}} \,-m|\). To that end, we first estimate \(|{{\,\mathrm{Im}\,}}P(z, {{\underline{G}} \,}) |\) by exploiting a crucial cancellation arising from taking the imaginary part, which yields stronger bounds on \(|{{\,\mathrm{Im}\,}}P(z,{{\underline{G}} \,}) |\) than are possible for \(|P(z, {{\underline{G}} \,}) |\).
-
2.
To estimate \(|{{\,\mathrm{Im}\,}}({{\underline{G}} \,} - m) |\) from \(|{{\,\mathrm{Im}\,}}P(z, {{\underline{G}} \,}) |\), we have to invert a self-consistent equation associated with \({{\,\mathrm{Im}\,}}P\). This equation is only stable provided that \(|{{\underline{G}} \,} - m |\) is small enough.
-
3.
The main work is to derive a strong enough bound on \(|{{\underline{G}} \,} - m |\) to ensure the stability of the self-consistent equation for \({{\,\mathrm{Im}\,}}({{\underline{G}} \,} - m)\). The precision required for this step is much higher than that obtained in [19]. Our starting point is the same as in [19, 24]: estimating high moments \({\mathbb {E}}|P |^{2n}\) of \(P \equiv P(z,{\underline{G}} \,)\) using the cumulant expansion. Note that P is constructed in such a way that the expectation \({\mathbb {E}}P(z, {{\underline{G}} \,})\) is very small by a near-exact cancellation (see Lemma 2.9). In the high moments, the interactions between different factors of P and \({{\overline{P}} \,}\), corresponding to the fluctuations of P, give rise to error terms whose control is the key difficulty of the proof. They cannot be estimated naively and have to be re-expanded to arbitrarily high order using a recursive application of the cumulant expansion. These error terms typically contain the partial derivative \(\partial _2 P\) of P in the second argument \({{\underline{G}} \,}\). As soon as P is differentiated, the cancellation built into P is lost. However, we nevertheless need to exploit remnants of this cancellation that are inherited by these higher-order terms containing derivatives of P. We track them by rewriting the partial derivative \(\partial _2 P\) in terms of the derivative \(\partial _w P = \partial _1 P + \partial _2 P \partial _w {{\underline{G}} \,}\) and an error term, and then use that \(\partial _w\) commutes with the derivative \(\frac{\partial }{\partial H_{ij}}\) from the cumulant expansion to obtain a form where the cancellation from the next cumulant expansion is obvious also for the derivative of P.
Let us explain the above points in more detail. The proof of (3.1) contains two steps. The main step is to bound the high moments of P in Proposition 6.1. We start with
We expand the first term on the right-hand side by Lemma 2.1 to get
Note that the polynomial P is designed such that
and for the same reason there are cancellations between the second and third terms on right-hand side of (3.2). It turns out that the most dangerous terms on right-hand side of (3.2) are contained within the first sum. One representative error term, arising from \(k = 3\) and \(s = 2\) in (3.2), is
which involves the interaction of P and \({{\overline{P}} \,}\) and hence depends on the fluctuations of P.
To get a sharp enough estimate of (3.3), it is not enough to take absolute value inside the expectation and then estimate \(|\partial _2 {{\overline{P}} \,}|\) and \(|N^{-1}(G^{*2})_{jj}|\) by Lemmas 2.11 and 2.8 respectively. Instead, the key idea is to rewrite the error term, so that it becomes amenable to another expansion step, asFootnote 1
which comes from the approximations
which of course have to be justified. Ignoring the error terms generated in this process, we find that (3.3) is reduced to
Since \(\partial _{{{\overline{w}} \,}}\) and \(\partial /\partial H_{ij}\) commute, we can again expand the first term on the right-hand side with Lemma 2.1. In this way the operator \(\partial _{{{\overline{w}} \,}}\) plays no role in our computation, and we can get the desired estimate using the smallness of \({\mathbb {E}} {{\overline{P}} \,}\).
A major difficulty in the above argument results from the fact that we need to track carefully the algebraic structure of the error terms arising from repeated applications of simplifications of the form (3.4). In particular, such terms occur inside expectations multiplying lots of other terms, and we need to ensure that such approximations remain valid in general expressions. In order to achieve this, we implement the ideas in [12, 14] to construct a hierarchy of Schwinger–Dyson equations for a sufficiently large class of polynomials in the entries of the Green’s function.
A desired bound for P, Proposition 6.1, together with the stability analysis of the self-consistent equation associated with P (Lemma 6.2 below), yields the key estimate
where we recall that \({{\,\mathrm{Re}\,}}z = L_0 + {\mathcal {Z}} + \kappa \). This estimate is crucial in establishing the stability of the self-consistent equation associated with \({{\,\mathrm{Im}\,}}P\) (see Lemma 6.4). More precisely, a Taylor expansion shows
As \(\partial _2^2 P(z,m)\approx 2\), taking the imaginary part and rearranging terms yields
It can be showed that \(|{{\,\mathrm{Re}\,}}\partial _2 P(z,m) | \asymp \sqrt{\kappa }\), and we move this factor to the right-hand side of (3.6) to obtain a recursive estimate of \({{\,\mathrm{Im}\,}}({\underline{G}} \,-m)\). The third term on right-hand side of (3.6) says that in order for this estimate to work, we need
which is exactly (3.5).
The final step in showing (3.2) is to bound the high moments of \({{\,\mathrm{Im}\,}}P\) in Proposition 6.3. As \({{\,\mathrm{Im}\,}}P\) is much smaller than P near the edge, compared to \({\mathbb {E}} |P|^{2n}\), we obtain a much smaller bound for \({\mathbb {E}} |{{\,\mathrm{Im}\,}}P |^{2n}\). The proof is similar to that of Proposition 6.1, but contains significantly fewer expansions. Combining Proposition 6.3 and Lemma 6.4 leads to our desired estimate of \({{\,\mathrm{Im}\,}}G\), which is
As we prove the above for z satisfying \({{\,\mathrm{Im}\,}}m \ll \frac{1}{N\eta }\), we get (3.1) as desired.
4 Proof of Theorem 1.2
In this section we prove Theorem 1.2. The key result is the following upper bound on the largest eigenvalue of H. The proof is postponed to Sect. 6.
Proposition 4.1
Denoting by \(\mu _N\) the largest eigenvalue of H, we have
We also need the following result to estimate the eigenvalues away from the spectral edges. The proof is postponed to Sect. 9.
Lemma 4.2
Let \(\rho \) denote the empirical eigenvalue density of H, and set
We have
for all \(I \subset I_1\) and \(I=I_2\).
Proof of Theorem 1.2
We prove (1.8) for \(i \in \{\lfloor {N/2} \rfloor -1,\ldots ,N-1\}\), and the same analysis works for the other half of the spectrum. Let \(i \in \{\lfloor {N/2} \rfloor -1,\ldots ,N-1\}\) and suppose first that
Then trivially we have \(\gamma _i \in I_2\) with very high probability. In addition, by the Cauchy interlacing theorem we have \(\lambda _i \leqslant \mu _N\), and together with Proposition 4.1 we obtain
Thus by the triangle inequality we get
Next, suppose (4.3) does not hold, namely
for some \(a\in (0,3)\). Let \(\nu \) be the empirical eigenvalue density of A. By the Cauchy interlacing theorem,
for any \(I \subset {\mathbb {R}}\). Together with (4.2), we have
for all \(I \subset I_1\) or \(I=I_2\). Let \(f(E):=\varrho ([E,\infty ))\). Then
where in the last step we used (4.5). By the definition of \(I_2\) we get \( |f(\gamma _i)-f(\lambda _{i})| \prec N^{-\delta }(|I_2|+a)^{3/2}.\) Together with the uniform square root behaviour of the density of \(\varrho \) near L from Lemma 2.11 we therefore have
with very high probability, where \(c > 0\) is a constant. Thus
with very high probability. Since \(f(x)\asymp (L-x)^{3/2}\) for \(x \in I_1\), we deduce that \(L-\gamma _i\asymp L-\lambda _{i}\) with very high probability. Moreover, by Lemma 2.11 we have \(f'(x)\asymp (L-x)^{1/2}\) for \(x \in I_1\), which implies \(f'(\lambda _i)\asymp f'(\gamma _i)\) with very high probability, and hence that \(f'(x)\asymp f'(\gamma _i)\) with very high probability for any x between \(\lambda _{i}\) and \(\gamma _i\). Thus the mean value theorem yields
Using the above relation, together with (4.4) and Lemma 2.12, we conclude that
We then take the expectation using Lemma 2.4, which yields
Combining the above two formulas we have (1.8) as desired. \(\square \)
5 Abstract polynomials and the construction of \(P_0\)
Convention
Throughout this section, \(z \in {{\mathbf {S}}}\) is deterministic.
In this section we construct the polynomial \(P_0\) and prove Lemma 2.9. It was essentially proved in [19, Proposition 2.9]; here we follow a more systematic approach, based on a class of abstract polynomials in the Green’s function entries, which provides an explicit proof. We shall generalize this class further in Sect. 7.
5.1 Abstract polynomials, part I
We start by introducing a notion of formal monomials in a set of formal variables, which are used to construct \(P_0\). Here the word formal refers to the fact that these definitions are purely algebraic and we do not assign any values to variables or monomials.
Definition 5.1
Let \(\{i_1,i_2,\ldots \}\) be an infinite set of formal indices. To \(\sigma , \nu _1 \in {\mathbb {N}}\), \(\theta \in {\mathbb {R}}\), \(x_1, y_1, \ldots , x_\sigma , y_\sigma \in \{i_1, \ldots , i_{\nu _1}\}\), and a family \((a_{i_1,\ldots ,i_{\nu _1}})_{1\leqslant i_1,\ldots ,i_{\nu _1}\leqslant N}\) of uniformly bounded complex numbers we assign a formal monomial
We denote \(\sigma (T) = \sigma \), \(\nu _1(T) = \nu _1\), \(\theta (T) = \theta \), and \(\nu _2(T) :=\sum _{k = 1}^\sigma {{\mathbf {1}}}_{x_k \ne y_k}\). Thus, \(\sigma (T)\) is the degree of T and \(\nu _2(T)\) is the number of off-diagonal Gs. We denote by \({\mathcal {T}}\) the set of formal monomials T of the form (5.1).
Definition 5.2
We assign to each monomial \(T \in {\mathcal {T}}\) with \(\nu _1 = \nu _1(T)\) its evaluation
which is a random variable depending on an \(\nu _1\)-tuple \((i_1,\ldots ,i_{\nu _1})\in \{1,2,\ldots ,N\}^{\nu _1}\). It is obtained by replacing, in the formal monomial T, the formal indices \(i_1,\ldots ,i_{\nu _1}\) with the integers \(i_1,\ldots ,i_{\nu _1}\) and the formal variables \(G_{xy}\) with elements \(G_{xy}\) of the Green’s function (2.1) with parameter z. We define
Defining the random variable
we have the following result, whose proof is given in Sect. 10.1 below.
Lemma 5.3
For any fixed \(T \in {\mathcal {T}}\) we have
Remark 5.4
When \(\nu _2(T)\ne 1\), Lemma 5.3 is a straightforward consequence of Lemma 2.8 and Proposition 2.5. When \(\nu _2(T)=1\), naively applying the Ward identity shows
In this case, therefore, Lemma 5.3 extracts an additional factor of \(\sqrt{\Gamma }\).
In the sequel we also need the subset
of formal monomials without off-diagonal entries. We define an averaging map \({\mathcal {M}}\) from \({\mathcal {T}}_0\) to the space of random variables through
for \(T =a_{i_1,\ldots ,i_{\nu _1}}N^{-\theta }G_{x_1x_1}G_{x_2x_2}\ldots G_{x_{\sigma }x_{\sigma }} \in {\mathcal {T}}_0\). The interpretation of \({\mathcal {M}}(T)\) is that it replaces all diagonal entries of G in T by their average \({{\underline{G}} \,}\) and then applies \({\mathcal {S}}\). Note that it is only applied to monomials \(T \in {\mathcal {T}}_0\) without off-diagonal entries. The following result is proved in Sect. 10.2 below.
Lemma 5.5
For any fixed \(T \in {\mathcal {T}}_0\) there exists \(k \in {\mathbb {N}}\) and \(T^{(1)},\ldots ,T^{(k)} \in {\mathcal {T}}_0\) such that
Each \(T^{(l)}\) satisfies \(\sigma (T^{(l)})-\sigma (T) \in 2{\mathbb {N}}+4\),
Lemma 5.5 leads to the following result.
Lemma 5.6
Fix \(T \in {\mathcal {T}}_0\). Fix \(r \in {\mathbb {N}}_+\). Then there exists deterministic and bounded \(b_1,\ldots ,b_{r}\) such that
satisfies
Proof
When \(r=1\), the Lemma is trivially true from Lemma 5.5. When \(r \geqslant 2\), the proof is essentially a repeated use of Lemma 5.5. More precisely, by Lemma 5.5,
for some fixed \(k \in {\mathbb {N}}\), where each \(T^{(l)}\) satisfies \(\sigma (T^{(l)})-\sigma (T) \in 2{\mathbb {N}}+4\), \(\nu _1(T^{(l)})=\nu _1(T)+1\) and \(\theta (T^{(l)})=\theta (T)+1+\beta (\sigma (T^{(l)})-\sigma (T)-2)\). As a result, \({\mathbb {E}} {\mathcal {S}}(T^{(l)})=O_{\prec }\big (N^{\nu _1(T)-\theta (T)-2\beta }\big )\) for each l. Now we apply Lemma 5.5 to each \(T^{(l)}\) on RHS of (5.8), and get
for some fixed \(k_l \in {\mathbb {N}}\), where each \(T^{(l,l_1)}\) satisfies \(\sigma (T^{(l,l_1)})-\sigma (T) \in 2{\mathbb {N}}+8\), \(\nu _1(T^{(l,l_1)})=\nu _1(T)+2\) and \(\theta (T^{(l)})=\theta (T)+2+\beta (\sigma (T^{(l,l_1)})-\sigma (T)-4)\). Moreover, by our conditions on \(\theta (T^{(l)})\), \(\nu _1(T^{(l)})\) and \(\theta (T^{(l)})\), we can write
for some deterministic and bounded \(b_{l,1},\ldots ,b_{r,1}\). Combining (5.8)–(5.10), we have
Note that \({\mathbb {E}} {\mathcal {S}}(T^{(l,l_1)})=O_{\prec }\big (N^{\nu _1(T)-\theta (T)-4\beta }\big )\) for each \((l,l_1)\). Thus we can again apply Lemma 5.5 to each \(T^{(l,l_1)}\) on RHS of (5.11). Repeating the above steps finitely many times completes the proof. \(\square \)
Note that we in particular have \({\mathcal {M}}(1,T)={\mathcal {M}}(T)\) through (5.7).
5.2 The construction of \(P_0\) and Proof of Lemma 2.9
We compute
and we shall find a polynomial \(Q_0\) such that
We then set \(P_0(z,x)=1+zx+Q_0(x)\). Using Lemma 2.1 with \(h=H_{ij}\) and \(f=f_{ji}(H)=G_{ji}\), we have
where \(\ell \) is a fixed positive integer to be chosen later, and \({\mathcal {R}}^{(ji)}_{\ell +1}\) is a remainder term defined analogously to \({\mathcal {R}}_{\ell +1}\) in (2.4). One can follow, e.g. the proof of Lemma 3.4 (iii) in [16], and readily check that
for \(\ell \equiv \ell (\beta ) \) large enough. From now on, we always assume the remainder term in cumulant expansion is negligible.
Now let us look at each \({\widetilde{X}}_k\). For \(k=1\), by the differential rule (2.2) and \({\mathcal {C}}_{2}(H_{ij})=1/N\) for \(i \ne j\), we have
For \(k=2\), the most dangerous term is
and by \({\mathcal {C}}_3(H_{ij})=O(N^{-1-\beta })\), we see that \(\nu _1(T) = 2\), \(\theta (T) = 2 + \beta \), and \(\nu _2(T)=1\). Thus by Lemma 5.3 we have
Other terms in \({\widetilde{X}}_2\) also satisfy the same bound. Similar estimates can also be done for all even k, which yield
For odd \(k \geqslant 3\), we split
where terms in \({\widetilde{X}}_{k,1}\) contain no off-diagonal entries of G. Use Lemma 5.3, we easily find
By Lemma 2.2, we see that
where \(a^{(k)}_{ij}\) is deterministic and uniformly bounded. Combining with (5.12)–(5.14), we have
where
To handle the right-hand side of (5.15) we invoke Lemma 5.6. Naively, we have
for each n. By Lemma 5.6, we can write
Thus (5.15) becomes
Thus we can set
and note that \(Q_0\) is a polynomial of degree \(2\lceil {\beta ^{-1}} \rceil \). This concludes the proof of Lemma 2.9.
Remark 5.7
After the construction of \(P_0\) (and consequently P), we shall construct a more general class of abstract polynomials associated with P in Sect. 7.2 below.
6 Proof of Proposition 4.1
Convention
Throughout this section,
where \(w=\kappa +\mathrm {i}\eta \) deterministic.
The proof of Proposition 4.1 consists of two steps; in the first we first estimate \({\underline{G}} \,\) and in the second we apply this estimate to obtain a more precise bound of \({{\,\mathrm{Im}\,}}{\underline{G}} \,\).
6.1 Estimate of \({\underline{G}} \,\)
Define the spectral domain
As a guide to the reader, the lower bound on \(\kappa \) is chosen to be slightly smaller than the scale \(\frac{1}{\sqrt{N}q}\) on which the extreme eigenvalues fluctuate; analogously, the lower bound on \(\eta \) is chose to be slightly smaller than the scale \(N^{-5/8} q^{-1/4}\), which is the solution of the equation \(\frac{\eta }{\sqrt{\kappa }} = \frac{1}{N \eta }\) with \(\kappa = \frac{1}{\sqrt{N}q}\). Using that \({{\,\mathrm{Im}\,}}m(z) \asymp \frac{\eta }{\sqrt{\kappa }}\) (see Lemma 2.11), this choice of lower bound on \(\eta \) will allow us to rule out the presence of eigenvalues [see (6.6) below], and hence establish rigidity.
Recall the definition of \(\tau \) in Lemma 2.11, and note that the lower bound on \(\kappa \) ensures, with very high probability,
for all \(w \in {{\mathbf {Y}}}\). The main technical step is the following bound for \(P(z,{\underline{G}} \,)\), whose proof is postponed to Sect. 7.
Proposition 6.1
Let \(w \in {{\mathbf {Y}}}\). Suppose \(|{\underline{G}} \,-m| \prec \Psi \) for some deterministic \(\Psi \in [\sqrt{\kappa }N^{-\delta },1]\). Then
Lemma 6.2
Suppose \(\varepsilon :{{\mathbf {Y}}}\rightarrow [N^{-1},N^{-\delta }]\) is a function so that
for all \(w \in {{\mathbf {Y}}}\). Suppose \(\varepsilon (w)\) is Lipschitz continuous with Lipschitz constant N and moreover that for each fixed \(\kappa \) the function \(\eta \rightarrow \varepsilon (\kappa +\mathrm {i}\eta )\) is nonincreasing for \(\eta >0\). Then
Proof
See [19, Proposition 2.11]. \(\square \)
Combining Proposition 6.1 and Lemma 6.2, we find that for any deterministic \(\Psi \) that does not depend on \(\eta \) we obtain the implication
Using the initial estimate \(|{\underline{G}} \,-m| \prec 1\) from Proposition 2.5, we therefore conclude the key bound
6.2 Estimate of \({{\,\mathrm{Im}\,}}{\underline{G}} \,\)
Define the subset
In this section we show that
for all \(w \in {{\mathbf {Y}}}_*\). This immediately implies that whenever \(\kappa +\mathrm {i}\eta \in {{\mathbf {Y}}}_*\), with very high probability there is no eigenvalue in the interval \((L_0+{\mathcal {Z}}+\kappa -\eta ,L_0+{\mathcal {Z}}+\kappa +\eta )\). In addition, [9, Lemma 4.4] implies
and hence the largest eigenvalue \(\mu _N\) of H satisfies (4.1), and Proposition 4.1 is proved.
What remains, therefore, is the proof of (6.6). In analogy to Proposition 6.1, we have the following estimate for \( {{\,\mathrm{Im}\,}}P(z, {\underline{G}} \,)\), whose proof is postponed to Sect. 8.
Proposition 6.3
Let \(w \in {{\mathbf {Y}}}_*\). Suppose \(| {{\,\mathrm{Im}\,}}{\underline{G}} \,- {{\,\mathrm{Im}\,}}{m}| \prec \Phi \) for some deterministic \(\Phi \equiv \Phi \in [N^{-1-\delta }\eta ^{-1},1]\). Then
Lemma 6.4
Let \(w \in {{\mathbf {Y}}}_*\). Suppose that
for some deterministic \(\varepsilon \in [N^{-1},N^{-\delta }]\). Then
Proof
A Taylor expansion gives
Note that \(\partial ^k_2 P(z,m) \prec 1\), and, recalling the definition of P, we find from (2.5) and by Lemma 2.11 that
for all \(k \geqslant 1\), where the last inequality holds for any \(w \in Y_*\) we have
This implies, for all \(k \geqslant 2\),
where in the second step we used (6.4). Taking imaginary part of (6.7) and rearranging the terms, we have
Note that \(|{{\,\mathrm{Im}\,}}\partial _2 P(z,m)| \asymp |{{\,\mathrm{Im}\,}}m| \ll \sqrt{\kappa }\), and by (2.7) we have \(|\partial _2 P(z,m)|\asymp \sqrt{\kappa }\). Thus
and together with (6.4) and (6.8) we have
This yields the claim. \(\square \)
From Proposition 6.3 and Lemma 6.4 we obtain the implication
Iterating (6.11) \(O(1/\delta )\) times and recalling Definition 2.3 yields
for all \(w \in {{\mathbf {Y}}}_*\). Since
we thus conclude (6.6). This concludes the proof of Proposition 4.1.
7 Proof of Proposition 6.1
Convention
Throughout this section, z is given by (6.1), where \(w \in {{\mathbf {Y}}}\) is deterministic.
Fix \(n \in {\mathbb {N}}_+\) and set
We shall show, for any fixed \(n \in {\mathbb {N}}\), that
from which Proposition 6.1 follows by Chebyshev’s inequality. The rest of this section is therefore devoted to the proof of (7.1).
Set
and abbreviate
Note that argument z of G is random, and
and as a result
We define the parameter
Recalling the random variable \(\Gamma \) from (5.3), we find
Moreover, we have
The next lemma collects basic estimates for the derivatives of P.
Lemma 7.1
Under the assumptions of Proposition 6.1, for any fixed \(k \in {\mathbb {N}}_+\) we have
and
Proof
By the mean value theorem,
for some \(\xi \) between m and \({\underline{G}} \,\). Then the first estimate in (7.8) is proved using Lemma 2.11 and (6.3). The second estimate in (7.8) is proved by Lemmas 2.6 and 2.8. By (7.4) and (7.8), one easily checks that
and combing with (7.7) one concludes (7.9). \(\square \)
7.1 The first expansion
By \((H-z)G=I\), we have
We use Lemma 2.1 to calculate the last term. By setting \(h=H_{ij}\), \(f=f_{ji}(H)=G_{ji}P^{n-1}P^{*n}\), we get
where, as in (5.12), we choose a large enough \(\ell \in {\mathbb {N}}_+\) such that the remainder term is is negligible. By splitting the differentials in (7.10) basing on if \(P,{{\overline{P}} \,}\) are differentiated, we have
We have the following result, which handles the terms on right-hand side of (7.11) and directly implies (7.1).
Lemma 7.2
Let (I–IV) be as in (7.11). We have
as well as
The rest of Sect. 7 is devoted to showing Lemma 7.2. To simplify notation, we drop the complex conjugates in (I–IV) (which play no role in the subsequent analysis), and estimate the quantities
and
7.2 Abstract polynomials, part II
In order to estimate (7.14) and (7.15), we introduce the following class of abstract polynomials, which generalizes the class \({\mathcal {T}}\) from Definition 5.1.
Definition 7.3
Let \(\{i_1,i_2,\ldots \}\) be an infinite set of formal indices. To integers \(s,k, \nu _1,\nu _3 \in {\mathbb {N}}\), digits \(\nu _4,\nu _5 \in \{0,1\}\) satisfying \(\nu _4 \leqslant \nu _5\), a real number \(\theta \in {\mathbb {R}}\), formal indices \(x,y,x_1, y_1, \ldots , x_\sigma , y_\sigma \in \{i_1, \ldots , i_{\nu _1}\}\), and a family \((a_{i_1,\ldots ,i_{\nu _1}})_{1\leqslant i_1,\ldots ,i_{\nu _1}\leqslant N}\) of uniformly bounded complex numbers we assign a formal monomial
We denote \(\sigma (V) = s+k\), \(\nu _i(V) = \nu _i\) for \(i = 1,3,4,5\), \(\theta (V) = \theta \), and
We denote by \({\mathcal {V}}\) the set of formal monomials V of the form (7.16).
We extend the evaluation from Definition 5.2 to the set \({\mathcal {V}}\), and denote the evaluation of V as in (7.16) by \(V_{i_1, \ldots , i_{\nu _1}}\). We also extend the operation \({\mathcal {S}}\) from (5.2) to \({\mathcal {V}}\).
The next lemma is an analogue of Lemma 5.3, whose proof is postponed to Sect. 10.3.
Lemma 7.4
Let \(V \in {\mathcal {V}}\) and abbreviate \(\nu _i = \nu _i(V)\) and \(\theta = \theta (V)\). Suppose that \(\nu _2 \ne 0\).
-
(i)
We have
$$\begin{aligned} {\mathbb {E}} {\mathcal {S}} (V)&\prec N^{\nu _1-\theta }(\Psi +\sqrt{\kappa +\eta })^{\nu _4} (N\eta )^{-\nu _5}\Upsilon {\mathbb {E}} |P^{\nu _3}| \\&\quad +\sum _{t=1}^{\nu _3} N^{\nu _1-\theta }(\Psi +\sqrt{\kappa +\eta })^{\nu _4} \Upsilon ^{\nu _5}((\Psi +\sqrt{\kappa +\eta }) \Upsilon )^{t}{\mathbb {E}} |P^{\nu _3-t}|. \end{aligned}$$ -
(ii)
Moreover, when \(\nu _4(V)=\nu _5(V)=0\), we have the stronger estimate
$$\begin{aligned} {\mathbb {E}} {\mathcal {S}} (V)\prec & {} N^{\nu _1-\theta }\Upsilon {\mathbb {E}} |P^{\nu _3}|+ N^{\nu _1-\theta }\Upsilon ^{2}{\mathbb {E}} |P^{\nu _3-1}| \\&+\sum _{t=2}^{\nu _3} N^{\nu _1-\theta }((\Psi +\sqrt{\kappa +\eta }) \Upsilon )^{t}{\mathbb {E}} |P^{\nu _3-t}|. \end{aligned}$$
In the sequel, we also need the subset
In analogy to (5.5), we define an averaging map \({\mathcal {M}}\) from \({\mathcal {V}}_0\) to the space of random variables through
for
The following is an analogue of Lemma 5.5, whose proof is given in Sect. 10.4.
Lemma 7.5
Let \(V \in {\mathcal {V}}_0\). There exist \(V^{(1)},\ldots ,V^{(k)} \in {\mathcal {V}}_0\) such that, abbreviating \(\nu _i = \nu _i(V)\) and \(\theta = \theta (V)\),
where k is fixed, and each \(V^{(l)}\) satisfies \( V^{(l)} \in {\mathcal {V}}_0\), \(\sigma (V^{(l)})- \sigma (V) \in 2{\mathbb {N}}+4\), \(\nu _i(V^{(l)})=\nu _i(V)\) for \(i=2,3,4,5\),
As a result, each \(V^{(l)}\) satisfies
Repeatedly using Lemma 7.5, and together with (7.6), we obtain the following result.
Lemma 7.6
Let \(V \in {\mathcal {V}}_0\) and abbreviate \(\nu _i = \nu _i(V)\), \(\theta = \theta (V)\), and \(\sigma =\sigma (V)\). Then there exist deterministic uniformly bounded \(b_1,\ldots ,b_{\lceil {\beta ^{-1}} \rceil }\) such that
satisfies
Finally, we have the following extension of Lemma 5.6, which is proved in Sect. 10.5.
Lemma 7.7
Fix \(r,u,v\in {\mathbb {N}}\). Let \(T \in {\mathcal {T}}_0\) and let \({\mathcal {M}}(r,T)\) be as in Lemma 5.6. Then
7.3 The computation of (IV’) in (7.14)
We write \(\text {(IV')}=\sum _{k=1}^lX_k\), where
7.3.1 The estimate of \(X_1\)
By (7.4) and \({\mathcal {C}}_2(H_{ij})=N^{-1}(1+O(\delta _{ij}))\), we have
Estimating the last term using Lemma 7.1, we conclude
By \(HG=zG+I\) and \(z \prec 1\), we deduce that \({\underline{HG}} \, \prec 1\). In addition, it is easy to check that \({\underline{G^3}} \, \prec \Upsilon (N\eta )^{-1}\). Thus the first term on right-hand side of (7.19) can be estimated by
As a result,
where in the second step we used Hölder’s inequality. Since
for all \(w \in {{\mathbf {Y}}}\), we have \(X_1 \prec {\mathcal {E}}^2 {\mathcal {P}}^{2n-2}\) as desired.
7.3.2 The estimate of \(X_2\)
Let us split
Since \({\mathcal {C}}_3(H_{ij})=O(N^{-1-\beta })\), we have
where in the second and third step we used Lemmas 2.8 and 7.1 respectively. Note that
and similarly
Thus by (7.20) we get
as desired. As for the term \(X_{2,2}\), we see from (7.3) that the most dangerous term is
where \(a_{ij}\) is deterministic and uniformly bounded. Note that we write
where \(\nu _1(V)=2,\theta (V)=2+\beta ,\nu _3(V)=2n-2,\nu _4(V)=\nu _5(V)=1\). By Lemma 7.4 (i), we have
One can easily check
for all \(t\geqslant 1\). Thus we have
For \(X_{2,2,2}\), we can again apply Lemma 2.1 for \(h=H_{ij}\) and get
Note that
for all fixed \(s\geqslant 0\). Together with (7.9) and the trivial bound \(N^{-1} \prec {\mathcal {E}}\), we see that
where in the last step we also used \(\Upsilon \prec {\mathcal {E}}\). Similar steps also work for \(X_{2,2,3}\). As a result, we have \((7.23) \prec \sum _{r=2}^{2n}{\mathcal {E}}^r {\mathcal {P}}^{2n-r}\). Other terms in \(X_{2,2}\) can be estimated in a similar fashion, which leads to
Combining with (7.22) we get
as desired.
7.3.3 The computation of \(X_3\)
Let us split
where
for \(s=1,2,3\).
Step 1 When \(s=1,3\), it is easy to see from (7.3) that
Using (7.9), we can deduce
Thus
where in the second step we used Lemma 2.8. As in (7.21) and (7.24), we have
for all \( t \geqslant 0\). Thus \(X_{3,s} \prec \sum _{r=2}^{2n} {\mathcal {E}}^r {\mathcal {P}}^{2n-r}\) for \(s=1,3\).
Step 2 Let us consider
Similar as in the previous steps, we can show that
By Lemma 7.1 and (7.4), we have
and
Together with Lemma 2.8 we get
As the last two terms can be estimated by \(O_{\prec }({\mathcal {E}}^2{\mathcal {P}}^{2n-2})\), we have
Step 3 Let us compute \(X_{3,2,1}\). We write
Note that \(V \in {\mathcal {V}}_0\) with \(\nu _1(V)=2,\theta (V)=2+2\beta ,\nu _3(V)=2n-2\) and \(\nu _4(V)=\nu _5(V)=1\). By Lemma 7.6 we have
where
and \(b_2,\ldots ,b_{\lceil {\beta ^{-1}} \rceil }\) are bounded. We can estimate the right-hand side of (7.28) by \(\sum _{r=2}^{2n} O_{\prec }({\mathcal {E}}^{r} {\mathcal {P}}^{2n-2})\), so that
Step 4 Let us consider the term \({\mathbb {E}} {\mathcal {M}}(V)\) in (7.29). Explicitly,
where
is bounded by Lemma 2.2. Since
we have
In addition,
Thus,
Step 5 We expand the term (A) again by Lemma 2.1, and get
By Lemma 7.1, whenever the derivative \(\partial ^k/\partial H_{ij}^k\) on the right-hand side hits \({\underline{G}} \,^3P^{2n-2}\), the corresponding term can be bounded by \(O_{\prec }\big (\sum _{r=2}^{2n} {\mathcal {E}}^r {\mathcal {P}}^{2n-r}\big )\). Furthermore, since \(\partial ^k/\partial H_{ij}^k\) commutes with \(\partial _{w}\),
The analysis of \(Y_k\) is similar to that of \({\widetilde{X}}_k\) in Sect. 5.2. For \(k=1\), by (7.3), (7.31) and Lemma 7.4 (i), we have
For \(k=2\), by (7.3) and (7.31) we see that the most dangerous term is
Since \({\mathcal {C}}_3(H_{ij})=O(N^{-1-\beta })\), we see that \(\nu _1({\widetilde{V}}) = 2\), \(\nu _2({\widetilde{V}})=1\), \(\nu _4({\widetilde{V}})=0\), \(\nu _5({\widetilde{V}})=1\), and \(\theta ({\widetilde{V}}) = 2 + 3 \beta \). Thus by, Lemma 7.4 (i),
where in the last step we again used \(w \in {{\mathbf {Y}}}\) and Hölder’s inequality. Other terms in \(Y_2\) also satisfy the same bound. A similar estimate can also be obtained for all even k, which yields
For odd \(k \geqslant 3\), we split
where by definition terms in \(Y_{k,1}\) contain no off-diagonal entries of G or \(G^2\). Use Lemma 7.4 (i), we can again show that
By Lemma 2.2, we see that
where \(a^{(k)}_{ij}\) is deterministic and uniformly bounded. Combining with (7.33)–(7.34), we obtain
where
By Lemma 7.7, we have
Since \(s \geqslant 2\), one readily checks that the last two terms can be bounded by \(O_{\prec }(\sum _{r=2}^{2n} {\mathcal {E}}^r {\mathcal {P}}^{2n-r})\). Thus (7.35) reads
Note that by construction, \(T^{(s)}\) in (7.36) is the same as in (5.16). From Lemma 7.7, we see that the term \({\mathcal {M}}(\lceil {\beta ^{-1}-2s+2} \rceil ,T^{(s)})\) in (7.37) is the same as in (5.18), which implies
Thus (7.37) reduces to
Final Step By (7.32) and (7.38), we see that there is a cancellation between \(\mathrm {(A)}\) and \(\mathrm {(B)}\), which leads to
The first two terms on right-hand side of (7.39) are stochastically dominated by
and one can check that \(\Upsilon /(N\eta )\gg {\mathcal {E}}^2\) and \(N^{-1-2\beta } \gg {\mathcal {E}}^2\), so that we need to keep track of these terms in order to obtain a further cancellation.
So far we have been dealing with \({\mathbb {E}} {\mathcal {M}}(V)\) in (7.29), and other terms in (7.29) can be handled in the same way as in Steps 4 and 5. Compared to \({\mathbb {E}}{\mathcal {M}}(V)\), each \(N^{-2\beta }b_{l} N^{-l\beta } {\mathbb {E}}P'N^{-1}{\underline{G^2}} \,\,{\underline{G}} \,^{2+2l} P^{2n-2}\) contains an additional factor \(N^{-l\beta }\). Similarly to (7.39) and (7.40), it can be shown that
for all \(l \geqslant 2\). As a result, we have
where \(b_1\) is defined as in (7.30).
Next, we consider the other terms on right-hand side of (7.26). Similarly to (7.41), we can also show that
as well as
Note that this results in two cancellations on right-hand side of (7.26), and we have
As we have already estimated \(X_{3,1}\) and \(X_{3,3}\) in Step 1, we conclude that
Remark 7.8
The crucial step in analysing \(X_3\) is the computation of \(X_{3,2,1}\) in (7.41). As in (7.27), we can write \(X_{3,2,1}={\mathbb {E}}{\mathcal {S}}(V)\), with \(V\in {\mathcal {V}}_0\), \(\nu _1(V)-\theta (V)=-2\beta \), \(\nu _3(V)=2n-2\), and \(\nu _4(V)=\nu _5(V)=1\). Since
the formula (7.41) implies the estimate
The argument for \(X_{3,2,1}\) can be repeated for general \({\mathbb {E}} {\mathcal {S}}(V)\), which allows one to show the following result.
Lemma 7.9
Let \(V\in {\mathcal {V}}_0\), with \(\nu _1(V)-\theta (V)\leqslant -2\beta \), \(\nu _3(V)=2n-2\), and \(\nu _4(V)=\nu _5(V)=1\). Then
7.3.4 Conclusion
After the steps in Sects. 7.3.1–7.3.3, it remains to estimate \(X_k\) for \(k\geqslant 4\).
When \(k\geqslant 4\) is even, the estimate of \(X_k\) is similar to that of \(X_2\) in Sect. 7.3.2. In fact, by Lemma 2.2, we see that there will be additional factors of \(N^{-\beta }\) in \(X_k\) when \(k \geqslant 4\), which makes the estimate easier. Using Lemma 7.4 (i), one can show that
When \(k\geqslant 4\) is odd, the estimate of \(X_k\) is similar to that of \(X_3\) in Sect. 7.3.3. By Lemma 2.2, we see that there will be additional factors of \(N^{-(k-2)\beta }\) in \(X_k, k\geqslant 4\). Using Lemmas 7.1, 7.4 and 7.9, one can show that
As a result, we arrive at
where \(b_1=-(4n-2)N^{2\beta -1}\sum _{i,j}{\mathcal {C}}_4(H_{ij})\) is bounded.
7.4 The computation of (II’) in (7.14)
Using Lemma 2.1 with \(h=H_{ij}\), we have
For each k, we write
Each \(Z_k\) can be handled again by applying Lemma 2.1 with \(h=H_{ij}\). One easily shows that
By \({\mathcal {C}}_2(H_{ij})=N^{-1}(1+O(\delta _{ij}))\), we have
Combining with (7.44) and (7.45), we have
The analysis of \({\widehat{X}}_{k}\) is similar to those of \(X_k\) in Sect. 7.3, and we only sketch the key steps.
For \(k=2\), we see from (7.3) that the most dangerous term in \({\widehat{X}}_{2}\) is
which is very close to the left-hand side of (7.23). We can apply Lemma 7.4 (i) and show that (7.47) is bounded by \(O_{\prec }(\sum _{r=2}^{2n} {\mathcal {E}}^r {\mathcal {P}}^{2n-r})\). Similarly, we can also handle all the other terms in \({\widehat{X}}_{2}\), which leads to
For \(k=3\), by the differential rule (7.3), we see that the most dangerous term in \({\widehat{X}}_{3}\) is
which is very close to the right-hand side of (7.25). Similarly to (7.26), we have
and the right-hand side can be computed similarly to \(X_{3,2,1}, X_{3,2,2}, X_{3,2,3}\) in (7.26). As a result, we can show that
where \(b_1=-(4n-2)N^{2\beta -1}\sum _{i,j}{\mathcal {C}}_4(H_{ij})\).
For \(k \geqslant 4\), the argument is similar to that in Sect. 7.3.4. We can show that
Combining the above with (7.46)–(7.49), we have
Now observe the cancellation between (7.43) and (7.50), which leads to
as desired.
7.5 The estimate of (7.15)
From the construction of \(P_0\) in Sect. 5.2, we can easily show that
and in this section we shall see that the analogue holds when the factor \(P^{2n-1}\) is added inside the expectations. Let us write
and analyse each \(X_k^{(1)}\).
Let us first consider the case when k is odd. For \(k=1\), it is easy to see from (7.3) and Lemma 2.8 that
For odd \(k \geqslant 3\), we see from (7.3) and Lemma 2.8 that
where \(a^{(k)}_{ij}\) is deterministic and uniformly bounded.
For even k, we follow a similar strategy as in Sect. 7.3.2. We see from (7.3) and Lemma 2.8 that
where \(a^{(k)}_{ij}\) is deterministic and uniformly bounded. The first term on right-hand side of (7.53) can be written as \({\mathbb {E}} {\mathcal {S}}(V)\), where \(V\in {\mathcal {V}}\), \(\nu _2(V)\ne 0\) and \(\nu _4(V)=\nu _5(V)=0\). Thus we can apply Lemma 7.4 (ii) to estimate this term, and show that it is bounded by \( O_{\prec }(\sum _{r=1}^{2n} {\mathcal {E}}^r {\mathcal {P}}^{2n-r})\). This implies
Combining (7.51), (7.52) and (7.54), we have
where we recall the definition of \({\mathcal {S}}(T)\) in (5.2). Observe that from the above steps, \(T^{(s)}\) in (7.55) is the same as in (5.16). To handle \({\mathbb {E}}[ {\mathcal {S}}(T^{(s)})P^{2n-1}]\), we introduce the following analogue of Lemmas 5.6 and 7.7.
Lemma 7.10
Let \(T \in {\mathcal {T}}_0\) with \(\nu _1(T)-\theta (T) \leqslant -2\beta \). Fix \(r \in {\mathbb {N}}\) and let \({\mathcal {M}}(r,T)\) be as in Lemma 5.6. Then
Proof
The proof analogous to those of Lemmas 5.6 and 7.7. We use the identity
to replace the diagonal entries in \({\mathcal {S}}(T)\), and then expand the terms containing H using Lemma 2.1. We omit the details. \(\square \)
By Lemma 7.10 we have, for any \(s \in \{2,3,\ldots ,\lceil {\ell /2} \rceil \}\),
Together with (7.55) we have
From Lemma 7.10, we see that the term \({\mathcal {M}}(\lceil {\beta ^{-1}-2s+2} \rceil ,T^{(s)})\) in (7.56) is the same as in (5.18), which implies
Thus
and together with (7.15) we conclude that
as desired. This concludes the proof of Lemma 7.2 and hence also that of Proposition 6.1.
8 Proof of Proposition 6.3
Convention
Throughout this section, z is given by (6.1), where \(w \in {{\mathbf {Y}}}_*\) is deterministic.
Let us fix \(n \in {\mathbb {N}}_+\) and set
We shall show that
from which Proposition 6.3 follows by Chebyshev’s inequality. We shall see that the proof of (8.1) is much simpler than that of (7.1), as it does not require a secondary expansion as in Sect. 7.3.3. We define the parameter
Recall the definition of \(\Gamma \) from (5.3). It is easy to check that
In addition, recall the definitions of \(P'\), Q and \(Q_0\) from (7.2). With the help of (6.4), we obtain the following improved version of Lemma 7.1.
Lemma 8.1
We have
and
By \(zG=HG-I\), we have
where in the second step we used that H has real entries.
Remark 8.2
Although we used that the entries of H are real in (8.4), our argument easily extends to complex entries of H. To see how, for any holomorphic \(f:{\mathbb {C}}_+\rightarrow {\mathbb {C}}\) we define \(J f(z) :=\frac{1}{2 \mathrm {i}} (f(z) - f({{\overline{z}} \,}))\). We view all quantities appearing in our arguments as functions of z and use the operator J instead of \({{\,\mathrm{Im}\,}}\). Then it is easy to check that in both real and complex cases, Proposition 6.3 as well as all its consequences remain true if we replace \({{\,\mathrm{Im}\,}}\) by J everywhere. Note that \({{\,\mathrm{Im}\,}}{{\underline{G}} \,} = J {{\underline{G}} \,}\) and \({{\,\mathrm{Im}\,}}P = J P\), but in general \({{\,\mathrm{Im}\,}}G_{ij} \ne J G_{ij}\). An alternative point of view is to regard all of our quantities as functions of z and H, and to take the imaginary part with respect to the Hermitian conjugation of z and H.
Similarly to (7.11), we can use Lemma 2.1 on the last term of (8.4), and get
We shall prove the following result, which directly implies (8.1).
Lemma 8.3
Let (V)–(VII) be as in (8.5). Then
and
8.1 Proof of (8.7)
Define
so that \(\mathrm {(VII)}=\sum _{k=1}^{\ell }X_{k}^{(2)}\). Note that for \(f:{\mathbb {R}} \rightarrow {\mathbb {C}}\) and h real, \(\frac{\mathrm {d}{{\,\mathrm{Im}\,}}f(h)}{\mathrm {d}h}={{\,\mathrm{Im}\,}}\frac{\mathrm {d}f(h)}{\mathrm {d}h}\), so that the derivatives in (8.8) can be computed through (7.3). Let us estimate each \(X_k^{(2)}\).
For any fixed \(k \in {\mathbb {N}}_+\), it is easy to see from (8.3) that
By (7.3) and Proposition 2.5, we see that
where in the last step we estimated \({{\,\mathrm{Im}\,}}G_{xx}\) by \(O_\prec ({{\,\mathrm{Im}\,}}{{\underline{G}} \,})\), using its spectral decomposition and Lemma 2.6. Here we see the crucial effect of taking imaginary part of P, which results \({{\,\mathrm{Im}\,}}{\underline{G}} \,\) on right-hand side of (8.9) instead of \({\underline{G}} \,\). Note that \({{\,\mathrm{Im}\,}}{\underline{G}} \, \leqslant \Phi +{{\,\mathrm{Im}\,}}m\asymp \Phi +\eta /\sqrt{\kappa }\), and together with Lemma 2.8 we have
Thus
By Cauchy-Schwarz and (6.5) we have \(\Theta ^{1/2}\prec \Psi +\frac{1}{N\eta }+\frac{\eta }{\sqrt{\kappa }}\prec \Psi +\frac{1}{N\eta }\), thus
Together with \( (\sqrt{\kappa }\Theta )^{s} \prec {\mathcal {E}}_{{{\,\mathrm{Im}\,}}}^{s} \) for all \(s\geqslant 0\), we get
for all \(r\geqslant 1\). Combining the above estimate with (8.10), we have
This concludes the proof of (8.7).
8.2 Proof of (8.6)
The proof is similar to the estimate of (7.15) in Sect. 7.5. Define
so that \(\mathrm {(VI)}=\sum _{k=1}^{\ell }X_k^{(3)}\). We analyse each \(X_k^{(3)}\).
Consider first the case when k is odd. For \(k=1\), it is easy to see from (7.3) and Lemma 2.8 that
When \(k \geqslant 3\) is odd, we see from (7.3) and Lemma 2.8 that
where \(a^{(k)}_{ij}\) is deterministic and uniformly bounded.
For even k, we see from (7.3) and Lemma 2.8 that
where \(a^{(k)}_{ij}\) is deterministic and uniformly bounded. Note that the analogue of (8.13) has appeared in (7.53). To handle this term, we use the following result.
Lemma 8.4
Fix an even \(k \geqslant 2\). Let \(\big (a^{(k)}_{ij}\big )_{i,j=1}^N\) be deterministic and uniformly bounded. Then
Proof
The proof essentially follows from the strategy of showing Lemmas 5.3 and 7.4. We use the identity
to replace the \(G_{ij}\) in the equation, and then expand the terms containing H using Lemma 2.1. We omit the details. \(\square \)
Lemma 8.4 immediately implies
Combining (8.11)–(8.14), we have
Here we recall the definition of \({\mathcal {S}}(T)\) in (5.2), and observe that \(T^{(s)}\) above is the same as in (5.16). To handle the last relation, we introduce the following analogue of Lemmas 5.6 and 7.7.
Lemma 8.5
Let \(T \in {\mathcal {T}}_0\) with \(\nu _1(T)-\theta (T) \leqslant 0\). Fix \(r \in {\mathbb {N}}\), and let \({\mathcal {M}}(r,T)\) be as in Lemma 5.6. Then
Proof
The proof is similar to those of Lemmas 5.6 and 7.7. We use the identity
to replace the diagonal entries in \({\mathcal {S}}(T)\), and then expand the terms containing H using Lemma 2.1. We omit the details. \(\square \)
By Lemma 8.5, we have, for any \(s \in \{2,3,\ldots ,\lceil {\ell /2} \rceil \}\),
Thus
From Lemma 8.5, we see that the term \({\mathcal {M}}(\lceil {\beta ^{-1}-2s+2} \rceil ,T^{(s)})\) above is the same as in (5.18), which implies
Thus
In addition, note that
From the definition of \(\mathrm {(V)}\) in (8.5), we conclude that
as desired. This concludes the proof of Lemma 8.3, and hence also that of Proposition 6.3.
9 Proof of Lemma 4.2
Convention
Throughout this section, z is given by (6.1), where w is deterministic and contained in
where \(c>0\) is fixed.
The key in proving Lemma 4.2 is the following result.
Proposition 9.1
Suppose \(|{\underline{G}} \,-m| \prec \Psi \) for some deterministic \(\Psi \in [N^{-1},1]\). Then
uniformly for all \(w \in {{\mathbf {D}}}\).
The stability analysis of P was dealt for the region \({{\mathbf {Y}}}\) in Lemma 6.2, and one easily checks that the same result holds for the region \({{\mathbf {D}}}\). This leads to the next lemma.
Lemma 9.2
Lemma 6.2 holds provided that \({{\mathbf {Y}}}\) is replaced with \({{\mathbf {D}}}\).
Combining Proposition 9.1 and Lemma 9.2, we obtain the implication
and thus
uniformly for all \(w \in {{\mathbf {D}}}\). By the rigidity estimate (9.2), together with a standard analysis using Helffer-Sjöstrand formula (e.g. [12, Proposition 3.2]), one immediately concludes the proof of Lemma 4.2.
The rest of the section is devoted to the proof of Proposition 9.1. It is simpler than that of Proposition 6.1, and we only give a sketch. A detailed proof of a slightly weaker result can be found in [19, Proposition 2.9].
9.1 Proof of Proposition 9.1
Fix \(n \in {\mathbb {N}}_+\) and set
We shall show that
and Proposition 9.1 is obtained by Chebyshev’s inequality.
We shall see that the proof of (9.3) is much simpler than that of (7.1), as it does not require a secondary expansion as in Sect. 7.3.3. Recall the definitions of \(P'\), Q and \(Q_0\) from (7.2), and recall the definition of \(\Upsilon \) from (7.5). We have the bound
In addition, note that Lemma 7.1 remains true for \(w \in {{\mathbf {D}}}\).
Similarly to (7.11), we have
The following result directly implies (9.3).
Lemma 9.3
We have
as well as
We now sketch the proof of (9.6). The proof of (9.7) follows in a similar fashion. Let us first consider (XI). We write \(\text {(XI)}=\sum _{k=1}^lX^{(4)}_k\), where
For \(k=1\), one can repeat the steps in Sect. 7.3.1 and show that
Note that we have the bound (9.4), which implies
For \(k=2\), one can follow the steps in Section 2 of [19], and show that \(X_2^{(4)} \prec \sum _{r=2}^{2n} \Upsilon ^{r} {\mathcal {P}}^{2n-r}\). Thus,
A similar strategy works for all even \(k \geqslant 4\). This gives
For \(k=3\), we split \(X_3^{(4)}=X_{3,1}^{(4)}+X_{3,2}^{(4)}+X_{3,3}^{(4)}\), where
for \(s=1,2,3\). Similarly to Step 1 of Sect. 7.3.3, we can show that
where in the last step we used
Now consider \(X_{3,2}^{(4)}\). Similarly to (7.26), we have
Thus
A similar strategy works for all odd \(k \geqslant 5\). Note that (9.10) implies the bound
By Lemma 2.2, we see that, compared to \(X^{(4)}_3\), there will be additional factors of \(N^{-(k-2)\beta }\) in \(X^{(4)}_k\) for all \(k\geqslant 4\). Thus we can shown that
Using the above relation, together with (9.8)–(9.10), we get
The computation of (IX) is similar, and we can show that
By Proposition 2.5, we have
Thus, there is a cancellation between the leading order terms of (IX) and (XI), which implies
as desired. This concludes the proof of Lemma 9.3, and also that of Lemma 4.2.
10 Proof of the improved estimates for abstract polynomials
In this section we repeatedly use the following identity.
Lemma 10.1
We have
Proof
The resolvent identity \((H-z)G=I\) shows
and from which the proof follows. \(\square \)
Let f(G) be a function of the entries of G. We compute \({\mathbb {E}} f(G)G_{ij}\) through
and we shall see that the last two terms above cancel each other up to leading order, by Lemma 2.1. As a result, we can replace \({\mathbb {E}} f(G)G_{ij}\) by a slightly nicer quantity \({\mathbb {E}} f(G)\delta _{ij}{\underline{G}} \,\). This is the idea that we use throughout this section.
In each of the following subsections, the assumptions on z are given by the assumptions of the corresponding lemma being proved.
10.1 Proof of Lemma 5.3
As discussed in Remark 5.4, it suffices to look at the case \(\nu _2=1\).
Without loss of generality, let \(T_{i_1,\ldots ,i_{\nu _1}}=a_{i_1,\ldots ,i_{\nu _1}}N^{-\theta }G_{i_1i_2}G_{x_2x_2}\ldots G_{x_\sigma x_\sigma }\), where \(x_2,\ldots ,x_\sigma \in \{i_1,\ldots ,i_{\nu _1}\}\), and \(a_{i_1,\ldots ,i_{\nu _1}}\) is uniformly bounded. Using Lemma 10.1 for \(i=i_1\) and \(j=i_2\), we have
By Lemma 2.1 and estimating the remainder term for large enough \(\ell \), the second last term in (10.1) becomes
Similarly, the last term in (10.1) becomes
Let us estimate each \(X_k^{(5)}\) and \(X^{(6)}_k\).
For \(k=1\), by \({\mathcal {C}}_2(H_{ij})=N^{-1}(1+O(\delta _{ij}))\) and Lemma 2.8 we have
and
Notice the cancellation between the above two equations. This gives
For \(k=2\), the most dangerous type of term in \(X_2^{(6)}\) contains only one off-diagonal entry of G, e.g.
Note that (10.3) can be written as \({\mathbb {E}}{\mathcal {S}}(T')\), where \(T' \in {\mathcal {T}}\) and \(\nu _1(T') = \nu _1(T) + 1\), \(\theta (T') = \theta (t) + 1 + \beta \), and \(\sigma (T')=\sigma (T)+1\). When a term in \(X_2^{(6)}\) contains at least two off-diagonal entries of G, one can use Lemma 2.8 to show that it is bounded by \(O_{\prec }(N^{\theta -\nu _1}{\mathbb {E}} \Gamma )\). A similar argument works for all \(X_k^{(5)}\) and \(X^{(6)}_k\) when \(k \geqslant 2\).
To sum up, we have
for some fixed integer m. Each \(T^{(l)}\) satisfies \(\nu _1(T^{(l)})=\nu _1(T)+1\), \(\sigma (T^{(l)}) \geqslant \sigma (T)+1\), \(\theta (T^{(l)})=\theta (T)+1+\beta (\sigma (T^{(l)})-\sigma (T))\) and \(\nu _2(T^{(l)})=1\), which implies
Note that we can repeat (10.4) for each \({\mathbb {E}}\, {\mathcal {S}}(T^{(l)})\), and get
and each \({\mathbb {E}} \,{\mathcal {S}}(T^{(l,l')})\) satisfies \( {\mathbb {E}}\, {\mathcal {S}}(T^{(l,l)})\prec N^{\nu _1(T)-\theta (T)-2\beta }. \) Repeating the step (10.4) \(\lceil {\beta ^{-1}} \rceil \) times concludes the proof of Lemma 5.3.
10.2 Proof of Lemma 5.5
Let \(T_{i_1,\ldots ,i_{\nu _1}}=a_{i_1,\ldots ,i_{\nu _1}}N^{-\theta }G_{x_1x_1}G_{x_2x_2}\ldots G_{x_\sigma x_\sigma }\), where \(x_1,\ldots ,x_\sigma \in \{i_1,\ldots ,i_{\nu _1}\}\), and \(a_{i_1,\ldots ,i_{\nu _1}}\) is uniformly bounded. Using Lemma 10.1 we have
Now let us expand the last two terms by Lemma 2.1. As in Sect. 10.1, we shall see a cancellation among the leading terms, which gives
For the terms on right-hand side of (10.5) that are not in \({\mathcal {T}}_0\), we can use Lemma 5.3 and show that they are bounded by \(O_{\prec }(N^{\nu _1(T)-\theta (T)}({\mathbb {E}}\Gamma +N^{-1}))\). As a result, we find
for some fixed integer m. Each \(T^{(l)}\) satisfies \(T^{(l)} \in {\mathcal {T}}_0\), \(\nu _1(T^{(l)})=\nu _1(T)+1\), \(\sigma (T^{(l)})- \sigma (T)\in 2{\mathbb {N}}+4\), and \(\theta (T^{(l)})=\theta (T)+1+\beta (\sigma (T^{(l)})-\sigma (T)-2)\). We can then repeat (10.6) on the term
After \(k-1\) times of repetition we get the desired result. This concludes the proof of Lemma 5.5.
10.3 Proof of Lemma 7.4
(i) Let V be of the form (7.16). By Lemma 2.8, we see that the result is trivially true for \(\nu _2\geqslant 2\), and hence we assume \(\nu _2=1\). Define
By the definition of \(\nu _2\), we consider two cases.
Case 1 The contribution of \(\nu _2\) comes from \(G_{x_1y_1}G_{x_2y_2}\ldots G_{x_ky_k}\). Without loss of generality, we assume \(x_1 \ne y_1\), and \(x_1=i_1,y_1=i_2\). Furthermore, we denote \({\widehat{V}}_{i_1,\ldots ,i_{\nu _1}}=V_{i_1,\ldots ,i_{\nu _1}}/G_{i_1i_2}\). From Lemma 10.1 we know that
By Lemma 2.1 and estimating the remainder term for large enough \(\ell \), the second last term in (10.7) becomes
and we denote the first sum by \(\sum _{k=1}^\ell X^{(7)}_k\). Similarly, the last term in (10.7) becomes
and we denote the first sum by \(\sum _{k=1}^\ell X^{(8)}_k\). Similarly to Sect. 10.1, we see that when expanded by Lemma 2.1, the leading terms of \(X_1^{(7)}\) and \(X_1^{(8)}\) cancel, and together with Lemma 7.1 we can show that
For \(k=2\), the most dangerous type of term in \(X_2^{(8)}\) contains \(\nu _3\) factors of P, and only one off-diagonal entry of G or \(G^2\), e.g.
Note that this term can be written as \({\mathbb {E}}{\mathcal {S}}(V')\), where \(V' \in {\mathcal {V}}\), \(\nu _1(V') = \nu _1(V) + 1\), \(\theta (V') = \theta (V) + 1 + \beta \), \(\sigma (V')=\sigma (V)+1\), and \(\nu _i(V')=\nu _i(V)\) for \(i=2,3,4,5\). When a term in \(X_2^{(8)}\) contains at least two factors of off-diagonal entries of G or \(G^2\), or the differential \(\partial ^2/\partial H^2_{i_1x}\) hits \(P^{\nu _3}\), one can easily use Lemma 7.1 to show that it is bounded by \(O_{\prec }(N^{\nu _1(V)-\theta (V)-2}{\mathbb {E}} |P^{\nu _3}|)\). A similar argument works for all \(X_k^{(5)}\) and \(X^{(6)}_k\) when \(k \geqslant 2\).
To sum up, we have
for some fixed integer m. Each \(V^{(l)}\) satisfies \(\nu _1(V^{(l)})=\nu _1(V)+1\), \(\sigma (V^{(l)}) \geqslant \sigma (V)+1\), \(\theta (V^{(l)})=\theta (V)+1+\beta (\sigma (V^{(l)})-\sigma (V))\), and \(\nu _i(V^{(l)})=\nu _i(V)\) for \(i=2,3,4,5\). Thus Lemma 7.1 implies
Note that we can repeat (10.10) for each \({\mathbb {E}} \,{\mathcal {S}}(V^{(l)})\) on right-hand side of (10.10). Doing this \(\lceil {(2\beta )^{-1}} \rceil \) times concludes the proof.
Case 2 The contribution to \(\nu _2\) comes from \(N^{-1}(G^2)_{xy}\). Without loss of generality, we assume \(x=i_1,y=i_2\), and we denote \({\widetilde{V}}_{i_1,\ldots ,i_{\nu _1}}=V_{i_1,\ldots ,i_{\nu _1}}/(G^2)_{i_1i_2}\). Note that
and hence
We can then expand the first two terms on right-hand side of (10.11) using Lemma 2.1. The first term on right-hand side of (10.11) gives
and we abbreviate the first sum above by \( \sum _{k=1}^{\ell }X_{k}^{(9)}\). The second term on right-hand side of (10.11) gives
and we abbreviate the first sum above by \( \sum _{k=1}^{\ell }X_{k}^{(10)}\). By (7.4), we see that
and
Thus there is a cancellation between \( X_1^{(9)}\) and \( X_1^{(10)}\), which shows
The first term on right-hand side of (10.13) is the leading term, and it no longer contains \((G^2)_{i_1i_2}\). The rest of the proof is analogues to Case 1. We omit the details.
(ii) Let \(V \in {\mathcal {V}}\) satisfy \(\nu _2(V)\ne 0\) and \(\nu _4(V)=\nu _5(V)=0\). From the result in (i), we have the bound
so that we only need to improve the bound for the term \(t=1\). Once again it suffices to assume \(\nu _2(V)=1\), and \(x_1=i_1\), \(y_1=i_2\). We denote \({\widehat{V}}_{i_1,\ldots ,i_{\nu _1}}=V_{i_1,\ldots ,i_{\nu _1}}/G_{i_1i_2}\). As in (10.7)–(10.9), we have
Let us pick a term \({\mathcal {X}}\) in \(\sum _{k=1}^\ell X^{(7)}_k+\sum _{k=1}^\ell X^{(8)}_k\), which, we recall, are given by the sums in (10.8) and (10.9). When \(\nu _3({\mathcal {X}}) \ne \nu _3(V)-1\), we handle this term as in the proof of (i). When \(\nu _3({\mathcal {X}})=\nu _3(V)-1\), then from (7.4), we must have \(\nu _4({\mathcal {X}})=\nu _5({\mathcal {X}})=1\). Thus from (i), we have
as desired. This concludes the proof of Lemma 7.4.
10.4 Proof of Lemma 7.5
Let \(V=a_{i_1,\ldots ,i_{\nu _1}}N^{-\theta }(P'N^{-1}(G^2)_{xx})^{\nu _4}G_{x_1x_1}G_{x_2x_2}\ldots G_{x_{k}x_{k}}{\underline{G}} \,^sP^{\nu _3}\). We abbreviate \({\widehat{V}}_{i_1,\ldots ,i_{\nu _1}}:=V_{i_1,\ldots ,i_{\nu _1}}/G_{x_1x_1}\), and denote
Using Lemma 10.1 we have
Now let us expand the last two terms by Lemma 2.1. As in Sect. 10.1, we shall see a cancellation among the leading terms. For other terms that are not in \({\mathcal {T}}_0\), we can use Lemma 5.3 and show that they are bounded by \(O_{\prec }({\mathcal {E}}_3(V))\). As a result, we can show that
for some fixed integer m. Each \(V^{(l)}\) satisfies \(V^{(l)} \in {\mathcal {V}}_0\), \(\nu _1(V^{(l)})=\nu _1(V)+1\), \(\sigma (V^{(l)})- \sigma (V)\in 2{\mathbb {N}}+4\), \(\theta (V^{(l)})=\theta (V)+1+\beta (\sigma (V^{(l)})-\sigma (V)-2)\), and \(\nu _i(V^{(l)})=\nu _i(V)\) for \(i=2,3,4,5\). One can then repeat (10.14) process on the term
After k times of repetition we conclude the proof of Lemma 7.5.
10.5 Proof of Lemma 7.7
The proof follows by repeatedly using the following result.
Lemma 10.2
Fix \(r,u,v\in {\mathbb {N}}\). For any fixed \(T \in {\mathcal {T}}_0\) there exists \(T^{(1)},\ldots ,T^{(k)} \in {\mathcal {T}}_0\), such that
where k is fixed. Each \(T^{(l)}\) satisfies \(\sigma (T^{(l)})-\sigma (T) \in 2{\mathbb {N}}+4\),
Proof of Lemma 10.2
We abbreviate the error, i.e. the last two terms on right-hand side of (10.15), by \({\mathcal {E}}_4\equiv {\mathcal {E}}_4(T,u,v)\).
Let \(T_{i_1,\ldots ,i_{\nu _1}}=a_{i_1,\ldots ,i_{\nu _1}}N^{-\theta }G_{x_1x_1}G_{x_2x_2}\ldots G_{x_kx_k}\), where \(x_1,\ldots ,x_k \in \{i_1,\ldots ,i_{\nu _1}\}\) and \(a_{i_1,\ldots ,i_{\nu _1}}\) is uniformly bounded. We abbreviate \({\widehat{T}}_{i_1,\ldots ,i_{\nu _1}}:=T_{i_1,\ldots ,i_{\nu _1}}/G_{x_1x_1}\). By Lemma 10.1, we have
Now we expand the last two terms in (10.16) using Lemma 2.1. Note that we have
The rest of the proof is analogous to that of Lemma 5.3. We omit the details. \(\square \)
Notes
Here \(\partial _{{{\overline{w}} \,}}\) denotes the antiholomorphic derivative in the complex variable w.
References
Alon, N.: Spectral techniques in graph algorithms. In: LATIN’98: Theoretical Informatics, pp. 206–215. Springer (1998)
Bauerschmidt, R., Huang, J., Knowles, A., Yau, H.T.: Edge rigidity and universality of random regular graphs of intermediate degree (2019). Preprint arXiv:1910.10121
Bourgade, P., Mody, K.: Gaussian fluctuations of the determinant of Wigner matrices. Electr. J. Prob. 24, 1–28 (2019)
Chung, F.R.K.: Spectral Graph Theory, No. 92. American Mathematical Society, Providence (1997)
de Monvel, A.B., Khorunzhy, A.: Asymptotic distribution of smoothed eigenvalue density. I. Gaussian random matrices. Random Oper. Stoch. Equ. 7, 1–22 (1999)
de Monvel, A.B., Khorunzhy, A.: Asymptotic distribution of smoothed eigenvalue density. II. Wigner random matrices. Random Oper. Stoch. Equ. 7, 149–168 (1999)
Erdős, L., Knowles, A., Yau, H.-T.: Averaging fluctuations in resolvents of random band matrices. Ann. H. Poincaré 14, 1837–1926 (2013)
Erdős, L., Knowles, A., Yau, H.-T., Yin, J.: Spectral statistics of Erdős–Rényi graphs II: eigenvalue spacing and the extreme eigenvalues. Commun. Math. Phys. 314, 587–640 (2012)
Erdős, L., Knowles, A., Yau, H.-T., Yin, J.: Spectral statistics of Erdős–Rényi graphs I: local semicircle law. Ann. Prob. 41, 2279–2375 (2013)
Erdős, L., Yau, H.-T., Yin, J.: Rigidity of eigenvalues of generalized Wigner matrices. Adv. Math. 229, 1435–1515 (2012)
Gustavsson, J.: Gaussian fluctuations of eigenvalues in the GUE. Ann. Inst. Henri Poincaré (B) 41, 151–178 (2005)
He, Y.: Bulk eigenvalue fluctuations of sparse random matrices. Preprint arXiv:1904.07140
He, Y., Knowles, A.: Mesoscopic eigenvalue statistics of Wigner matrices. Ann. Appl. Prob. 27, 1510–1550 (2017)
He, Y., Knowles, A.: Mesoscopic eigenvalue density correlations of Wigner matrices. Prob. Theory Relat. Fields 177, 147–216 (2020)
He, Y., Knowles, A., Marcozzi, M.: Local law and eigenvector delocalization for supercritical Erdős–Rényi graphs. Ann. Prob. 47, 3278–3302 (2019)
He, Y., Knowles, A., Rosenthal, R.: Isotropic self-consistent equations for mean-field random matrices. Prob. Theory Relat. Fields 171, 203–249 (2018)
Hoory, S., Linial, N., Wigderson, A.: Expander graphs and their applications. Bull. Am. Math. Soc. 43(4), 439–561 (2006)
Huang, J., Landon, B., Yau, H.-T.: Bulk universality of sparse random matrices. J. Math. Phys. 56, 123301 (2015)
Huang, J., Landon, B., Yau, H.-T.: Transition from Tracy–Widom to Gaussian fluctuations of extremal eigenvalues of sparse Erdős–Rényi graphs. Ann. Prob. 48, 916–962 (2020)
Khorunzhy, A.M., Khoruzhenko, B.A., Pastur, L.A.: Asymptotic properties of large random matrices with independent entries. J. Math. Phys. 37, 5033–5060 (1996)
Landon, B., Sosoe, P.: Applications of mesoscopic CLTs in random matrix theory (2018). Preprint arXiv:1811.05915
Landon, B., Yau, H.-T.: Convergence of local statistics of Dyson Brownian motion. Commun. Math. Phys. 355, 949–1000 (2017)
Lee, J.O., Yin, J.: A necessary and sufficient condition for edge universality of Wigner matrices. Duke Math. J. 163, 117–173 (2014)
Lee, J.O., Schnelli, K.: Local law and Tracy–Widom limit for sparse random matrices. Prob. Theory Relat. Fields 171, 543–616 (2018)
O’Rourke, S.: Gaussian fluctuations of eigenvalues in Wigner random matrices. J. Stat. Phys. 138, 151–178 (2010)
Sinai, Y., Soshnikov, A.: A refinement of Wigner’s semicircle law in a neighborhood of the spectrum edge. Funct. Anal. Appl. 32, 114–131 (1998)
Soshnikov, A.: Universality at the edge of the spectrum in Wigner random matrices. Commun. Math. Phys. 207, 697–733 (1999)
Tao, T., Vu, V.: Random matrices: universality of local eigenvalue statistics up to the edge. Commun. Math. Phys. 298, 549–572 (2010)
Tracy, C., Widom, H.: Level-spacing distributions and the Airy kernel. Commun. Math. Phys. 159, 151–174 (1994)
Tracy, C., Widom, H.: On orthogonal and symplectic matrix ensembles. Commun. Math. Phys. 177, 727–754 (1996)
Acknowledgements
The authors would like to thank Zhigang Bao, Jiaoyang Huang, and Benjamin Schlein for helpful discussions. Y.H. gratefully acknowledges partial support from the NCCR SwissMAP and the Swiss National Science Foundation through the Grant 200020_172623 “Dynamical and energetic properties of Bose-Einstein condensates”. A.K. gratefully acknowledges funding from the European Research Council (ERC) under the European Union’s Horizon 2020 research and innovation programme (grant agreement No. 715539_RandMat) and from the Swiss National Science Foundation through the NCCR SwissMAP grant.
Funding
Open Access funding provided by Universität Zürich.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Proof of Lemma 2.12
Proof of Lemma 2.12
We prove the result for \(i \in \{1,2,\ldots ,\lceil {N/2} \rceil \}\). The same analysis works for the other half of the spectrum. Let us denote
and recall that \(P_0(z,x)=P(z,x)-{\mathcal {Z}} x^2\), and \(m_0\) satisfies \(P_0(z,m_0(z))=0\). Recall the definition of \( \widetilde{{{\mathbf {S}}}}\) from (2.6), and define \(F: \widetilde{{{\mathbf {S}}}}\times {\mathbb {C}} \rightarrow {\mathbb {C}}\) by
Since \(|m_0(z)|\asymp 1\) for \(z \in \widetilde{{{\mathbf {S}}}}\), and \(R(x),R'(x)=O(1/q^2)\) uniformly for \(|x|\leqslant 100\), it is easy to check that
with very high probability. By implicit function theorem, we can define a map \(\Delta :\widetilde{{{\mathbf {S}}}}\rightarrow {\mathbb {C}}\) satisfying
By (2.5), it is east to check that
uniformly for \(z \in \widetilde{{{\mathbf {S}}}}\). Set \(M=m_0/\sqrt{1+\Delta }\). We have
and (A.1) implies
Combining the above two relations gives
Thus
Let \({\mathcal {E}}_5:=1/(\sqrt{N}q^3)\). Then (A.2) and (A.3) imply
Let us write \(z=E+\mathrm {i}\eta \). We have
We set \({\widetilde{L}}:=L_0(1+{\mathcal {Z}}/2)\), and note that
Since \(\varrho _0\) has square root behaviour near the edge, we have
For any \(i \in \{1,2,\ldots ,\lceil {N/2} \rceil \}\), we have
and
where \({\widetilde{\gamma }}_{0,i}:=\gamma _{0,i}(1+{\mathcal {Z}}/2)\). Thus
We claim that
which together with the trivial estimate
implies the desired result.
If \(\gamma _i<-{\widetilde{L}}\), then \(\gamma _i \geqslant -L\) and (A.4) imply \(\gamma _i+{\widetilde{L}} \prec {\mathcal {E}}_5\). Thus (A.5) shows \({\mathcal {E}}_6 \prec {\mathcal {E}}_5^{3/2}\). We also have
which implies \({\widetilde{\gamma }}_{0,i}+{\widetilde{L}} \prec {\mathcal {E}}_5\). The claim then follows from \(\gamma _i+{\widetilde{L}} \prec {\mathcal {E}}_5\) and a triangle inequality.
If \(\gamma _i \in [ -{\widetilde{L}}, {\widetilde{\gamma }}_{0,i}]\) then it suffices to assume \({\widetilde{\gamma }}_{i,0}+{\widetilde{L}}\geqslant {\mathcal {E}}_5\). We have
and together with (A.5) we get
If \(\gamma _i >{\widetilde{\gamma }}_{0,i}\) then it suffices to assume \(\gamma _i+{\widetilde{L}} \geqslant {\mathcal {E}}_5\). We have
which together with (A.5) implies the claim. This concludes the proof of Lemma 2.12.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
He, Y., Knowles, A. Fluctuations of extreme eigenvalues of sparse Erdős–Rényi graphs. Probab. Theory Relat. Fields 180, 985–1056 (2021). https://doi.org/10.1007/s00440-021-01054-4
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00440-021-01054-4