1 Introduction

Clustering problems are ubiquitous in data science. The main goal is to find a partition of the data into clusters in such form that the members in the same cluster are more similar than the members in different clusters. At the same time it is necessary to balance the clusters sizes to avoid the trivial solution of one cluster per member.

A large body of work has focused on the stochastic block model, a random network model with a planted cluster structure, we refer the reader to [2] for a survey on recent developments. We will focus on case of two balanced communities. Let n be an even natural number and \(G \sim {\mathcal {G}}(n;p,q)\) be a random graph on n nodes drawn as follows: Randomly partition the set of n vertices V in two equally sized communities \(V = S_1 \cup S_2\). For every pair of vertices, an edge is placed with probability p if they belong to the same community \(S_i\) and with probability \(q<p\) otherwise, all independent. The goal is to exactly recover the partition \(\{S_1,S_2\}\) from the graph alone. Let the matrix \(A \in {\mathbb {R}}^{n\times n}\) denote the adjacency matrix of the graph G. Considering a label vector \(x\in \{\pm 1\}^n\) representing community membership of notesFootnote 1. The maximum likelihood estimator for the node labels x is given by the program below [2],

$$\begin{aligned} \begin{aligned} \max _{x} \quad&x^{T}Ax\\ \text {s.t.} \quad&{\mathbf {1}}^{T}x = 0\\&x\in \{\pm 1\}^n \\ \end{aligned} \end{aligned}$$
(1)

Here \({\mathbf {1}}\) denotes all-ones vector. Since it is well known that the problem (1) is NP-Hard [6], we consider the standard semidefinite relaxation [7].

$$\begin{aligned} \begin{aligned} \max _{X \in {\mathbb {R}}^{n\times n}} \quad&\mathrm {Tr}(AX)\\ \text {s.t.} \quad&X_{ii} = 1\\&X \succeq 0 \\&\mathrm {Tr}(X{\mathbf {J}}) = 0 \end{aligned} \end{aligned}$$
(2)

Where X is a surrogate variable for \(xx^{T}\) and \({\mathbf {J}}\) denotes all-ones matrix. The following theorem gives the sharp phase transition for the community detection problem with two balanced communities.

Theorem 1

(Exact recovery threshold [1, 4, 8, 10]) Let \(G \sim {\mathcal {G}}(n;p,q)\) with \(p =\alpha \frac{\log n}{n}\), \(q=\beta \frac{\log n}{n}\) and planted communities \(\{S_1,S_2\}\). Then,

  1. (I)

    For \(\sqrt{\alpha }-\sqrt{\beta } <\sqrt{2}\), no algorithm can exactly recover the partition with high probability.

  2. (II)

    For \(\sqrt{\alpha }-\sqrt{\beta } >\sqrt{2}\), with high probability: The semidefinite program (2) has a unique solution given by \(X^\natural = x^{\natural } (x^\natural ) ^T\) where \(x^\natural \) corresponds to the memberships of the true communities, thus achieving exact recovery.

Although polynomial time, semidefinite programs tend to be computationally costly. A powerful tool to overcome computational complexity is that of sketching (we refer the reader to [13] for an instance of this idea in least squares, and [5, 14] for semidefinite optimization). In the particular framework addressed in this paper, Mixon and Xie [9] have recently proposed a sketching approach wherein a potentially significantly smaller semidefinite program is solved, its size depends on the community structure strength. Our main contribution is to resolve in the positive a conjecture in [9] regarding the dependency of the size of the resulting semidefinite program and the community structure strength. We now describe the sketching approach in [9], which consists of a three step process, and a tuning parameter \(0< \gamma <1 \).

  • (Step 1) Given a graph with vertex set V. Subsample a smaller vertex set \(V^{\natural }\) by sampling each node in V independently at random with probability \(\gamma \).

  • (Step 2) Solve the community detection problem in the subgraph induced by \(V^{\natural }\).

  • (Step 3) For each node v not in \(V^{\natural }\) use a majority vote procedure among the neighbours of v in \(V^{\natural }\) to infer its community membership.

The main goal of this paper is to determine the minimum value of \(\gamma \) such that the approach above exactly recovers both communities with high probability. The computational savings come from the reduced size of the semidefinite program and so the parameter \(\gamma \) governs the computational cost of the algorithm (we refer the reader to [3] for the dependency of the computational cost of semidefinite programming on the number of variables).

Mixon and Xie [9] conjectured that, as long as

$$\begin{aligned} \gamma > \frac{2}{\left( \sqrt{\alpha }-\sqrt{\beta }\right) ^2}, \end{aligned}$$

the sketching approach works with high probability. Our main result provides a positive answer for this conjecture. In particular, for \(\gamma =1\), we recover the threshold in Theorem 1 (see part II).

2 An oracle bound

As described above, the sketching approach consists of three steps: Sampling, solving the community detection problem for a smaller sampled graph and then recovering the entire communities using a majority vote procedure. In this section, we analyze the Step 3 and prove that it works, for a certain range of the parameter \(\gamma \), as long as we know the smaller communities in Step 2. The analysis is described in the proposition below, we refer to it as an oracle bound because it assumes the knowledge of the communities in Step 2.

Proposition 2

Let \(G \sim {\mathcal {G}}(n;p,q)\) with planted communities \(\{S_1,S_2\}\) and with \(p=\alpha \frac{\log n}{n}\) and \(q=\beta \frac{\log n}{n}\) satisfying \(p>q\). Draw a vertex set \(V^{\natural }\) at random by sampling each node of the graph G independently at random with probability \(\gamma \). Let \(R_1,R_2\) be the planted communities in the sampled graph, i.e, \(R_i = S_i \cap V^{\natural }\) for both \(i \in \{1,2\}\). Moreover, let e(vS) be the number of edges of G between the vertex v and the set \(S \subset V(G)\) where V(G) is the vertex set of the graph G. Now, consider

$$\begin{aligned} {\hat{S}}_1= & {} R_1 \cup \{ v\in V(G) {\backslash } V^{\natural }: e(v,R_1)> e(v,R_{2})\} . \\ {\hat{S}}_2= & {} R_2 \cup \{ v\in V(G) {\backslash } V^{\natural }: e(v,R_2) > e(v,R_{1})\} . \end{aligned}$$

Then there exists absolute constants \(C,c>0\) such that, with probability \(1-Cn^{-c((\alpha +\beta )\frac{\gamma }{2}-\gamma \sqrt{\alpha \beta }-1)}\), \(({\hat{S}}_1,{\hat{S}}_2)=(S_1,S_2)\). In particular, \(({\hat{S}}_1,{\hat{S}}_2)=(S_1,S_2)\) with probability \(1-o(1)\), as long as

$$\begin{aligned} \gamma > \frac{2}{\left( \sqrt{\alpha }-\sqrt{\beta }\right) ^2}. \end{aligned}$$

The next lemma will play a key role in the proof of Proposition 2, it is similar to Lemma 8 in [1] but it deals with almost balanced communities, this is crucial to our analysis.

Lemma 1

Suppose \(\alpha> \beta >0\). Let X and Y be two independent random variables with \(X \sim \text {Binom}(K_1, \alpha \frac{\log n}{n})\) and \(Y \sim \text {Binom}(K_2,\beta \frac{\log n}{n})\), where \(K_1 = \frac{n\gamma }{2} + o(n)\) and \(K_2 = \frac{n\gamma }{2} + o(n)\) as \(n \rightarrow \infty \). Then,

$$\begin{aligned} {\mathbb {P}}(X-Y \le 0) \le n^{- ((\alpha +\beta )\frac{\gamma }{2} -\gamma \sqrt{\alpha \beta }) + o(1)}. \end{aligned}$$

We present a simple and direct proof of this lemma.

Proof

Let \(\varepsilon >0\). We proceed with the Laplace transform method, for all \(t\ge 0\) we write

$$\begin{aligned} {\mathbb {P}}(X-Y \le 0) \le {\mathbb {P}}(X-Y \le \varepsilon ) \le e^{t\varepsilon }{\mathbb {E}}e^{-t(X-Y)} := e^{-\psi (t)}, \end{aligned}$$
(3)

where \(\psi (t):= -t\varepsilon - \log {\mathbb {E}}e^{-t(X-Y)}\). Now we use the fact that the function \(\psi (t)\) is additive for sums of independent random variables together with the formula for the moment generating function of a binomial distribution (Example 3.32 in [12])

$$\begin{aligned} \log {\mathbb {E}}e^{-t(X-Y)} = K_1 \log (1 - p(1-e^{-t})) + K_2\log (1-q(1-e^{t})), \end{aligned}$$

where \(p = \alpha \frac{\log n}{n}\) and \(q= \beta \frac{\log n}{n}\). Using the elementary inequality, \(\log (1-x)\le -x\), valid for all \(0\le x \le 1\), we get

$$\begin{aligned} \psi (t) \ge -\varepsilon t + K_1p(1-e^{-t}) + K_2q(1-e^{t}). \end{aligned}$$

We pick \(t^{*} = \log ((2K_2q)^{-1}(-\varepsilon + \sqrt{\varepsilon ^2 + 4K_1K_2pq}))\) in order to optimize the right hand side. The second term in the right hand side becomes

$$\begin{aligned} K_1p(1-e^{-t^{*}})= K_1p \left( 1-\frac{2K_2q}{-\varepsilon + \sqrt{\varepsilon ^2 + 4K_1K_2pq}}\right) . \end{aligned}$$

We are interested in the behaviour of \(\psi (t^{*})\) when \(\varepsilon \rightarrow 0^{+}\), so we take the limit both sides in the equality above

$$\begin{aligned} \lim _{\varepsilon \rightarrow 0 ^{+}} K_1p(1-e^{-t^{*}}) = K_1p - \sqrt{K_1K_2pq}. \end{aligned}$$

Similarly, we get

$$\begin{aligned} \lim _{\varepsilon \rightarrow 0 ^{+}} K_2q(1-e^{t^{*}}) = K_2q - \sqrt{K_1K_2pq}. \end{aligned}$$

Now we can take the limit as \(\varepsilon \rightarrow 0^{+}\) in inequality 3 to obtain

$$\begin{aligned} {\mathbb {P}}(X-Y \le 0) \le e^{\lim _{\varepsilon \rightarrow 0 ^{+}}-\psi (t^{*})} \le e^{-(K_1p +K_2q -2\sqrt{K_1K_2pq})}. \end{aligned}$$

Recall that \(K_1 = \frac{n\gamma }{2} + o(n)\), \(K_2 = \frac{n\gamma }{2}+o(n)\), \(p = \alpha \frac{\log n}{n}\) and \(q= \beta \frac{\log n}{n}\). Then,

$$\begin{aligned} {\mathbb {P}}(X-Y \le 0) \le e^{-\log (n)(\gamma \frac{\alpha +\beta }{2}-\gamma \sqrt{\alpha \beta }+o(1))}. \end{aligned}$$

\(\square \)

We end this section with the proof of Proposition 2.

Proof

We denote the success event by \({\mathcal {E}}\), i.e, the event that the communities are recovered and we condition on the event that \(V^{\natural }\) has been drawn. By union bound we can write,

$$\begin{aligned} {\mathbb {P}}({\mathcal {E}}^{c} \mid V^{\natural }) \le P_1 + P_2. \end{aligned}$$

Here \(P_1:= \sum _{v \in S_1} \mathbbm {1}_{\{v \in V(G) {\backslash }V^{\natural }\}}{\mathbb {P}}(e(v,R_1) - e(v,R_{2})\le 0)\) and \(P_2\) is defined analogously.

Observe that now the probability in the right hand side of \(P_1\) is equal to

$$\begin{aligned} {\mathbb {P}}\left( \sum _{j=1}^{K_1}B_j^{(p)} - \sum _{j=1}^{K_{2}}B_j^{(q)}\le 0 \right) , \end{aligned}$$

where \(K_i = \mid R_i\mid \) and for all j, the random variables \(B_j^{p} \sim \text {Ber}(p)\) and \(B_j^{q} \sim \text {Ber}(q)\) are all independent. We set \(X:= \sum _{j=1}^{K_1}B_j^{(p)} \sim \text {Binom}(K_1, \alpha \frac{\log n}{n}) \) and \(Y:= \sum _{j=1}^{K_{2}}B_j^{(p)} \sim \text {Binom}(K_{2}, \beta \frac{\log n}{n})\). In order to apply Lemma 1, we denote the event in which both \(K_1\) and \(K_{2}\) lie in the interval \(\frac{n\gamma }{2}(1\pm \frac{1}{\sqrt{\log n}}) \) by \({\mathcal {A}}\). So we can bound \(P_1\) by

$$\begin{aligned} P_1 \le \sum _{v \in S_1} \mathbbm {1}_{\{v \in V(G) {\backslash }V^{\natural }\}}(\mathbbm {1}_{\{{\mathcal {A}}^{c}\}}+ \mathbbm {1}_{\{{\mathcal {A}}\}}{\mathbb {P}}(X-Y\le 0 \mid {\mathcal {A}})). \end{aligned}$$

We use the crude bound \(\mathbbm {1}_{\{v \in V(G) {\backslash } V^{\natural }\}} \le 1\) and write

$$\begin{aligned} P_1 \le \frac{n}{2}(\mathbbm {1}_{\{{\mathcal {A}}^{c}\}}+ \mathbbm {1}_{\{{\mathcal {A}}\}}{\mathbb {P}}(X-Y\le 0\mid {\mathcal {A}})). \end{aligned}$$

It is easy to see that the same bound holds for \(P_2\), so

$$\begin{aligned} {\mathbb {P}}({\mathcal {E}}^{c}\mid V^{\natural }) \le P_1+P_2\le n(\mathbbm {1}_{\{{\mathcal {A}}^{c}\}}+ \mathbbm {1}_{\{{\mathcal {A}}\}}{\mathbb {P}}(X-Y\le 0\mid {\mathcal {A}})). \end{aligned}$$

We take the expectation with respect to \(V^{\natural }\) both sides to obtain,

$$\begin{aligned} {\mathbb {P}}({\mathcal {E}}^{c})\le n({\mathbb {P}}({\mathcal {A}}^{c}) + {\mathbb {E}}_{V^{\natural }}{\mathbb {P}}(X-Y\le 0\mid {\mathcal {A}})). \end{aligned}$$
(4)

By Chernoff’s small deviation inequality (Exercise 2.3.5 [11]), there is an absolute constant \(c>0\) such that

$$\begin{aligned} {\mathbb {P}}({\mathcal {A}}^{c})\le 2{\mathbb {P}}\left( \mid K_1 - \frac{n\gamma }{2}\mid > \frac{n\gamma }{2\sqrt{\log n}}\right) \le 2e^{-c\frac{\gamma n}{\log n}} = o\left( \frac{1}{n}\right) . \end{aligned}$$
(5)

By Lemma 1,

$$\begin{aligned} {\mathbb {E}}_{V^{\natural }}{\mathbb {P}}(X-Y\le 0\mid {\mathcal {A}}) \le n^{- ((\alpha +\beta )\frac{\gamma }{2} -\gamma \sqrt{\alpha \beta }) + o(1)}. \end{aligned}$$
(6)

By the assumption on \(\gamma \), \((\alpha +\beta -2\sqrt{\alpha \beta })\frac{\gamma }{2}>1\). Therefore, there exists an \(\varepsilon >0\) such that

$$\begin{aligned} {\mathbb {P}}(X-Y\le 0\mid {\mathcal {A}}) \le n^{-1 -\varepsilon +o(1)} = o\left( \frac{1}{n}\right) . \end{aligned}$$

Then we combine inequalities 5 and 6 with inequality 4 to complete the proof. \(\square \)

3 Exact recovery in the subsampled nodes

In the sampling procedure in Step 1, the unknown communities \(S_1 \cap V^{\natural }\) and \(S_2 \cap V^{\natural }\) are no longer guaranteed to be balanced, therefore we cannot directly use the optimization program (2) because the maximum likelihood estimator is no longer (1). However, thanks to the authors in [8], similar semidefinite programs can be used to handle this case. We follow the approach in [8].

To begin with, it is straightforward to see that if the communities have sizes K and \(n-K\), the maximum likelihood estimator becomes

$$\begin{aligned} \begin{aligned} \max _{x} \quad&x^{T}Ax\\ \text{ s.t. } \quad&\mathbf {1}^{T}x = (2K-n)\\ {}&x\in \{\pm 1\}^n \\ \end{aligned} \end{aligned}$$
(7)

Therefore we can relax the problem in the same as before, we set \(X:=xx^{T}\) and write

$$\begin{aligned} \begin{aligned} \max _{X \in {\mathbb {R}}^{n\times n}} \quad&\mathrm {Tr}(AX)\\ \text {s.t.} \quad&X_{ii} = 1\\&X \succeq 0 \\&\mathrm {Tr}(X{\mathbf {J}}) = (2K-n)^2 \end{aligned} \end{aligned}$$
(8)

We should remark that the formulation (8) requires the knowledge of the sizes of the communities. To overcome this problem, we consider a Lagrangian formulation

$$\begin{aligned} \begin{aligned} \max _{X \in {\mathbb {R}}^{n\times n}} \quad&\mathrm {Tr}(AX) -\lambda ^{*}\mathrm {Tr}(X{\mathbf {J}})\\ \text {s.t.} \quad&X_{ii} = 1\\&X \succeq 0 \\ \end{aligned} \end{aligned}$$
(9)

The intuition is that the Lagrange multiplier \(\lambda ^{*}\) adjusts the sizes of the communities. An important insight from [8] is the following: There exists a value of \(\lambda ^{*}\) that works for all values K, so the optimization program (9) can be used to recover unbalanced communities with unknown sizes. Indeed, the following proposition reflects it. We use the notation \(G \sim {\mathcal {G}}(n_1,n_2,p,q)\) to denote a random graph drawn exactly in the same way as before with the exception that now the planted communities have sizes \(n_1\) and \(n_2\) satisfying \(n_1+n_2 =n\) but \(n_1\) is not necessarily equal to \(n_2\).

Proposition 3

[8] Let \(G \sim {\mathcal {G}}(K,n-K,p,q)\) with planted communities \(\{S_1,S_2\}\) and with \(p=\alpha \frac{\log n}{n}\) and \(q=\beta \frac{\log n}{n}\) satisfying \(p>q\). Then, for \(\sqrt{\alpha }-\sqrt{\beta } > \sqrt{2}\), the semidefinite program (9) with \(\lambda ^{*} = \left( \frac{\alpha -\beta }{\log \alpha -\log \beta } \right) \frac{\log n}{n}\) exactly recovers the communities with probability \(1-Cn^{-c(\frac{1}{2}(\sqrt{\alpha }-\sqrt{\beta })^2-1)}\), where \(C,c>0\) are absolute constants.

4 Main theorem

We shall proceed to the main result of this paper. We combine the ideas in Sects. 2 and 3 to establish a complete analysis of the sketching procedure.

Theorem 4

(Main result) Let \(G \sim {\mathcal {G}}(n;p,q)\) with planted communities \(\{S_1,S_2\}\) and with \(p=\alpha \frac{\log n}{n}\) and \(q=\beta \frac{\log n}{n}\) satisfying \(p>q\). Draw a vertex set \(V^{\natural }\) at random by sampling each node of the graph G independently at random with probability \(\gamma \). Denote, for \(i \in \{1,2\}\), \({\hat{R}}_i\) to be the maximum likelihood estimators of \(R_i = S_i \cap V^{\natural }\) obtained by running the semidefinite program 9 with the input matrix A being the adjacency matrix of the graph \(H \subset G\) induced by \(V^{\natural }\) and the parameter \(\lambda ^{*}\) chosen as follows: In the event that \(\mid V^{\natural }\mid \ge 2\), set \(\lambda ^{*} = \frac{\alpha _H - \beta _H}{\log \alpha _H - \log \beta _H}\frac{\log \mid V^{\natural }\mid }{\mid V^{\natural }\mid }\), where \(\alpha _{H}:= \frac{p \mid V^{\natural }\mid }{\log \mid V^{\natural }\mid }\) and \(\beta _{H}:= \frac{q\mid V^{\natural }\mid }{\log \mid V^{\natural }\mid }\), otherwise set \(\lambda ^{*}=0\). Now take

$$\begin{aligned} {\hat{S}}_1 = {\hat{R}}_1 \cup \{ v \in V(G){\backslash } V^{\natural }: e(v,{\hat{R}}_1)>e(v,{\hat{R}}_{2})\}. \\ {\hat{S}}_2 = {\hat{R}}_2 \cup \{ v \in V(G){\backslash } V^{\natural }: e(v,{\hat{R}}_2)>e(v,{\hat{R}}_{1})\}. \end{aligned}$$

Then there exists absolute constants \(C,c>0\) such that, with probability \(1-Cn^{-c((\alpha +\beta )\frac{\gamma }{2}-\gamma \sqrt{\alpha \beta }-1)}\), \(({\hat{S}}_1,{\hat{S}}_2)=(S_1,S_2)\). In particular, with probability \(1-o(1)\), \(({\hat{S}}_1,{\hat{S}}_2)=(S_1,S_2)\) as long as

$$\begin{aligned} \gamma > \frac{2}{\left( \sqrt{\alpha }-\sqrt{\beta }\right) ^2}. \end{aligned}$$

Proof

Observe that after sampling the vertex set V(G) of the graph, the induced subgraph \(H \subset G\) is a random graph with law \(H \sim {\mathcal {G}}(S_1\cap V^{\natural }, S_2\cap V^{\natural },p,q)\). We claim that there exists a \(\lambda ^{*}\) such that the optimization program 9 recovers both communities \(S_1\cap V^{\natural }\) and \(S_2\cap V^{\natural }\) with the desired probability. The proof of the theorem easily follows from the claim by applying Proposition 2 and union bound.

Now, we proceed to prove the claim. In order to apply Proposition 3 we need to check that, with sufficiently large probability,

$$\begin{aligned} \sqrt{\alpha _{H}}- \sqrt{\beta _{H}} > \sqrt{2}, \end{aligned}$$
(10)

where \(\alpha _{H}:= \frac{p \mid V^{\natural }\mid }{\log \mid V^{\natural }\mid }\) and \(\beta _{H}:= \frac{q\mid V^{\natural }\mid }{\log \mid V^{\natural }\mid }\) if \(\mid V^{\natural }\mid \ge 2\) and zero otherwise. Recall, by definition, \(p=\alpha \frac{\log n}{n}\) and \(q=\beta \frac{\log n}{n}\). The degenerate event \(\mid V^{\natural }\mid \le 1\) (empty set or single vertex) occurs with exponentially small probability. Indeed, observe \(\mid V^{\natural }\mid \) is a sum of n i.i.d random variables with Bernoulli distribution with mean \(\gamma \), so

$$\begin{aligned} {\mathbb {P}}(\mid V^{\natural }\mid \le 1) = (1-\gamma )^{n}+n(1-\gamma )^{n-1}\gamma \le 2e^{-\gamma (n-1)+\log n}, \end{aligned}$$

and

$$\begin{aligned} {\mathbb {P}}\left( \alpha _{H}= \alpha \frac{\mid V^{\natural }\mid \log n}{n\log \mid V^{\natural }\mid } \cap \mid V^{\natural }\mid \ge 2\right) = 1-2e^{-\gamma (n-1) +\log n}. \end{aligned}$$

An analogous fact holds for \(\beta _{H}\), so the event

$$\begin{aligned} \left\{ \sqrt{\alpha _H}-\sqrt{\beta _{H}}= \sqrt{\frac{\mid V^{\natural }\mid \log n}{n\log \mid V^{\natural }\mid }}\left( \sqrt{\alpha }-\sqrt{\beta }\right) \right\} \cap \left\{ \mid V^{\natural }\mid \ge 2\right\} , \end{aligned}$$

occurs with exponentially large probability. Since \(\frac{\log n}{\log \mid V^{\natural }\mid }\ge 1\) (when the quotient makes sense), it is enough to prove that, with the desired probability,

$$\begin{aligned} \left( \sqrt{\frac{\mid V^{\natural }\mid }{n}} \right) \left( \sqrt{\alpha }-\sqrt{\beta }\right) > \sqrt{2}. \end{aligned}$$

By assumption, there exists a \(\delta >0\) such that \(\sqrt{\alpha }-\sqrt{\beta }\ge \sqrt{\frac{2}{\gamma }}(1+\delta )\) and by the small Chernoff deviation inequality, for every \(\varepsilon >0\), \({\mathbb {P}}\left( \frac{\mid V^{\natural }\mid }{n}\ge \gamma -\varepsilon \right) \ge 1- 2e^{-c\varepsilon ^2n\gamma ^3}\). Putting these three facts together, we obtain, for every \(\varepsilon >0\),

$$\begin{aligned} \left( \sqrt{\frac{\mid V^{\natural }\mid \log n}{n\log \mid V^{\natural }\mid }} \right) \left( \sqrt{\alpha }-\sqrt{\beta }\right) \ge \sqrt{2}(1+\delta )\sqrt{1-\frac{\varepsilon }{\gamma }}, \end{aligned}$$

with exponentially large probability. We choose \(\varepsilon >0\) small enough to guarantee that \((1+\delta )\sqrt{1-\frac{\varepsilon }{\gamma }} >\sqrt{1+\delta }\) and then inequality (10) is satisfied with the desired probability. The claim now follows from Proposition 3. \(\square \)