Abstract
In this manuscript, we study the limiting distribution for the joint law of the largest and the smallest singular values for random circulant matrices with generating sequence given by independent and identically distributed random elements satisfying the so-called Lyapunov condition. Under an appropriated normalization, the joint law of the extremal singular values converges in distribution, as the matrix dimension tends to infinity, to an independent product of Rayleigh and Gumbel laws. The latter implies that a normalized \(\textit{condition number}\) converges in distribution to a Fréchet law as the dimension of the matrix increases.
1 Introduction
1.1 Singular values and condition number
The condition number was independently introduced by von Neumann and Goldstine in (Goldstine and von Neumann 1951; von Neumann et al. 1947), and by Turing in (Turing 1948) for studying the accuracy in the solution of a linear system in the presence of finite-precision arithmetic. Roughly speaking, the condition number measures how much the output value of a linear system can change by a small perturbation in the input argument, see (Smale 1985; Wo 1977) for further details.
Let \(\mathcal {A}\in \mathbb {C}^{n\times m}\) be a matrix of dimension \(n\times m\). We denote the singular values of \(\mathcal {A}\) in non-decreasing order by \(0\le \sigma ^{(n,m)}_{1}(\mathcal {A})\le \cdots \le \sigma ^{(n,m)}_{n}(\mathcal {A})\). That is to say, they are the square roots of the eigenvalues of the n-square matrix \(\mathcal {A}\mathcal {A}^*\), where \(\mathcal {A}^*\) denotes the conjugate transpose matrix of \(\mathcal {A}\). The condition number of \(\mathcal {A}\), \(\kappa (\mathcal {A})\), is defined as
It is known that
where
and \(\Vert \cdot \Vert\) denotes the operator norm. For \(n=m\), the smallest singular value \(\sigma ^{(n,n)}_{1}(\mathcal {A})\) measures the distance between the matrix \(\mathcal {A}\) and the set of singular matrices. We refer to Chapter 1 in (Bürgisser and Cucker 2013) for further details.
The extremal singular values are ubiquitous in the applications and have produced a vast literature in geometric functional analysis, mathematical physics, numerical linear algebra, time series, statistics, etc., see for instance (Bose and Saha 2018; Davis et al. 2016; Edelman 1989; Rider and Sinclair 2014; Rudelson and Vershynin 2010; Tao and Vu 2010) and Chapter 5 in (Vershynin 2012). The study of the largest and the smallest singular values has been very important in the study of sample correlation matrices, we refer to (Heiny and Mikosch 2017, 2018, 2019, 2021) for further details. Moreover, Poisson statistics for the largest eigenvalues in random matrices ensembles (Wigner ensembles for instance) are studied in (Soshnikov 2004, 2006).
Calculate or even estimate the condition number of a generic matrix is a difficult task, see (Sankar et al. 2006). In computational complexity theory, it is of interest to analyze the random condition number, that is, when the matrix \(\mathcal {A}\) given in (1) is a random matrix. In (Edelman 1988), it is computed the limiting law of the condition number of a random rectangular matrix with independent and identically distributed (i.i.d. for short) standard Gaussian entries. Moreover, the exact law of the condition number of a \(2\times n\) matrix is derived. In (Azaïs and Wschebor 2004), for real square random matrices with i.i.d. standard Gaussian entries, no asymptotic lower and upper bounds for the tail probability of the condition number are established. Later, in (Edelman and Sutton 2005), the results are generalized for non-square matrices and analytic expressions for the tail distribution of the condition number are obtained. Lower and upper bounds for the condition number (in p-norms) and the so-called average “loss of precision” are studied in (Szarek 1991) for real and complex square random matrices with i.i.d. standard Gaussian entries. In (Viswanath and Trefethen 1998), it is studied the case of random lower triangular matrices \(\mathcal {L}_n\) of dimension n with i.i.d. standard Gaussian random entries and shown that \((\kappa (\mathcal {L}_n))^{1/n}\) converges almost surely to 2 as n tends to infinity. In (Pérez et al. 2014), using a Coulomb fluid technique, it is derived asymptotics for the cumulative distribution function of the condition number for rectangular matrices with i.i.d. standard Gaussian entries. More recently, distributional properties of the condition number for random matrices with i.i.d. Gaussian entries are established in (Anderson and Wells 2009; Chen and Dongarra 2005; Shakil and Ahsanullah 2018) and large deviation asymptotics for condition numbers of sub-Gaussian distributions are given in (Singull et al. 2021). We recommend (Bürgisser and Cucker 2013; Cucker 2016; Demmel 1988; Edelman 1989) for a complete and current descriptions of condition number for random matrices.
1.2 Random circulant matrices
The (random) circulant matrices and (random) circulant type matrices are an important object in different areas of pure and applied mathematics, for instance compressed sensing, cryptography, discrete Fourier transform, extreme value analysis, information processing, machine learning, numerical analysis, spectral analysis, time series analysis, etc. For more details we refer to (Aldrovandi 2001; Bose et al. 2011, 2012; Davis 1994; Gray 2006; Kra and Simanca 2012; Rauhut 2009) and the monograph on random circulant matrices (Bose and Saha 2018). Some topics that have been studied are spectral norms, extremal distributions, the so-called limiting spectral distribution for random circulant matrices and random circulant-type matrices and process convergence of fluctuations, see (Bose et al. 2002, 2009, 2010, 2011a, b, 2012a, b, 2020).
Let \(\nu _0,\ldots ,\nu _{n-1}\in \mathbb {C}\) be any given complex numbers. We say that an \(n\times n\) complex matrix \(\text {circ}(\nu _0,\ldots ,\nu _{n-1})\) is circulant with generating elements \(\{\nu _0,\ldots ,\nu _{n-1}\}\) if it has the following structure:
Let \(\mathsf {i}\) be the imaginary unit and \(\omega _n=\exp (\mathsf {i}\cdot 2\pi /n)\) be a primitive n-th root of unity. Define the Fourier unitary matrix of order n by \(\mathcal {F}_n:=\frac{1}{\sqrt{n}}(\omega ^{kj}_n)_{k,j\in \{0,\ldots ,n-1\}}\). It is well-known that \(\text {circ}(\nu _0,\ldots ,\nu _{n-1})\) can be diagonalized as follows:
where \(\mathcal {D}_n:=\text {diag}(\lambda ^{(n)}_1,\ldots ,\lambda ^{(n)}_n)\) is a diagonal matrix with entries satisfying
By (2) we note that \((\lambda ^{(n)}_k)_{k\in \{0,\ldots ,n-1\}}\) are the eigenvalues of \(\text {circ}(\nu _0,\ldots ,\nu _{n-1})\). For \(\xi _0,\ldots ,\xi _{n-1}\) being random elements, we say that \(\mathcal {C}_n:=\text {circ}(\xi _0,\ldots ,\xi _{n-1})\) is an \(n\times n\) random circulant matrix. Then its eigenvalues are given by
Since any circulant matrix is a normal matrix, its singular values are given by \((|\lambda ^{(n)}_k|)_{k\in \{0,\ldots ,n-1\}}\), where the symbol \(|\cdot |\) denotes the complex modulus. Then the extremal singular values of \(\mathcal {C}_n\) can be written as
By (1) and (4) it follows that the random condition number of \(\mathcal {C}_n\) is given by
We stress that the random variables \(\sigma ^{(n)}_{\max }\) and \(\sigma ^{(n)}_{\min }\) are not independent. Thus, it is a challenging problem to compute (or estimate) the distribution of condition numbers of circulant matrices for general i.i.d. random coefficients \(\xi _0,\ldots ,\xi _{n-1}\).
1.3 Main results
The problem of computing the limiting distribution of the condition number for square matrices with i.i.d. (real or complex) standard Gaussian random entries is completely analyzed in Chapter 7 of (Edelman 1989). In this manuscript we focus on the computation of the limiting distribution of \(\kappa (\mathcal {C}_n)\) for \(\xi _0,\ldots ,\xi _{n-1}\) being i.i.d. real random variables satisfying the so-called Lyapunov condition, see Hypothesis 1.1 below. In fact, the limiting distribution is a Fréchet distribution that belongs to the class of the so-called extreme value distributions (Galambos 1987). Non-asymptotic estimates for the condition number for random circulant and Toeplitz matrices with i.i.d. standard Gaussian random entries are given in (Pan 2001, 2015). The approach and results of (Pan 2001, 2015) are different in nature from our results given in Theorem 1.2.
We assume the following integrability condition. It appears for instance in the so-called Lyapunov Central Limit Theorem, see Section 7.3.1 in (Ash 2000). Along this manuscript, the set of non-negative integers is denoted by \(\mathbb {N}_0\).
Hypothesis 1.1 (Lyapunov condition)
We assume that \((\xi _j)_{j\in \mathbb {N}_0}\) is a sequence of i.i.d. non-degenerate real random variables on some probability space \((\Omega ,\mathcal {F}, \mathbb {P})\) with zero mean and unit variance. If there exists \(\delta >0\) such that \(\mathbb {E}\left[ |\xi _0|^{2+\delta }\right] <\infty\), where \(\mathbb {E}\) denotes the expectation with respect to \(\mathbb {P}\), we say that \((\xi _j)_{j\in \mathbb {N}_0}\) satisfies the Lyapunov integrability condition.
We note that a sequence of i.i.d. non-degenerate sub-Gaussian random variables satisfies the Lyapunov condition. Before state the main result and its consequences, we introduce some notation. For shorthand and in a conscious abuse of notation, we use indistinctly the following notations for the exponential function: \(\exp (a)\) or \(e^{a}\) for \(a\in \mathbb {R}\). We denote by \(\ln (\cdot )\) the Napierian logarithm function and we use the same notation \(|\cdot |\) for the complex modulus and the absolute value.
The main result of this manuscript is the following.
Theorem 1.2 (Joint asymptotic distribution of the smallest and the largest singular values)
Assume that Hypothesis 1.1 is valid. Then it follows that
where
are the Rayleigh distribution and Gumbel distribution, respectively, and the normalizing constants are given by
The proof of Theorem 1.2 is based in the Davis–Mikosch method used to prove that a normalized maximum of the periodogram converges in distribution to the Gumbel law, see Theorem 2.1 in (Davis and Mikosch 1999). This method relies on Einmahl’s multivariate extension of the so-called Komlós–Major–Tusnády approximation, see (Einmahl 1989). However, the laws of the largest singular value \(\sigma ^{(n)}_{\max }\) and the smallest singular value \(\sigma ^{(n)}_{\min }\) are strongly correlated, and hence the computation of the condition number law is a priori hard. Thus, using the Davis–Mikosch method we compute the limiting law of the random vector
whose components are not independent. Applying the Continuous Mapping Theorem we deduce the limiting law of the condition number \(\kappa (\mathcal {C}_n)\), see Corollary 1.3 Item (iv). Along the lines of (Heiny and Mikosch 2021), one can obtain convergence of the point processes of the singular values in the setting of Theorem 1.2. We remark that (3) resembles the discrete Fourier transform and it is related with the so-called periodogram, which have been used in many areas of apply science. The maximum of the periodogram has been already studied for instance in (Davis and Mikosch 1999; Kokoszka and Mikosch 2000; Lin and Liu 2009; Turkman and Walker 1984) and under a suitable rescaling, the Gumbel distribution appears as a limiting law. The asymptotic behavior of Fourier transforms of stationary ergodic sequences with finite second moments is analyzed and shown that asymptotically the real part and the imaginary part of the Fourier transform decouple in a product of independent Gaussian distributions, see (Peligrad and Wu 2010). The so-called quenched central limit theorem for the discrete Fourier transform of a stationary ergodic process is obtained in (Barrera and Peligrad 2016) and the central limit theorem for discrete Fourier transforms of function times series are given in (Cerovecki and Hörmann 2017). In addition, in (Cerovecki et al. 2022) the maximum of the periodogram is studied in times series with values in a Hilbert space. However, up to our knowledge, Theorem 1.2 is not immediately implication of these results. As a consequence of Theorem 1.2 we obtain the limiting distribution of the (normalized) largest singular value, the smallest singular value and the (normalized) condition number as Corollary 1.3 states. For a \(n\times n\) symmetric random Toeplitz matrix satisfying Hypothesis 1.1, it is shown in (Sen and Virág 2013) that the largest eigenvalue scaled by \(\sqrt{2n\ln (n)}\) converges in \(L^{2+\delta }\) as \(n\rightarrow \infty\) to the constant \(\Vert \textsf {Sin}\Vert ^2_{2\rightarrow 4}=0.8288\ldots\).
We point out that Theorem 1 in (Bryc and Sethuraman 2009) yields that the Gumbel distribution is the limiting distribution for the (renormalized) largest singular value for symmetric circulant random matrices with generating i.i.d. sequence (half of its entries) satisfying Hypothesis 1.1. Also, for suitable normalization the limiting law of the largest singular value of Hermitian Gaussian circulant matrices has Gumbel distribution, see Corollary 5 in (Meckes 2009).
Recall that the square root of a Exponential distribution with parameter \(\lambda\) has Rayleigh distribution with parameter \((2\lambda )^{-1/2}\). The exponential law appears as the limiting distribution of the minimum modulus of trigonometric polynomials, see Theorem 1 in (Yakir and Zeitouni 2021) for the Gaussian case and Theorem 1.2 of (Cook and Nguyen 2021) for the sub-Gaussian case.
The Fréchet distribution appears as a limiting distribution of the largest eigenvalue (rescaled) for random real symmetric matrices with independent and heavy tailed entries, see Corollary 1 in (Auffinger et al. 2009) and (Basrak et al. 2021) for the non-i.i.d. case. The Fréchet distribution in Corollary 1.3 Item (iv) has cumulative distribution \(F(t)=\exp (-t^{-2}) \mathbbm {1}_{\lbrace t>0 \rbrace }\) with shape parameter 2, scale parameter 1 and location parameter 0. Moreover, it does not possess finite variance. For descriptions of extreme value distributions and limiting theorems we refer to (Galambos 1987).
Corollary 1.3 (Asymptotic distribution of the largest singular value, the smallest singular value and the condition number)
Let the notation and hypothesis of Theorem 1.2 be valid. The following holds.
-
(i)
The normalized maximum \(\frac{\sigma ^{(n)}_{\max }-\mathfrak {a}_n}{\mathfrak {b}_n}\) and the minimum \(\sigma ^{(n)}_{\min }\) are asymptotically independent.
-
(ii)
The normalized maximum \(\frac{\sigma ^{(n)}_{\max }-\mathfrak {a}_n}{\mathfrak {b}_n}\) converges in distribution as \(n\rightarrow \infty\) to a Gumbel distribution, i.e.,
$$\lim \limits _{n\rightarrow \infty }\mathbb {P}\left( \frac{\sigma ^{(n)}_{\max }-\mathfrak {a}_n}{\mathfrak {b}_n}\le y \right) =G(y) \quad \text { for any }\quad y\in \mathbb {R}.$$ -
(iii)
The minimum \(\sigma ^{(n)}_{\min }\) converges in distribution as \(n\rightarrow \infty\) to a Rayleigh distribution, i.e.,
$$\lim \limits _{n\rightarrow \infty }\mathbb {P}\left( \sigma ^{(n)}_{\min }\le x \right) =R(x) \quad \text { for any }\quad x\ge 0.$$ -
(iv)
The condition number \(\kappa (\mathcal {C}_n)\) converges in distribution as \(n\rightarrow \infty\) to a Fréchet distribution, i.e.,
$$\lim \limits _{n\rightarrow \infty } \mathbb {P}\left( \frac{\kappa (\mathcal {C}_n)}{\sqrt{\frac{1}{2}n\ln (n)}}\le z \right) =F(z)\quad \text { for any }\quad z>0,$$where \(F(z)=\exp \left( -z^{-2}\right) \mathbbm {1}_{\lbrace z>0 \rbrace }\).
In the sequel, we briefly compare our results with the literature about the limiting law of the condition number, the smallest singular value and the largest singular value.
Remark 1.4 (Frechet's distributions as limiting distributions for condition numbers)
The Fréchet distribution with shape parameter 2, scale parameter 2 and location parameter 0 is the limiting distribution as the dimension growths of \(\kappa (\mathcal {A}_n)/n\), where \(\mathcal {A}_n\) is a square matrix of dimension n with i.i.d. complex Gaussian entries, see Theorem 6.2 in (Edelman 1988). When \(\mathcal {A}_n\) has real i.i.d. Gaussian entries, the limiting law of \(\kappa (\mathcal {A}_n)/n\) converges in distribution as n tends to infinity to a random variable with an explicit density, see Theorem 6.1 in (Edelman 1988). We stress that such density does not belong to the Fréchet family of distributions. We also point out that the distribution of the so-called Demmel condition for (real and complex) Wishart matrices are given explicitly in (Edelman 1992).
Remark 1.5 (A word about the smallest singular value \(\sigma _1(\mathcal {A}_n)\))
The behavior of the smallest singular value appears naturaly in numerical inversion of large matrices. For instance, when the random matrix \(\mathcal {A}_n\) has complex i.i.d. Gaussian entries, for all n the random variable of \(n\sigma _1(\mathcal {A}_n)\) has the Chi-square distribution with two degrees of freedom, see Corollary 3.3 in (Edelman 1988). For a random matrix \(\mathcal {A}_n\) with real i.i.d. Gaussian entries, it is shown in Corollary 3.1 in (Edelman 1988) that \(n\sigma _1(\mathcal {A}_n)\)converges in distribution as the dimension increases to a random variable with an explicit density. For further discussion about the smallest singular values, we refer to (Bai and Yin 1993; Bose and Hachem 2020; Gregoratti and Maran 2021; Heiny and Mikosch 2018; Huang and Tikhomirov 2020; Kostlan 1988; Tao and Vu 2010; Tatarko 2018).
Remark 1.6 (A word about the largest singular value \(\sigma _n(\mathcal {A}_n)\))
A lot is known about the behavior of the largest singular value. As an illustration, for a random matrix \(\mathcal {A}_n\) with real i.i.d. Gaussian entries, it is shown in Lemma 4.1 in (Edelman 1988) that \((1/n)\sigma _n(\mathcal {A}_n)\) converges in probability to 4 as n growths while for the random matrix \(\mathcal {A}_n\) with complex i.i.d. Gaussian entries, \((1/n)\sigma _n(\mathcal {A}_n)\) converges in probability to 8. We stress that the Gumbel distribution is the limiting law of the spectral radius of Ginibre ensembles, see Theorem 1.1 in (Rider 2014). Recently, it is shown in Theorem 4 in (Arenas-Velilla and Pérez-Abreu 2021) that the Gumbel distribution is the limiting law of the largest eigenvalue of a Gaussian Laplacian matrix. For further discussion, we recommend to (Auffinger et al. 2009; Bai et al. 1988; Bryc and Sethuraman 2009; Cerovecki et al. 2022; Davis and Mikosch 1999; Heiny and Mikosch 2018; Meckes 2009; Soshnikov 2004, 2006).
We continue with the proof of Corollary 1.3.
Proof of Corollary 1.3
Items (i)-(iii) follow directly from Theorem 1.2 and the Continuous Mapping Theorem (see Theorem 13.25 in (Klenke 2014)). In the sequel, we prove Item (iv). By (5) we have
Theorem 1.2 with the help of the Continuous Mapping Theorem, limits
and the Slutsky Theorem (see Theorem 13.18 in (Klenke 2014)) implies
where \(\mathfrak {G}\) and \(\mathfrak {R}\) are random variables with Gumbel and Rayleigh distributions, respectively. The Slutsky Theorem with the help of (10) and (11) implies
Lemma 3.4 in Appendix A implies (iv) that \(\frac{\sqrt{2}}{\mathfrak {R}}\) possesses Fréchet distribution F. This finishes the proof of Item (iv).\(\square\)
Recently, the condition number for powers of matrices has been studied in (Huang and Tikhomirov 2020). As a consequence of Theorem 1.2 we have the following corollary which, in particular, gives the limiting distribution of the condition number for the powers of \(\mathcal {C}_n\).
Corollary 1.7 (Asymptotic distribution of p-th power of the maximum, the minimum and the condition number)
Let the notation and hypothesis of Theorem 1.2 be valid and take \(p\in \mathbb {N}\). The following holds.
-
(i)
Asymptotic distribution of the p-th power of the normalized maximum. For any \(y\in \mathbb {R}\) it follows that
$$\lim \limits _{n\rightarrow \infty }\mathbb {P}\left( \frac{\sigma ^{(n)}_{\max }(\mathcal {C}^p_n)-A_n(p)}{B_n(p)}\le y \right) =G\left( y\right) ,$$where \(A_n(p)=\mathfrak {a}^p_n\) and \(B_n(p)=p\mathfrak {b}^p_n(2^p-1)\) for all \(n\ge 3\).
-
(ii)
Asymptotic distribution of the p-th power of the minimum. For any \(x\ge 0\) it follows that
$$\lim \limits _{n\rightarrow \infty }\mathbb {P}\left( \sigma ^{(n)}_{\min }(\mathcal {C}^p_n)\le x \right) =R\left( x^{1/p}\right) .$$ -
(iii)
Asymptotic distribution of the p-th power of the condition number. For any \(z>0\) it follows that
$$\lim \limits _{n\rightarrow \infty } \mathbb {P}\left( \frac{\kappa (\mathcal {C}^p_n)}{(\frac{1}{2}n\ln (n))^{p/2}}\le z \right) =F\left( z^{1/p}\right) .$$
Proof
By (2) we have \(\mathcal {C}^p_n=\mathcal {F}^*_n \mathcal {D}^p_n \mathcal {F}_n\) for any \(p\in \mathbb {N}\). Therefore, \(\mathcal {C}^p_n\) is a normal matrix and then
Consequently, the statements (ii)-(iii) are direct consequences of Theorem 1.2 with the help of Lemma 3.3 in Appendix A and the Continuous Mapping Theorem.
In the sequel, we prove (i). For \(p=1\), Item (i) follows directly from Item (ii) of Corollary 1.3. For any \(p\in \mathbb {N}\setminus \{1\}\) set
Let \(y\in \mathbb {R}\) and observe that \(B_n y+A_n\rightarrow \infty\) as \(n\rightarrow \infty\) due to \(\mathfrak {b}_n/ \mathfrak {a}_n\rightarrow 0\) as \(n\rightarrow \infty\). Then for n large enough we have
We claim that
Indeed, notice that
where \(\delta _n:=\mathfrak {b}_n/ \mathfrak {a}_n\). Since \(\delta _n\rightarrow 0\) as \(n\rightarrow \infty\), L’Hôpital’s rule yields
Hence, (13) with the help of Lemma 3.3 in Appendix A implies
By (8), a straightforward computation yields \(B_n=B_n(p)=p\mathfrak {b}^p_n(2^p-1)\) for all \(n\ge 3\). This finishes the proof of Item (i).\(\square\)
The following proposition states the standard Gaussian case. It is an immediately consequence of Theorem 1.2. However, since the random variables are i.i.d. with standard Gaussian distribution, the computation of the law of (9) and its limiting distribution can be carried out explicitly. We decide to include an alternative and instructive proof in order to illustrate how we identify the right-hand side of (6).
Proposition 1.8 (Joint asymptotic distribution of the minimum and the maximum: the Gaussian case)
Let \((\xi _j)_{j\in \mathbb {N}_0}\) be a sequence of i.i.d. real Gaussian coefficients with zero mean and unit variance. Then it follows that
where the functions R and G are given in (7), and the normalizing constants \(\mathfrak {a}_n\), \(\mathfrak {b}_n\) are defined in (8).
Proof
Let \(q_n:=\lfloor n/2\rfloor\). For \(1\le k\le q_n\) we note that \(\lambda ^{(n)}_k = \overline{\lambda ^{(n)}_{n-k}}\), where \(\overline{\lambda }\) denotes the complex conjugate of a complex number \(\lambda\). Then we have
The complex modulus of \(\lambda ^{(n)}_k\) is given by \(|\lambda ^{(n)}_{k} |=\sqrt{c_{k,n}^2+s_{k,n}^2}\), where \(c_{k,n} := \sum _{j=0}^{n-1} \xi _j \cos \left( {2\pi \frac{kj}{n}}\right)\) and \(s_{k,n}:= \sum _{j=0}^{n-1} \xi _j \sin \left( {2\pi \frac{kj}{n}}\right)\) for \(0\le k \le q_n\). Note that \(s_{0,n}=0\). By straightforward computations we obtain for any \(n\in \mathbb {N}\)
-
(i)
\(\mathbb {E}\left( c_{k,n} \right) =\mathbb {E}\left( s_{k,n} \right) =0\) for \(0\le k\le q_n\),
-
(ii)
\(\mathbb {E}\left( c^2_{0,n} \right) =n\), \(\mathbb {E}\left( c_{k,n}^2 \right) =\mathbb {E}\left( s_{k,n}^2 \right) =\frac{n}{2}\) for \(1\le k\le q_n\),
-
(iii)
\(\mathbb {E}\left( c_{k,n}\cdot s_{l,n} \right) =\mathbb {E}\left( c_{k,n}\cdot c_{l,n} \right) =\mathbb {E}\left( s_{k,n}\cdot s_{l,n} \right) = 0\) for \(0\le l<k\le q_n\).
Then (i), (ii) and (iii) implies that \(\frac{1}{\sqrt{n}}\left( {c_{0,n},c_{1,n},s_{1,n},\ldots ,c_{q_n,n},s_{q_n,n}}\right)\) is a Gaussian vector such that its entries are not correlated, i.e., it has independent Gaussian entries. Thus, the random vector \(\frac{1}{n}\big (c_{0,n}^2,c^2_{1,n}+s^2_{1,n},\ldots ,c^2_{q_n,n}+s^2_{q_n,n}\big ) =\frac{1}{n}\big (|\lambda ^{(n)}_{0}|^2,|\lambda ^{(n)}_{1}|^2,\ldots ,|\lambda ^{(n)}_{q_n}|^2\big )\) has independent random entries satisfying
and
For n being an odd number, (19), (20) and Lemma 3.1 in Appendix A imply
We observe that \(\lim \limits _{n\rightarrow \infty } \frac{x^2}{n}=0\) and \(\lim \limits _{n\rightarrow \infty }\frac{(\mathfrak {b}_n y+\mathfrak {a}_n)^2}{n}=\infty\) for any \(x,y\in \mathbb {R}\). Recall that for any non-negative numbers u and v such that \(u<v\) it follows that \(\mathbb {P}\left( u<\textsf {E}_1\le v \right) =\exp (-u)-\exp (-v)\). Then, a straightforward calculation yields
Hence, for any x, y it follows that
The preceding limit with the help of (21) implies limit (17). Similar reasoning yields the proof when n is an even number. This finishes the proof of Proposition 1.8.\(\square\)
The rest of the manuscript is organized as follows. Section 2 is divided in five subsections. In Sect. 2.1 we prove that the asymptotic behavior of \(|\lambda ^{(n)}_0|\) can be removed from our computations. In Sect. 2.2 we establish that the original sequence of random variables can be assumed to be bounded. In Sect. 2.3 we provide a procedure in which we smooth (by a Gaussian perturbation) the bounded sequence obtained in Sect. 2.2. In Sect. 2.4 we prove the main result for the bounded and smooth sequence given in Sect. 2.3. Finally, in Sect. 2.5 we summarize all the results proved in previous subsections and show Theorem 1.2. Finally, there is Appendix 1 which collects technical results used in the main text.
2 Komlós–Major–Tusnády approximation
In this section, we show that Theorem 1.2 can be deduced by a Gaussian approximation in which computations can be carried out. Roughly speaking, we show that
where \(\text {o}_n(1)\rightarrow 0\) as \(n\rightarrow \infty\), \((\mathfrak {a}_n)_{n\in \mathbb {N}}\) and \((\mathfrak {b}_n)_{n\in \mathbb {N}}\) are the sequences defined in (8) in Theorem 1.2, and \(\sigma ^{(n,\mathcal {N})}_{\min }\) and \(\sigma ^{(n,\mathcal {N})}_{\max }\) denote the smallest and the largest singular values, respectively, of a random circulant matrix with generating sequence given by i.i.d. bounded and smooth random variables which can be well-approximated by a standard Gaussian distribution.
Along this section, for any set \(A\subset \Omega\), we denote its complement with respect to \(\Omega\) by \(A^{\text {c}}\). We also point out the following immediately relation:
2.1 Removing the singular value \(|\lambda ^{(n)}_0|\)
In this subsection we show that \(\left| \lambda ^{(n)}_0\right|\) can be removed from our computations. In other words, we only need to consider our computations over the array \(\left( {|\lambda ^{(n)}_k |}\right) _{k\in \{1,\ldots ,n-1\}}\) as the following lemma states.
Lemma 2.1 (Asymptotic behavior of \(|\lambda ^{(n)}_0|\) is negligible)
Assume that Hypothesis 1.1 is valid. Then for any \(x,y\in \mathbb {R}\) it follows that
Proof
For any \(n\in \mathbb {N}\) we set
Since \({A}\subset {B}\), we have
By (3) we have \(\lambda ^{(n)}_0=\sum _{j=0}^{n-1}\xi _j\). Since \(\xi _0,\ldots ,\xi _{n-1}\) are i.i.d. non-degenerate zero mean random variables with finite second moment, the Central Limit Theorem yields
where \(\text {N}(0,\mathbb {E}[|\xi _0|^{2})\) denotes the Gaussian distribution with zero mean and variance \(\mathbb {E}[|\xi _0|^{2}]\). We note that for any \(x,y\in \mathbb {R}\) the following limits holds
Hence (26), (27) and (28) with the help of Lemma 3.3 in Appendix 1 imply (25).\(\square\)
2.2 Reduction to the bounded case
In this subsection, we prove that it is enough to prove Theorem 1.2 for bounded random variables. That is, the random variables \((\xi _j)_{j\in \{0,\ldots ,n-1\}}\) in \(\lambda ^{(n)}_k = \sum _{j=0}^{n-1} \omega ^{kj}_n \xi _j\) can be considered bounded for all \(j\in \{0,\ldots ,n-1\}\). Following the spirit of Lemma 4 in (Bryc and Sethuraman 2009) we obtain the following comparison.
Lemma 2.2 (Truncation procedure)
Assume that Hypothesis 1.1 is valid. For each \(n\in \mathbb {N}\) and \(j\in \{0,\ldots n-1\}\) define the array of random variables \(\big (\widetilde{\xi }_j^{(n)}\big )_{j\in \{0,\ldots ,n-1\}}\) by
where \(s=2+\delta\) and \(\delta\) is the constant that appears in Hypothesis 1.1. For each \(k\in \{1,\ldots ,n-1\}\), set
Then it follows that
Proof
Let \(k\in \{1,\ldots ,n-1\}\) be fixed. Recall that \(\omega _n^{k} = \exp (\mathsf {i} k \cdot 2\pi /n)\). Note that \(w_n^{k} \ne 1\) and \(w_n^{kn}= 1\). Hence the geometric sum \(\sum _{j=0}^{n-1} \omega _n^{kj}=0\). Then
Indeed,
As a consequence of (32) and the triangle inequality we have the following estimate
Since \(\mathbb {E}\left( |\xi _0 |^s \right) <\infty\), we have \(\sum _{\ell =1}^\infty \mathbb {P}\left( |\xi _0 |^s > \ell \right) <\infty\), (see for instance Theorem 4.26 in (Klenke 2014)). Hence \(\sum _{\ell =1}^\infty \mathbb {P}\left( |\xi _0 | > \ell ^{1/s} \right) <\infty\). By the Borel–Cantelli Lemma (see Theorem 2.7 Item (i) in (Klenke 2014)), we have that
In other words, there exists an event \(\Omega ^*\) with \(\mathbb {P}\left( \Omega ^* \right) =1\) such that for each \(v\in \Omega ^*\) there is \(L(v)\in \mathbb {N}\) satisfying
Let \(v\in \Omega ^*\) and define \(n\ge 1+\max \lbrace L(v),|\xi _0(v) |^s,\ldots ,|\xi _{L(v)}(v) |^s \rbrace\). By (35) and the definition of n, we obtain
Combining (34) and (36) we deduce (31).\(\square\)
Next lemma allows us to replace the original array of random variables \(\left( {{\xi }_j^{(n)}}\right) _{j\in \{0,\ldots ,n-1\}}\) by array \((\widetilde{\xi }^{(n)}_j)_{j\in \{0,\ldots ,n-1\}}\) of bounded random variables defined in (29) of Lemma 2.2.
Lemma 2.3 (Reduction to the bound case)
Assume that Hypothesis 1.1 is valid. Let
where \((\widetilde{\lambda }^{(n)}_k)_{k\in \{1,\ldots , n-1\}}\) is defined in (30) of Lemma 2.2. Then for any \(x,y\in \mathbb {R}\) it follows that
Proof
For each \(n\in \mathbb {N}\) let
Lemma 2.2 implies that almost surely \(\mathbbm {1}_{A_n} - \mathbbm {1}_{B_n}\rightarrow 0\), as \(n\rightarrow \infty\). Hence, the Dominated Convergence Theorem yields
The preceding limit with the help of Lemma 2.1 and (24) and implies (38).\(\square\)
2.3 Smoothness
In this subsection we introduce a Gaussian perturbation in order to smooth the random variables \((\widetilde{\xi }^{(n)}_j)_{j\in \{1,\ldots ,n-1\}}\) defined in (29) of Lemma 2.2. Let \((N_j)_{j\in \{1,\ldots ,n-1\}}\) be an array of random variables with standard Gaussian distribution. Let \((\mathfrak {s}_n)_{n\in \mathbb {N}}\) be a deterministic sequence of positive numbers satisfying \(\mathfrak {s}_n\rightarrow 0\) as \(n\rightarrow \infty\) in a way that
This is precise in (52) below. We anticipate that \(\mathfrak {s}_n\approx n^{-\theta }\) for a suitable positive exponent \(\theta\). We define the array \((\gamma^{(n)}_k)_{k\in \{1,\ldots ,n-1\}}\) as follows
Then we set
where \((\widetilde{\lambda }^{(n)}_k)_{k\in \{1,\ldots ,n-1\}}\) is defined in (30) of Lemma 2.2.
Lemma 2.4
Assume that Hypothesis 1.1 is valid. Then for any \(x,y\in \mathbb {R}\) and a suitable sequence \((\mathfrak {s}_n)_{n\in \mathbb {N}}\) that tends to zero as \(n\rightarrow \infty\), it follows that
where \(\widetilde{\sigma }^{(n)}_{\min }\) and \(\widetilde{\sigma }^{(n)}_{\max }\) are given in (37) of Lemma 2.3 and R, G are defined in Theorem 1.2.
Proof
Let \(k\in \{1,\ldots ,n-1\}\). The triangle inequality yields
By (44) we have
where \(\beta _n\), \(\sigma ^{(n,\mathcal {N})}_{\min }\) and \(\sigma ^{(n,\mathcal {N})}_{\max }\) are defined in (42). We claim that
Indeed, by Proposition 1.8, Slutsky’s Theorem and (40) we have
in distribution, as \(n\rightarrow \infty\). The limit (47) can be also deduced from (1.1), p. 522 in (Davis and Mikosch 1999) or Chapter 10 in (Brockwell and Davis 1991).
Now, we have the necessary elements to conclude the proof of Lemma 2.4. By Lemma 2.5 we have
By (46) and (48) with the help of the Slutsky Theorem we deduce
The preceding limit with the help of (45) implies
As a consequence we conclude (43).\(\square\)
2.4 Bounded and smooth case
In the subsection we prove the following lemma.
Lemma 2.5 (Gaussian approximation for the bounded and smooth case)
Assume that Hypothesis 1.1 is valid. Let \(\beta _n\), \(\sigma ^{(n,\mathcal {N})}_{\min }\) and \(\sigma ^{(n,\mathcal {N})}_{\max }\) being as in (42). Then for any \(x,y\in \mathbb {R}\) and a suitable sequence \((\mathfrak {s}_n)_{n\in \mathbb {N}}\) that tends to zero as \(n\rightarrow \infty\), it follows that
where R and G are defined in Theorem 1.2.
We follow the ideas from (Davis and Mikosch 1999). To prove Lemma 2.5 we introduce some notation and the so-called Einmahl–Komlós–Major–Tusnády approximation. For each \(d\in \mathbb {N}\) and indexes \(i_j\in \{1,\ldots ,n-1\}\), \(j=1,\ldots ,d\) we define the Fourier frequencies \(w_{i_j} = \frac{2\pi i_j}{n}\), \(j=1,\ldots ,d\) and then the vector
where \({\text {T}}\) denotes the transpose operator. The next lemma is the main tool in the proof of Lemma 2.5. It allows us to reduce the problem to a perturbed Gaussian case.
Lemma 2.6 [Lemma 3.4 of (Davis 1999)]
Let \(d,n\in \mathbb {N}\) and denote by \(\widetilde{p}_n\) the continuous density function of the random vector
where \(\left( {N_\ell }\right) _{\ell\in \mathbb {N}_0}\) is a sequence of i.i.d. standard Gaussian random variables, independent of the sequence \(( \widetilde{\xi }_{\ell }^{(n)})_{\ell \in \{0,\ldots ,n-1\}}\) that is defined in (29) of Lemma 2.2, and \(s_n^2 = \text {Var}\left( {\widetilde{\xi }_0^{(n)}}\right) r_n^2\). If
for arbitrarily small \(\eta >0\), then the relation
holds uniformly for \(|x |^3=\text {o}^{(d)}_n \ {\left( \min \lbrace n^c,n^{1/2-1/(2+\delta )} \rbrace \right) }\), where the implicit constant in the \(\text {o}^{(d)}_n\)-notation depends on d, and \(\varphi _\Sigma\) is the density of a 2d-dimensional zero mean Gaussian vector with covariance matrix \(\Sigma\).
Proof of Lemma 2.5
Let \(x,y\in \mathbb {R}\). The idea is to apply Lemma 2.6 to the random sequence \(\left( \widetilde{\xi }_j^{(n)}+\mathfrak {s}_n N_j\right) _{j\in \mathbb {N}_0}\) with a suitable deterministic sequence \((\mathfrak {s}_n)_{n\in \mathbb {N}}\). We note that the variance of \(\widetilde{\xi }_0^{(n)}\), \(\text {Var}\left( {\widetilde{\xi }_0^{(n)}}\right)\), is bounded by \(\mathbb {E}[\xi ^2_0]=1\). Then we define
for sufficiently small \(\eta >0\). Since
where \(q_n:=\lfloor n/2\rfloor\) with \(\lfloor \cdot \rfloor\) being the floor function. Then we define
In what follows we compute the limit of (54) as \(n\rightarrow \infty\). In fact we prove that
For convenience, let
and observe that
By Lemma 3.2 in Appendix 1 for any fixed \(\ell \in \mathbb {N}\) we obtain
where
We claim that for every fixed \(d\in \mathbb {N}\) the following limit holds true
where the symbol \(\left( {\begin{array}{c}q_n\\ d\end{array}}\right)\) denotes the binomial coefficient. Indeed, by Lemma 2.6 we have
where \(I_{2d}\) denotes the \(2d\times 2d\) identity matrix, \(\varphi _{(1+\mathfrak {s}_n^2)I_{2d}}\) is the density of a 2d-dimensional Gaussian vector with zero mean and covariance matrix \((1+\mathfrak {s}_n^2)I_{2d}\), \(\text {o}_n(1)\rightarrow 0\) as \(n\rightarrow \infty\), and
By (8) and since \(x,y\in \mathbb {R}\) are fixed, there exists \(n_0=n_0(x,y)\) such that
Hence, (61) with the help of Lemma 3.5 in Appendix 1 yields
By (8), the Stirling formula (see formula (1) in (Robbins 1955)) and the fact that \(\mathfrak {s}_n\rightarrow 0\) as \(n\rightarrow \infty\) we deduce
As a consequence of (63) and (64) we obtain (60).
Now, we prove (55). By (58), (59), (60) and the fact that the events \((A^{(n)}_{k})_{k\in \{1,\ldots ,q_n\}}\) are independent, we have for any \(\ell \in \mathbb {N}\)
Sending \(\ell \rightarrow \infty\) in the preceding inequality yields
The preceding limit with the help of (57) implies (55).
Finally, by (53), (54) and (55) we obtain
where R and G are defined in (7).\(\square\)
2.5 Proof of Theorem 1.2
In the subsection, we stress the fact that Theorem 1.2 is just a consequence of what we have already stated up to here.
Proof of Theorem 1.2
Let \(x,y\in \mathbb {R}\). By (43) in Lemma 2.4 we have
where \(\widetilde{\sigma }^{(n)}_{\min }\) and \(\widetilde{\sigma }^{(n)}_{\max }\) are given in (37) of Lemma 2.3. The preceding limit with the help of (38) in Lemma 2.3 implies
This concludes the proof of Theorem 1.2.\(\square\)
Data availability
Data sharing not applicable to this article as no datasets were generated or analyzed during the current study.
References
Aldrovandi, R.: Special matrices of mathematical physics: stochastic, circulant and Bell matrices. World Scientific Publishing, (2001). https://doi.org/10.1142/4772
Anderson, W., Wells, M.: The exact distribution of the condition number of a Gaussian matrix. SIAM J. Matrix Anal. Appl. 31, no. 3, 1125-1130 (2009). https://doi.org/10.1137/070698932
Arenas-Velilla, S., Pérez-Abreu, V.: Extremal laws for Laplacian random matrices. ArXiv:2101.08318 (2021)
Ash, R.: Probability and measure theory. Second edition. With contributions by C. Doléans–Dade. Harcourt-Academic Press, Burlington, MA, (2000)
Auffinger, A., Ben Arous, G., Péché, S.: Poisson convergence for the largest eigenvalues of heavy tailed random matrices. Ann. Inst. Henri Poincaré Probab. Stat. 45, no. 3, 589-610 (2009). https://doi.org/10.1214/08-AIHP188
Azaïs, J., Wschebor, M.: Upper and lower bounds for the tails of the distribution of the condition number of a Gaussian matrix. SIAM J. Matrix Anal. Appl. 26, no. 2, 426-440 (2004). https://doi.org/10.1137/S0895479803429764
Bai, Z., Silverstein, J., Yin, Y.: A note on the largest eigenvalue of a large-dimensional sample covariance matrix. J. Multivar. Anal. 26, no. 2, 166-168 (1988). https://doi.org/10.1016/0047-259X(88)90078-4
Bai, Z., Yin, Y.: Limit of the smallest eigenvalue of a large dimensional sample covariance matrix. Ann. Probab. 21, no. 3, 1275-1294 (1993). https://doi.org/10.1214/aop/1176989118
Barrera, D., Peligrad, M.: Quenched limit theorems for Fourier transforms and periodogram. Bernoulli 22, no. 1, 275-301 (2016). https://doi.org/10.3150/14-BEJ658
Basrak, B., Cho, Y., Heiny, J., Jung, P.: Extreme eigenvalue statistics of \(m\)-dependent heavy-tailed matrices. Ann. Inst. Henri Poincaré Probab. Stat. 57, no. 4, 2100-2127 (2021). https://doi.org/10.1214/21-AIHP1152
Bose, A., Guha, S., Hazra, R., Saha, K.: Circulant type matrices with heavy tailed entries. Statist. Probab. Lett. 81, no. 11, 1706-1716 (2011). https://doi.org/10.1016/j.spl.2011.07.001
Bose, A., Hazra, R., Saha, K.: Poisson convergence of eigenvalues of circulant type matrices. Extremes 14, no. 4, 365-392 (2011a). https://doi.org/10.1007/s10687-010-0115-5
Bose, A., Hazra, R., Saha, K.: Spectral norm of circulant-type matrices. J. Theoret. Probab. 24, no. 2, 479-516 (2011b). https://doi.org/10.1007/s10959-009-0257-z
Bose, A., Hachem, W.: Smallest singular value and limit eigenvalue distribution of a class of non-Hermitian random matrices with statistical application. J. Multivariate Anal. 178, 104623, 24 pp (2020). https://doi.org/10.1016/j.jmva.2020.104623
Bose, A., Hazra, R., Saha, K.: Extremum of circulant type matrices: a survey. J. Indian Statist. Assoc. 50, no. 1-2, 21-49 (2012). http://repository.ias.ac.in/121167/
Bose, A., Hazra, R., Saha, K.: Product of exponentials and spectral radius of random \(k\)-circulants. Ann. Inst. Henri Poincaré Probab. Stat. 48, no. 2, 424-443 (2012a). https://doi.org/10.1214/10-AIHP404
Bose, A., Mitra, J., Sen, A.: Limiting spectral distribution of random \(k\)-circulants. J. Theoret. Probab. 25, no. 3, 771-797 (2012b). https://doi.org/10.1007/s10959-010-0312-9
Bose, A., Hazra, R., Saha, K.: Spectral norm of circulant type matrices with heavy tailed entries. Electron. Commun. Probab. 15, 299-313 (2010). https://doi.org/10.1214/ECP.v15-1554
Bose, A., Maurya, S., Saha, K.: Process convergence of fluctuations of linear eigenvalue statistics of random circulant matrices. Random Matrices Theory Appl. 2150032 (2020). https://doi.org/10.1142/S2010326321500325
Bose, A., Mitra, J.: Limiting spectral distribution of a special circulant. Statist. Probab. Lett. 60, no. 1, 111-120 (2002). https://doi.org/10.1016/S0167-7152(02)00289-4
Bose, A., Saha, K.: Random circulant matrices. Chapman and Hall/CRC Press, (2018). https://doi.org/10.1201/9780429435508
Bose, A., Subhra H., Saha, K.: Limiting spectral distribution of circulant type matrices with dependent inputs. Electron. J. Probab. 14, no. 86, 2463-2491 (2009). https://doi.org/10.1214/EJP.v14-714
Brockwell, P., Davis, R.: Time series: theory and methods. Springer-Verlag New York, (1991). https://doi.org/10.1007/978-1-4419-0320-4
Bryc, W., Sethuraman, S.: A remark on the maximum eigenvalue for circulant matrices. Inst. Math. Stat. (IMS) Collect. 5, 179-184 (2009). https://doi.org/10.1214/09-IMSCOLL512
Bürgisser, P., Cucker, F.: Condition: the geometry of numerical algorithms. Springer-Verlag Berlin Heidelberg, (2013). https://doi.org/10.1007/978-3-642-38896-5
Cerovecki, C., Characiejus, V., Hörmann, S.: The maximum of the periodogram of a sequence of functional data. J. Amer. Statist. Assoc. (2022). https://doi.org/10.1080/01621459.2022.2071720
Cerovecki, C., Hörmann, S.: On the CLT for discrete Fourier transforms of functional time series. J. Multivariate Anal. 154, 282-295 (2017). https://doi.org/10.1016/j.jmva.2016.11.006
Chen, Z., Dongarra, J.: Condition numbers of Gaussian random matrices. SIAM J. Matrix Anal. Appl. 27, no. 3, 603-620 (2005). https://doi.org/10.1137/040616413
Cook, N., Nguyen, H.: Universality of the minimum modulus for random trigonometric polynomials. Discrete Anal. 20, 46 pp (2021). https://doi.org/10.19086/da.28985
Cucker, F.: Probabilistic analyses of condition numbers. Acta Numer. 25, 321-382 (2016). https://doi.org/10.1017/S0962492916000027
Davis, P.: Circulant matrices. Amer. Math. Soc., (1994)
Davis, R., Mikosch, T.: The maximum of the periodogram of a non-Gaussian sequence. Ann. Probab. 27, no. 1, 522-536 (1999). https://doi.org/10.1214/aop/1022677270
Davis, R., Heiny, J., Mikosch, T., Xie, X.: Extreme value analysis for the sample autocovariance matrices of heavy-tailed multivariate time series. Extremes 19, no. 3, 517-547 (2016). https://doi.org/10.1007/s10687-016-0251-7
Demmel, J.: The probability that a numerical analysis problem is difficult. Math. Comput. 50, no. 182, 449-480 (1988). https://doi.org/10.2307/2008617
Edelman, A.: Eigenvalues and condition numbers of random matrices. Ph.D. dissertation and numerical analysis report 89-7, MIT, Cambridge, MA. Retrieved 2020.11.11, 21.05 h. EET zone (1989). http://math.mit.edu/~edelman/publications/eigenvalues_and_condition_numbers.pdf
Edelman, A.: Eigenvalues and condition numbers of random matrices. SIAM J. Matrix Anal. Appl. 9, no. 4, 543-560 (1988). https://doi.org/10.1137/0609045
Edelman, A.: On the distribution of a scaled condition number. Math. Comp. 58, no. 197, 185-190 (1992). https://doi.org/10.1090/S0025-5718-1992-1106966-2
Edelman, A., Sutton, B.: Tails of condition number distributions. SIAM J. Matrix Anal. Appl. 27, no. 2, 547-560 (2005). https://doi.org/10.1137/040614256
Einmahl, U.: Extensions of results of Komlós, Major, and Tusnády to the multivariate case. J. Multivariate Anal. 28, no. 1, 20-68 (1989). https://doi.org/10.1016/0047-259X(89)90097-3
Galambos, J.: The asymptotic theory of extreme order statistics, 2nd edn. Krieger Publishing Company, Florida (1987)
Goldstine, H., von Neumann, J.: Numerical inverting matrices of high order II. Proc. Amer. Math. Soc. 2, no. 2, 188-202 (1951). https://doi.org/10.2307/2032484
Gray, R.: Toeplitz and circulant matrices: A review. Found. Trends Commun. Inf. Theory 2, no. 3, (2006), 155-239. https://doi.org/10.1561/0100000006
Gregoratti, G., Maran, D.: Least singular value and condition number of a square random matrix with i.i.d. rows. Statist. Probab. Lett. 173, no. 109070, 7 pp (2021). https://doi.org/10.1016/j.spl.2021.109070
Heiny, J., Mikosch, T.: Almost sure convergence of the largest and smallest eigenvalues of high-dimensional sample correlation matrices. Stochastic Process. Appl. 128, no. 8, 2779-2815 (2018). https://doi.org/10.1016/j.spa.2017.10.002
Heiny, J., Mikosch, T.: Eigenvalues and eigenvectors of heavy-tailed sample covariance matrices with general growth rates: the iid case. Stochastic Process. Appl. 127, no. 7, 2179-2207 (2017). https://doi.org/10.1016/j.spa.2016.10.006
Heiny, J., Mikosch, T.: Large sample autocovariance matrices of linear processes with heavy tails. Stochastic Process. Appl. 141, 344-375 (2021). https://doi.org/10.1016/j.spa.2021.07.010
Heiny, J., Mikosch, T.: The eigenstructure of the sample covariance matrices of high-dimensional stochastic volatility models with heavy tails. Bernoulli 25, no. 4B, 3590-3622 (2019). https://doi.org/10.3150/18-BEJ1103
Heiny, J., Mikosch, T., Yslas, J.: Point process convergence for the off-diagonal entries of sample covariance matrices. Ann. Appl. Probab. 31, no. 2, 558-560 (2021). https://doi.org/10.1214/20-AAP1597
Huang, H., Tikhomirov, K.: A remark on the smallest singular value of powers of Gaussian matrices. Electron. Commun. Probab. 25, no. 10, 1-8 (2020). https://doi.org/10.1214/20-ECP285
Klenke, A.: Probability theory: A comprehensive course. Second edition, Springer-Verlag London, (2014). https://doi.org/10.1007/978-1-4471-5361-0
Kokoszka, P., Mikosch, T.: The periodogram at the Fourier frequencies. Stochastic Process. Appl. 86, no. 1, 49-79 (2000). https://doi.org/10.1016/S0304-4149(99)00086-1
Kostlan, E.: Complexity theory of numerical linear algebra. J. Comput. Appl. Math. 22, no. 2-3, 219-230 (1988). https://doi.org/10.1016/0377-0427(88)90402-5
Kra, I., Simanca, S.: On circulant matrices. Notices Amer. Math. Soc. 59, no. 3, 368-377 (2012). https://doi.org/10.1090/noti804
Lin, Z., Bai, Z.: Probability inequalities. Science Press Beijing, Beijing. Springer, Heidelberg, (2010). https://doi.org/10.1007/978-3-642-05261-3
Lin, Z., Liu, W.: On maxima of periodograms of stationary processes. Ann. Stat. 37, no. 5B, 2676-2695 (2009). https://doi.org/10.1214/08-AOS590
Meckes, M.: Some results on random circulant matrices. Inst. Math. Stat. Collections 5, 213-223 (2009). https://doi.org/10.1214/09-IMSCOLL514
Pan, V.: Structured matrices and polynomials: unified superfast algorithms. Birkhäuser Boston, (2001). https://doi.org/10.1007/978-1-4612-0129-8
Pan, V., Svadlenja, J., Zhao, L.: Estimating the norms of random circulant and Toeplitz matrices and their inverses. Linear Algebra Appl. 468, 197-210 (2015). https://doi.org/10.1016/j.laa.2014.06.027
Peligrad, M., Wu, W.: Central limit theorem for Fourier transforms of stationary processes. Ann. Probab. 38, no. 5, 2009-2022 (2010). https://doi.org/10.1214/10-AOP530
Pérez, I., Katzav, E., Vivo, P.: Phase transitions in the condition-number distribution of Gaussian random matrices. Phys. Rev. E 90, 050103-1–050103-5 (2014). https://doi.org/10.1103/PhysRevE.90.050103
Rauhut, H.: Circulant and Toeplitz matrices in compressed sensing. Proc. SPARS’09-Signal Processing with Adaptive Sparse Structured Representations, Saint-Malo, France, (2009). http://www.mathc.rwth-aachen.de/~rauhut/files/ToeplitzSPARS.pdf
Rider, B., Sinclair, C.: Extremal laws for the real Ginibre ensemble. Ann. Appl. Probab. 24, no. 4, 1621-1651 (2014). https://doi.org/10.1214/13-AAP958
Robbins, H.: A remark on Stirling’s formula. Amer. Math. Monthly 62, no. 1, 26-29 (1955). https://doi.org/10.2307/2308012
Rudelson, M., Vershynin, R.: Non-asymptotic theory of random matrices: extreme singular values. Proceedings of the International Congress of Mathematicians III, Hindustan Book Agency, New Delhi, 1576-1602 (2010). https://doi.org/10.1142/7920
Vershynin, R.: Compressed sensing theory and applications. Cambridge University Press, (2012). https://doi.org/10.1017/CBO9780511794308.006
Sankar, A., Spielman, D., Teng, S.: Smoothed analysis of the condition numbers and growth factors of matrices. SIAM J. Matrix Anal. Appl. 28, no. 2, 446-476 (2006). https://doi.org/10.1137/S0895479803436202
Sen, A., Virág, B.: The top eigenvalue of the random Toeplitz matrix and the sine kernel. Ann. Probab. 41, no. 6, 4050-4079 (2013). https://doi.org/10.1214/13-AOP863
Shakil, M., Ahsanullah, M.: A note on the characterizations of the distributions of the condition numbers of real Gaussian matrices. Spec. Matrices De Gruyter 6, no. 1, 282-296 (2018). https://doi.org/10.1515/spma-2018-0022
Singull, M., Uwamariya, D., Yang, X.: Large-deviation asymptotics of condition numbers of random matrices. J. Appl. Probab. 58, no. 4, 1114-1130 (2021). https://doi.org/10.1017/jpr.2021.13
Smale, S.: On the efficiency of algorithms of analysis. Bull. Amer. Math Soc. 13, no. 2, 87-121 (1985). https://doi.org/10.1090/S0273-0979-1985-15391-1
Soshnikov, A.: Poisson statistics for the largest eigenvalues of Wigner random matrices with heavy tails. Electron. Comm. Probab. 9, 82-91 (2004). https://doi.org/10.1214/ECP.v9-1112
Soshnikov, A.: Poisson statistics for the largest eigenvalues in random matrix ensembles. Mathematical physics of quantum mechanics 351-364, Lecture Notes in Phys. 690, Springer, Berlin, (2006). https://doi.org/10.1007/3-540-34273-7_26
Szarek, S.: Condition numbers of random matrices. J. Complexity 7, no. 2, 131-149 (1991). https://doi.org/10.1016/0885-064X(91)90002-F
Tao, T., Vu, V.: Random matrices: the distribution of the smallest singular values. Geom. Funct. Anal. 20, no. 1, 260-297 (2010). https://doi.org/10.1007/s00039-010-0057-8
Tatarko, K.: An upper bound on the smallest singular value of a square random matrix. J. Complexity 48, 119-128 (2018). https://doi.org/10.1016/j.jco.2018.06.002
Turing, A.: Rounding-off errors in matrix processes. Quart. J. Mech. Appl. Math. 1, no. 1, 287-308 (1948). https://doi.org/10.1093/qjmam/1.1.287
Turkman, K., Walker, M.: On the Asymptotic distributions of maxima of trigonometric polynomials with random coefficients. Adv. Appl. Probab. 16, no. 4, 819-842 (1984). https://doi.org/10.2307/1427342
Viswanath, D., Trefethen, L.: Condition numbers of random triangular matrices. SIAM J. Matrix Anal. Appl. 19, no. 2, 564-581 (1998). https://doi.org/10.1137/S0895479896312869
von Neumann, J., Goldstine, H.: Numerical inverting of matrices of high order. Bull. Amer. Math. Soc. 53, no. 11, 1021-1099 (1947). https://doi.org/10.1090/S0002-9904-1947-08909-6
Woźniakowski, H.: Numerical stability for solving nonlinear equations. Numer. Math. 27, 373-390 (1977). https://doi.org/10.1007/BF01399601
Yakir, O., Zeitouni, O.: The minimum modulus of Gaussian trigonometric polynomials. Israel J. Math. 1-24 (2021). https://doi.org/10.1007/s11856-021-2218-x
Acknowledgements
G. Barrera would like to express his gratitude to University of Helsinki, Department of Mathematics and Statistics, for all the facilities used along the realization of this work. The authors are grateful to the reviewers for the thorough examination of the paper, which has lead to a significant improvement.
Funding
Open Access funding provided by University of Helsinki including Helsinki University Central Hospital. The research of G. Barrera has been supported by the Academy of Finland, via the Matter and Materials Profi4 University Profiling Action, an Academy project (project No. 339228) and the Finnish Centre of Excellence in Randomness and STructures (project No. 346306).
Author information
Authors and Affiliations
Contributions
Both authors have contributed equally to the paper.
Corresponding author
Ethics declarations
Conflict of interests
The authors declare that they have no conflict of interest.
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendix A. Tools
Appendix A. Tools
The following section contains useful tools that help us to make this article more fluid. The following elementary lemma is crucial in the proof of Proposition 1.8.
Lemma 3.1
Let \(n\ge 2\) be a given integer number and let \(Y_1,\ldots ,Y_n\) be independent random variables. Define \(m^{(n)}:=\min \lbrace Y_1,\ldots ,Y_n \rbrace\) and \(M^{(n)}:=\max \lbrace Y_1,\ldots ,Y_n \rbrace\). Then
Since the proof of Lemma 3.1 is straightforward, we omit it.
Lemma 3.2 (Bonferroni's inequality)
Let \(\left( {\Omega ,\mathcal {F},\mathbb {P}}\right)\) be a probability space. Let \(A_1,\ldots ,A_n\) be events. Then for every \(\ell \in \mathbb {N}\),
where
The proof of Lemma 3.2 is given in Section 1.1 “Inclusion-Exclusion Formula” of (Lin and Bai 2010).
Lemma 3.3 (Continuity)
Let \((\Omega ,\mathcal {F}, \mathbb {P})\) be a probability space. Let \((X_n)_{n\in \mathbb {N}}\) be a sequence of random variables defined on \(\Omega\) and taking values in \(\mathbb {R}\). Assume that \(X_n\) converges in distribution to a random variable X, as \(n\rightarrow \infty\). Let x be a continuity point of the distribution function \(F_X\) of the random variable X and let \((a_n(x))_{n\in \mathbb {N}}\) be a deterministic sequence of real numbers such that \(a_n(x)\rightarrow x\) as \(n\rightarrow \infty\). Then
In addition, if \(F_X\) is a continuous function then
-
(i)
for any deterministic sequence \((a_n)_{n\in \mathbb {N}}\) such that \(a_n\rightarrow 0\) as \(n\rightarrow \infty\) it follows that
$$\begin{aligned} \lim \limits _{n\rightarrow \infty }\mathbb {P}(|X_n|\le a_n)= 0. \end{aligned}$$(71) -
(ii)
for any deterministic sequence \((a_n)_{n\in \mathbb {N}}\) satisfying \(a_n\rightarrow \infty\) as \(n\rightarrow \infty\) it follows that
$$\begin{aligned} \lim \limits _{n\rightarrow \infty }\mathbb {P}(|X_n|> a_n)= 0. \end{aligned}$$(72)
Proof
We start with the proof of (70). Let x be a continuity point of \(F_X\) and take \(\epsilon >0\). Then there exists \(n_\epsilon :=n_\epsilon (x)\) such that \(x-\epsilon<a_n(x)<x+\epsilon\) for all \(n\ge n_\epsilon\). By monotonicity we have
Hence, the Portmanteau Theorem (see Theorem 13.16 in (Klenke 2014)) implies for any \(\epsilon >0\)
Since x is a continuity point of \(F_X\), sending \(\epsilon \rightarrow 0\) we deduce (70).
We continue with the proof of Item (i) and Item (ii). By the Continuous Mapping Theorem we have \(|X_n|\rightarrow |X|\) in distribution, as \(n\rightarrow \infty\). On the one hand, (70) yields
which finishes the proof of Item (i).
On the other hand, let \(m>0\) be arbitrary. Then there exists \(n_{m}\in \mathbb {N}\) such that \(a_n>m\) for all \(n\ge n_m\). Hence,
which implies Item (ii) sending \(m\uparrow \infty\).\(\square\)
Lemma 3.4 (Fréchet distribution as the inverse of a scaled Rayleigh distribution)
Let X be a real random variable with the Rayleigh distribution \(R(x)=(1-\exp \left( {-x^2/2}\right) )\mathbbm {1}_{\lbrace x\ge 0 \rbrace }\). Let Y be a real random variable with the Fréchet distribution \(F(y)=\exp (-y^{-2})\mathbbm {1}_{\lbrace y>0 \rbrace }\). Then \(\sqrt{2}/X{\mathop {=}\limits ^{d}}Y\), where the symbol \({\mathop {=}\limits ^{d}}\) denotes equality in distribution.
Proof
Let \(y>0\) and note that
The preceding equality concludes the statement.\(\square\)
Lemma 3.5 (Exponential distribution as a sum of square independent Gaussian distributions)
Let \(X_1\) and \(X_2\) be two independent random variables with Gaussian distribution with zero mean and variance \(\sigma ^2>0\). Then \(X^2_1+X^2_2\) has Exponential distribution with parameter \(1/(2\sigma ^2)\), that is,
Proof
Since the random variables \(X_j/\sigma\), \(j=1,2\) have standard Gaussian distributions, the random variables \(X^2_j/\sigma ^2\), \(j=1,2\) have Chi-square distribution with one degree of freedom. Due to the independence assumption, we have that \(X^2_1/\sigma ^2+X^2_2/\sigma ^2\) has Chi-square distribution with two degrees of freedom. That is, for any \(x\ge 0\), we have \(\mathbb {P}\left( X^2_1/\sigma ^2+X^2_2/\sigma ^2\le x\right) =1-e^{-x/2}\). The preceding relation yields \(\mathbb {P}\left( X^2_1+X^2_2\le x\right) =1-\exp {\left( -\frac{x}{2\sigma ^2}\right) }\) for any \(x\ge 0\).\(\square\)
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Barrera, G., Manrique-Mirón, P. The asymptotic distribution of the condition number for random circulant matrices. Extremes 25, 567–594 (2022). https://doi.org/10.1007/s10687-022-00442-w
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10687-022-00442-w
Keywords
- Circulant random matrices
- Condition number
- Fréchet distribution
- Gumbel distribution
- Lyapunov integrability condition
- Rayleigh distribution
AMS 2000 Subject Classifications
- Primary—60F05
- 60G70
- 65F35
- Secondary—60G10
- 62E20
- 65F22