Abstract
We use the H-matrix technology to compute the approximate square root of a covariance matrix in linear cost. This allows us to generate normal and log-normal random fields on general point sets with optimal cost. We derive rigorous error estimates which show convergence of the method. Our approach requires only mild assumptions on the covariance function and on the point set. Therefore, it might be also a nice alternative to the circulant embedding approach which applies only to regular grids and stationary covariance functions.
Similar content being viewed by others
Change history
21 March 2019
The corrected version states the parameter range as p���������2��� instead of p������������. The effect is to disallow non-smooth norms such as the ���1-norm for the distance measure.
21 March 2019
The corrected version states the parameter range as p���������2��� instead of p������������. The effect is to disallow non-smooth norms such as the ���1-norm for the distance measure.
References
Babuška, I., Andersson, B., Smith, P.J., Levin, K.: Damage analysis of fiber composites. I. Statistical analysis on fiber scale. Comput. Methods Appl. Mech. Eng. 172(1–4), 27–77 (1999)
Bernardo, J.M., Berger, J.O., Dawid, A.P., Smith, A.F.M. (eds.): Bayesian Statistics 6, Proceedings of the 6th Valencia International Meeting held in Alcoceber, 6–10 June 1998. The Clarendon Press, Oxford University Press, New York, pp. x+867 (1999)
Börm, S.: Efficient Numerical Methods for Non-local Operators. Volume 14 of EMS Tracts in Mathematics. European Mathematical Society (EMS), Zürich (2010)
Chan, G., Wood, A.T.A.: Algorithm as 312: an algorithm for simulating stationary gaussian random fields. J. R. Stat. Soc. Ser. C (Appl. Stat.) 46(1), 171–181 (1997)
Dietrich, C.R., Newsam, G.N.: Fast and exact simulation of stationary Gaussian processes through circulant embedding of the covariance matrix. SIAM J. Sci. Comput. 18(4), 1088–1107 (1997)
Dölz, J., Harbrecht, H., Schwab, Ch.: Covariance regularity and h-matrix approximation for rough random fields. Numerische Mathematik, pp. 1–27 (2016)
Elishakoff, I. (ed): Whys and Hows in Uncertainty Modelling. In: Probability, Fuzziness and Anti-optimization, Volume 388 of CISM Courses and Lectures. Springer, Vienna (1999)
Frommer, A.: Monotone convergence of the Lanczos approximations to matrix functions of Hermitian matrices. Electron. Trans. Numer. Anal. 35, 118–128 (2009)
Graham, I.G., Kuo, F.Y., Nuyens, D., Scheichl, R., Sloan, I.H.: Quasi-Monte Carlo methods for elliptic PDEs with random coefficients and applications. J. Comput. Phys. 230(10), 3668–3694 (2011)
Grasedyck, L., Hackbusch, W.: Construction and arithmetics of \(H\)-matrices. Computing 70(4), 295–334 (2003)
Hackbusch, Wolfgang: Hierarchical Matrices: Algorithms and Analysis. Volume 49 of Springer Series in Computational Mathematics. Springer, Heidelberg (2015)
Harbrecht, H., Peters, M., Siebenmorgen, M.: Efficient approximation of random fields for numerical applications. Numer. Linear Algebra Appl. 22(4), 596–617 (2015)
Higham, N.J.: Computing real square roots of a real matrix. Linear Algebra Appl. 88(89), 405–430 (1987)
Higham, N.J.: Stable iterations for the matrix square root. Numer. Algorithms 15(2), 227–242 (1997)
Kenney, C., Laub, A.J.: Rational iterative methods for the matrix sign function. SIAM J. Matrix Anal. Appl. 12(2), 273–291 (1991)
Khoromskij, B.N., Litvinenko, A., Matthies, H.G.: Application of hierarchical matrices for computing the Karhunen–Loève expansion. Computing 84(1–2), 49–67 (2009)
Moret, I.: Rational Lanczos approximations to the matrix square root and related functions. Numer. Linear Algebra Appl. 16(6), 431–445 (2009)
Schmitt, B.A.: Perturbation bounds for matrix square roots and pythagorean sums. Linear Algebra Appl. 174, 215–227 (1992)
Author information
Authors and Affiliations
Corresponding author
Appendices
Appendix—Proof of Lemma 1
The following lemma is an elementary statement on holomorphic functions.
Lemma 9
Let \(f:O\rightarrow {{\mathbb {C}}}\) be a continuous function on the domain \(O\subset {{\mathbb {C}}}^n\) which is holomorphic in O in all variables \({\varvec{x}}_i\), \(i\in \{1,\ldots ,n\}\), i.e.,
is holomorphic in \(\big \{{\varvec{x}}_i\in {{\mathbb {C}}}\,:\,({\varvec{x}}_1,\ldots ,{\varvec{x}}_i,\ldots ,{\varvec{x}}_n)\in O\big \}\) for all \({\varvec{x}}_1,\ldots ,{\varvec{x}}_{i-1},{\varvec{x}}_{i+1},\ldots ,{\varvec{x}}_n\in {{\mathbb {C}}}\). Then, for all multi-indices \(\alpha \in {{\mathbb {N}}}_0^n\), the function \(\partial _{\varvec{x}}^\alpha f\) is holomorphic in O in all variables \({\varvec{x}}_i\), \(i\in \{1,\ldots ,n\}\) as defined above.
Proof
The result is proved by induction on \(|\alpha |_1\). Obviously, for \(|\alpha |_1=0\), \(\partial _{\varvec{x}}^\alpha f=f\) and the statement is true. Assume the statement holds for all \(|\alpha |_1\le k\) and choose some \(\alpha \in {{\mathbb {N}}}_0^n\) with \(|\alpha |_1=k+1\). Then, we have for some \(i\in \{1,\ldots ,n\}\) and some \(\alpha _0\in {{\mathbb {N}}}_0^n\) with \(|\alpha _0|_1=k\) that
Since, \(\partial _{\varvec{x}}^{\alpha _0}f\) is holomorphic in O in all variables by the induction hypothesis, obviously \(\partial _{\varvec{x}}^\alpha f\) is holomorphic in O at least in \({\varvec{x}}_i\) (derivatives of holomorphic functions are holomorphic). To prove the statement for all other variables, we may employ Cauchy’s integral formula to obtain
for some \(\varepsilon >0\) with \(B_\varepsilon ({\varvec{x}}_i)\subset {{\mathbb {C}}}\) being the ball with radius \(\varepsilon \). The integrand is holomorphic in all variables \({\varvec{x}}_j\), \(j\ne i\). Hence, we conclude that \(\partial _{\varvec{x}}^\alpha f({\varvec{x}})\) is holomorphic in all variables and prove the assertion. \(\square \)
The following result is elementary but technical.
Lemma 10
For \(n,p\in {{\mathbb {N}}}\), define the set \(M:=\big \{{\varvec{x}}\in {{\mathbb {C}}}^{n}\,:\,\mathrm{real}(\sum _{i=1}^n {\varvec{x}}_i^p) \le 0\big \}\). Then, there holds \(({{\mathbb {R}}}^n)_+:=\big \{{\varvec{x}}\in {{\mathbb {R}}}^n{\setminus }\{0\}\,:\,{\varvec{x}}_i\ge 0\big \}\cap M=\emptyset \) and
Proof
Let \({\varvec{x}}\in ({{\mathbb {R}}}^n)_+\), then we have \(\sum _{i=1}^n {\varvec{x}}_i^p>0\) and hence \({\varvec{x}}\notin M\). It is easy to see that the cone \(C_p:=\big \{r\exp (i\phi )\,:\,r>0,\,\phi \in (-\frac{\pi }{2p},\frac{\pi }{2p})\big \}\subset {{\mathbb {C}}}\) satisfies \(\mathrm{real}(x^p)>0\) for all \(x\in C_p\). Thus, we have that
satisfies \(C_p^n\cap M=\emptyset \).
Moreover, a simple geometric argument (see Figure 6) shows that all \(x>0\) satisfy
Since \(({{\mathbb {R}}}^n)_+\subseteq C_p^n\), this implies
This concludes the proof. \(\square \)
Products of asymptotically smooth functions are again asymptotically smooth. This is shown in the next lemma.
Lemma 11
Given two functions \(f,g:D\times D\rightarrow {{\mathbb {R}}}\) which are asymptotically smooth (1). Then, also their product fg satisfies (1).
Proof
To simplify the notation, we consider f, g as functions of one variable \({\varvec{z}}=({\varvec{x}},{\varvec{y}})\in D\times D\subset {{\mathbb {R}}}^{2d}\). For multi-indices \(\alpha ,\beta \in {{\mathbb {N}}}^{2d}\), define
Note that there holds \(\left( {\begin{array}{c}\alpha \\ \beta \end{array}}\right) \le \left( {\begin{array}{c}|\alpha |_1\\ |\beta |_1\end{array}}\right) \). This follows from the basic combinatorial fact that the number of possible choices of \(\beta _i\) elements out of a set of \(\alpha _i\) elements for all \(i=1,\ldots ,2d\) is smaller than the number of choices of \(|\beta |_1\) elements out of a set of \(|\alpha |_1\) elements.
The Leibniz formula together with the definition of asymptotically smooth function (1) show for \(\alpha \in {{\mathbb {N}}}^{2d}\)
where we used \((|\alpha |_1+1)^{2d}\le (2d\exp (2d))^{|\alpha |_1}\) and \({\widetilde{c}}_2=c_2/(2d\exp (2d))\). This concludes the proof.
The final lemma of this section proves the concatenations of certain asymptotically smooth functions are asymptotically smooth.
Lemma 12
Let \(g:D\times D\rightarrow {{\mathbb {R}}}\) be asymptotically smooth (1) with constants \(c_1,c_2>0\).
-
(i)
If \(c_g:=\sup _{{\varvec{x}}\in D\times D}g({\varvec{x}})<\infty \). Then, \(\exp \circ g\) satisfies (1) with constants \(\widetilde{c}_1:= \exp (c_g)\) and \({\widetilde{c}}_2:= c_2/(2\max \{1,c_1\})\).
-
(ii)
If g satisfies \(\partial _{\varvec{x}}^\alpha \partial _{\varvec{y}}^\alpha g({\varvec{x}},{\varvec{y}})\le C_g\) for all \(\alpha ,\beta \in {{\mathbb {N}}}_0^d\) and some \(C_g<\infty \) as well as \(g({\varvec{x}},{\varvec{y}})\ge C_g^{-1}|{\varvec{x}}-{\varvec{y}}|\), then, \(g^{1/q}\) satisfies (1) with \({{\widetilde{\varrho }}}_1=1/2\) and \({{\widetilde{\varrho }}}_2 =C_g^{-1}\) for all \(q\in {{\mathbb {N}}}\).
-
(iii)
If g satisfies the assumptions from (ii) and additionally \(g({\varvec{x}},{\varvec{y}})\ge c_0>0\) for all \({\varvec{x}},{\varvec{y}}\in D\), then \(g^{-1/q}\) satisfies (1) for all \(q\in {{\mathbb {N}}}\).
Proof
To simplify the notation, we consider g as a function of one variable \({\varvec{z}}=({\varvec{x}},{\varvec{y}})\in D\times D\subset {{\mathbb {R}}}^{2d}\). Define the set of all partitions of \(\{1,\ldots ,n\}\) as
For a multi-index \(\alpha \in {{\mathbb {N}}}^{2d}\), we define \(\widetilde{\alpha }\in \{1,\ldots ,2d\}^n\) by \({{\widetilde{\alpha }}}_i=j\) for all \(1+\sum _{k=1}^{j-1}\alpha _k\le i\le \sum _{k=1}^j\alpha _k\) and all \(1\le j\le 2d\) (e.g., \(\alpha =(2,3,1,1)\) yields \(\widetilde{\alpha }=(1,1,2,2,2,3,4)\)). With \(n=|\alpha |_1\) and some \(S\in P\in \varPi (n)\), we define
(the definition implies \(\partial _{{\varvec{z}}}^{\{1,\ldots ,n\}}g({\varvec{z}})=\partial _{\varvec{z}}^\alpha g({\varvec{z}})\).) With those definitions and given a function \(f:{{\mathbb {R}}}\rightarrow {{\mathbb {R}}}\), Faà di Bruno’s formula reads for a multi-index \(\alpha \in {{\mathbb {N}}}^{2d}\)
For (i), Faà di Bruno’s formula (37) and \(\partial _x^{|P|}\exp =\exp \) show for all multi indices \(\alpha \in {{\mathbb {N}}}^{2d}\) with \(n=|\alpha |_1\) that
The definition of asymptotically smooth (1) and \(\Vert g\Vert _{L^\infty (D\times D)}=c_g\) imply
With \(f(x):=(1-x)^{-1}\), \(x\in {{\mathbb {R}}}{\setminus }\{1\}\), we have \(\partial _x^k f(x) = k!(1-x)^{-1-k}\). Hence, the last factor can be written, using Faà di Bruno’s formula again, as
As the function \(h(x):=\exp ((1-x)^{-1})\), \(x\in {{\mathbb {C}}}\) is holomorphic at least for \(|x|<1\), Cauchy’s integral formula shows
Altogether, we conclude the proof of (i) by
For (ii), Faà di Bruno’s formula (37) shows again for \(q>1\)
where we used \(f(x):=x^{1/q}\) and \(|\partial _x^{|P|}f(x)|=|(1/q)(1/q-1)(1/q-2)\cdots (1/q-|P|+1)||x|^{1/q-|P|}\le |P|!|x|^{1/q-|P|}\) as well as the boundedness assumption on the derivatives of g from (ii). With \(r(x):=\exp (x)-1\) and \(f(x):=(1-x)^{-1}\), \(x\in {{\mathbb {R}}}\), the last factor satisfies
The function \(h(x):=f\circ r(x)= (2-\exp (x))^{-1}\), \(x\in {{\mathbb {C}}}\) is holomorphic at least for \(|x|\le 1/2\). As above, this implies
and thus concludes the proof of (ii).
For (iii), we conclude the proof as for (ii) by use of the estimate \(g(z)^{{-}1/q-|P|}\le c_0^{-1-n}\).
At last, we are ready to prove Lemma 1 which states that the covariance functions from (2) and (3) are asymptotically smooth (1).
Proof of Lemma 1
To see (1), consider \(\varrho (\cdot ,\cdot )\) from (2). We define for complex variables \({\varvec{x}}_i,{\varvec{y}}_i\in {{\mathbb {C}}}\)
whenever \((\cdot )^{1/p}\) is defined in \({{\mathbb {C}}}\). and consider \({\widetilde{\varrho }}({\varvec{x}},{\varvec{y}})\) which is \(\varrho ({\varvec{x}},{\varvec{y}})\) from (2) but with \(d({\varvec{x}}-{\varvec{y}})\) instead of \(|{\varvec{x}}-{\varvec{y}}|_p\). With the notation of Lemma 10, the above sum has positive real part in \(O:=\big \{({\varvec{x}},{\varvec{y}})\in {{\mathbb {C}}}^{2d}\,:\,{\varvec{x}}-{\varvec{y}}\notin M\big \}\). Thus, the function \(({\varvec{x}},{\varvec{y}})\mapsto d({\varvec{x}}-{\varvec{y}})\) is holomorphic in each variable in O. Since for \(a>0\), \({\varvec{x}}\mapsto {\varvec{x}}^\mu K_\mu (ax)\) is a holomorphic function on \({{\mathbb {C}}}{\setminus }({{\mathbb {R}}}_-\cup \{0\})\), and \(d({\varvec{x}}-{\varvec{y}})\) has positive real part, we deduce that \(({\varvec{x}},{\varvec{y}})\mapsto {\widetilde{\varrho }}({\varvec{x}},{\varvec{y}})\) is holomorphic in each variable in O. Thus, Lemma 9 proves that \(\partial _{\varvec{x}}^\alpha \partial _{\varvec{y}}^\beta {\widetilde{\varrho }}({\varvec{x}},{\varvec{y}})\) is holomorphic in O in all variables \({\varvec{x}}_i\) and \({\varvec{y}}_i\). Therefore, Cauchy’s integral formula applied in all variables shows
The balls \(B_{{\varvec{x}},i}\) and \(B_{{\varvec{y}},i}\) have to be chosen such that \(\prod _{i=1}^d B_{{\varvec{x}},i}\times \prod _{i=1}^d B_{{\varvec{y}},i}\subset O\). With Lemma 10, and for \(({\varvec{x}},{\varvec{y}})\in {{\mathbb {R}}}^{2d}\) such that \({\varvec{x}}-{\varvec{y}}\in ({{\mathbb {R}}}^n)_+\) (note that Lemma 10 implies \(({\varvec{x}},{\varvec{y}})\in O\)), this can be achieved by setting \(B_{{\varvec{x}},i}:= B_\varepsilon ({\varvec{x}}_i)\) and \(B_{{\varvec{y}},i}:=B_\varepsilon ({\varvec{y}}_i)\) with \(\varepsilon :=\sin (\pi /(2p))|{\varvec{x}}-{\varvec{y}}|/(2d+1)\). From this, we obtain the estimate
for all \(({\varvec{x}},{\varvec{y}})\in {{\mathbb {R}}}^{2d}\) such that \({\varvec{x}}-{\varvec{y}}\in ({{\mathbb {R}}}^n)_+\), where the first equality follows from \(d({\varvec{x}}-{\varvec{y}})=|{\varvec{x}}-{\varvec{y}}|_p\) for all \({\varvec{x}}-{\varvec{y}}\in ({{\mathbb {R}}}^n)_+\). To remove the restriction \({\varvec{x}}-{\varvec{y}}\in ({{\mathbb {R}}}^n)_+\), consider \(b\in \{0,1\}^d\) and define the function
Since we consider \(\varrho (\cdot ,\cdot )\) from (2), there holds \(\varrho \circ F_b=\varrho \). Since for all \({\varvec{x}},{\varvec{y}}\in {{\mathbb {R}}}^{d}\) with \({\varvec{x}}\ne {\varvec{y}}\), there exists some \(b\in \{0,1\}^d\) such that \(({\varvec{x}}_b,{\varvec{y}}_b):=F_b({\varvec{x}},{\varvec{y}})\) satisfies \({\varvec{x}}_b-{\varvec{y}}_b\in ({{\mathbb {R}}}^n)_+\), we prove (38) for all \({\varvec{x}},{\varvec{y}}\in {{\mathbb {R}}}^d\) with \({\varvec{x}}\ne {\varvec{y}}\). Finally, the fact \(\alpha !\beta ! \le |\alpha +\beta |_1!\), proves that \(\varrho (\cdot ,\cdot )\) from (2) is asymptotically smooth (1).
Next, consider the covariance function \(\varrho (\cdot ,\cdot )\) from (3). By definition \(\varvec{\varSigma }_{\varvec{x}}\) is continuous on \(\overline{D}\). Hence, \(\mathrm{det}(\varvec{\varSigma }_{\varvec{x}})\ge c_0>0\) for all \({\varvec{x}}\in D\). The assumption (4) implies that also \(\mathrm{det}(\varvec{\varSigma }_{\varvec{x}})\) has bounded derivatives in the sense of (4) (since \(\mathrm{det}(\varvec{\varSigma }_{\varvec{x}})\) is a polynomial in the matrix entries of \(\varvec{\varSigma }_{\varvec{x}}\)). Thus, Lemma 12 shows that the functions \(({\varvec{x}},{\varvec{y}})\mapsto \mathrm{det}(\varvec{\varSigma }_{\varvec{x}})^{1/4}\), \(({\varvec{x}},{\varvec{y}})\mapsto \mathrm{det}(\varvec{\varSigma }_{\varvec{y}})^{1/4}\), and \(({\varvec{x}},{\varvec{y}})\mapsto \mathrm{det}(\varvec{\varSigma }_{\varvec{x}}+\varvec{\varSigma }_{\varvec{y}})^{-q}\), \(q\in \{1/2,1\}\) satisfy (1). With \(\varvec{\varSigma }_{\varvec{x}}\), also all functions \(\widetilde{\varvec{\varSigma }}_{\varvec{x}}\) defined by considering only sub-matrices of \(\varvec{\varSigma }_{\varvec{x}}\) satisfy (4). Thus, Cramer’s rule and Lemma 11 show that the map \(({\varvec{x}},{\varvec{y}})\mapsto ((\varvec{\varSigma }_{\varvec{x}}+\varvec{\varSigma }_{\varvec{y}})^{-1})_{i,j}\) for all \(i,j\in \{1,\ldots ,d\}\) satisfies (1). From this, we conclude (again with Lemma 11), that \(({\varvec{x}},{\varvec{y}})\mapsto ({\varvec{x}}-{\varvec{y}})^T(\varvec{\varSigma }_{\varvec{x}}+\varvec{\varSigma }_{\varvec{y}})^{-1}({\varvec{x}}-{\varvec{y}})\) as sum and product of asymptotically smooth functions is asymptotically smooth (1). Finally, Lemma 12 shows that \(\varrho ({\varvec{x}},{\varvec{y}})\) satisfies (1). This concludes the proof.
Appendix—Proof of Proposition 1
The following lemmas state facts about the \(H^2\)-matrix block partitioning, which are well-known but cannot be found explicitly in the literature.
Lemma 13
Under Assumption 1, there exists a constant \(C_{B}>0\) which depends only on d, \(C_\mathrm{u}\), D, and \(B_{X_\mathrm{root}}\) such that all \(X\in {{\mathbb {T}}}_\mathrm{cl}\) satisfy
Moreover, all \((X,Y)\in {{\mathbb {T}}}\) satisfy
where \(C_{BB}>0\) depends only on \(C_B\), \(C_\mathrm{leaf}\), and D.
Proof
The first estimate (39a) follows from the fact that always the longest edge of a bounding box is halved. This means that the ratio \(L_\mathrm{max}/L_\mathrm{min}\) of the maximal and the minimal side length of a bounding box \(B_X\) stays bounded in terms of the corresponding ratio for \(B_{X_\mathrm{root}}\). Therefore, we have
To see the second estimate (39b), consider a given bounding box B with side lengths \(L_1,\ldots , L_d\). Due to Assumption 1 the balls \(Q_{\varvec{x}}\) with centre \({\varvec{x}}\) and radius \(C_\mathrm{u}^{-1}N^{-1/d}/2\) for all \({\varvec{x}}\in {{\mathcal {N}}}\) do not overlap. All balls \(Q_{\varvec{x}}\) with \({\varvec{x}}\in B\) are containted in a box with sidelengths \(L_\mathrm{max}+C_\mathrm{u}^{-1}N^{-1/d}\). Thus, the number \(m_B\) of \({\varvec{x}}\in {{\mathcal {N}}}\) contained in B can be bounded by
Since \(m_B\le 1\) if \(L_\mathrm{max}< C_\mathrm{u}^{-1}N^{-1/d}/2\) and since \(L_\mathrm{max}^d\simeq |B|\), we may improve the estimate to
where \(C_B\) depends only on d and \(C_\mathrm{u}\). On the other hand, Assumption 1 implies that any ball with radius \(C_\mathrm{u}N^{-1/d}\) contains at least one point \({\varvec{x}}\in {{\mathcal {N}}}\). Since each such ball fits inside a box with sidelength \(2C_\mathrm{u}N^{-1/d}\), we obtain
points of \({{\mathcal {N}}}\). This allows us to estimate \(m_B\ge C_{B}^{-1}|B|N-1\) and conclude (39b). The estimate (39c) follows from the fact \(|B_X|=|B_{X^\prime }|/2\) for all \(X\in \mathrm{sons}(X^\prime )\). For (40), we observe with (39b) that
for all \(X\in {{\mathbb {T}}}_\mathrm{cl}\) with hidden constants depending only on \(C_B\). Thus, with (39c), we have for all \(X\in {{\mathbb {T}}}_\mathrm{cl}\) with \(X\in \mathrm{sons}(X^\prime )\) that
Moreover, if additionally \(\mathrm{sons}(X)=\emptyset \), we have even \(2^{-\mathrm{level}(X)}\simeq |B_x|\lesssim C_\mathrm{leaf}/N\). By definition of the block-tree \({{\mathbb {T}}}\), a level difference between X and Y for \((X,Y)\in {{\mathbb {T}}}\) can only happen, if \(\mathrm{sons}(X)=\emptyset \) or \(\mathrm{sons}(Y)=\emptyset \). Assume \(\mathrm{sons}(X)=\emptyset \). In this case, we have \(\mathrm{level}(Y)\ge \mathrm{level}(X)\). Then, we have
with hidden constants depending only on \(C_B\) and D. This implies \(\mathrm{level}(Y)\le \mathrm{level}(X) + C\) for some constant \(C>0\) which depends only on \(C_\mathrm{leaf}\), D, and \(C_B\) from (39). From this we derive (40) by use of (39).
Lemma 14
Given the definition of \({{\mathbb {T}}}_\mathrm{far}\) in Sect. 3.1, there exists a constant \(C>0\) such that all \((X,Y)\in {{\mathbb {T}}}_\mathrm{far}\) satisfy
Proof
By Lemma 13, we have
For \((X,Y)\in \mathrm{sons}(X^\prime ,Y^\prime )\), we obtain additionally
By definition of the block-partitioning, for \((X,Y)\in {{\mathbb {T}}}_\mathrm{far}\) there holds that \(B_X,B_Y\) satisfy (5) and \(B_{X^\prime },B_{Y^\prime }\) do not satisfy (5). Altogether, this implies
where we used \(|\mathrm{level}(X^\prime )-\mathrm{level}(X)|\le 1\). This concludes the proof. \(\square \)
The following lemma gives some basic facts about tensorial Chebychev-interpolation (see, e.g., [3, Section 4.4])
Lemma 15
Let \(f:B\rightarrow {{\mathbb {R}}}\) for an axis parallel box \(B\subseteq {{\mathbb {R}}}^{2d}\) such that \(\partial _j^k f\in L^\infty (B)\) for all \(j=1,\ldots ,d\) and all \(0\le k\le p\). Then, the tensorial Chebychev-interpolation operator of order p, \(I_p:C(B) \rightarrow {{\mathcal {P}}}^p(B)\) satisfies
where
is the operator norm of the one dimensional Chebychev interpolation operator
Proof
It is well-known that the one dimensional Chebychev interpolation operator \(I_p^{{\varvec{x}}}\) satisfies the error estimate for any \(f\in C([-1,1])\)
with an operator norm given in (43). Consider \(B:=[-1,1]^{2d}\). Then, there holds with \(I_p^{{\varvec{x}}_i}\) denoting interpolation in the \({\varvec{x}}_i\)-variable \(i\in \{1,\ldots ,2d\}\)
Since, for any affine transformation \(A:{{\mathbb {R}}}^{2d}\rightarrow {{\mathbb {R}}}^{2d}\), we have \(I_p(f\circ A)= I_p(f)\circ A\), a standard scaling argument concludes the proof.
Proof of Proposition 1
We start by proving that \(\lambda _\mathrm{min}(\varvec{C}_p)>0\) if p satisfies (7). To that end, note
since the Frobenius norm is an upper bound for the spectral norm. By use of (6) (which is proved below) and (7), we conclude \( \lambda _\mathrm{min}(\varvec{C}_p)>0\).
To see (6), we first estimate the maximal depth of the tree \({{\mathbb {T}}}_\mathrm{cl}\). With (39b)–(39c), we obtain \(C_\mathrm{leaf}\le |X| \lesssim 2^{-\mathrm{level}(X)}\) for all \(X\in {{\mathbb {T}}}_\mathrm{cl}\) with \(\mathrm{sons}(X)\ne \emptyset \). Thus, there holds
Second, we bound the so-called sparsity constant
The H-matrix case can be found in [10, Lemma 4.5]. For the \(H^2\)-matrix case, the combination of (40) and (41) (from Lemma 14) shows that \((X,Y)\in {{\mathbb {T}}}_\mathrm{far}\) only if \(B_Y\) touches the (hyper-) annulus with center \(B_X\) and radii \(C^{-1}\mathrm{diam}(B_X)\) and \(C\mathrm{diam}(B_X)\). By comparing the volumes of this annulus and of \(B_Y\) and using the fact that all the bounding boxes are disjoint, we see that the number of Y such that \((X,Y)\in {{\mathbb {T}}}_\mathrm{far}\) is bounded in terms of C and the constants in (39).
For \(Y\in {{\mathbb {T}}}_\mathrm{cl}\) such that \((X,Y)\in {{\mathbb {T}}}_\mathrm{near}\), we have with (39)–(40)
Again, comparing the volumes of the ball with radius \(\mathrm{diam}(B_X)\) and of \(B_Y\), we see that the number of Y such that \((X,Y)\in {{\mathbb {T}}}_\mathrm{near}\) is bounded in terms of the constants in (39). Altogether, we bound \(C_\mathrm{sparse}\) uniformly in terms of the constants of Lemma 13. Now, [3, Lemma 3.38] proves the estimate for storage requirements and [3, Theorem 3.42] proves the estimate for matrix–vector multiplication.
It remains to prove the error estimate (see also [3, Section 4.6] for the integral operator case). To that end, note that since the near field \({{\mathbb {T}}}_\mathrm{near}\) is stored exactly, there holds
Given, \((i,j)\in I(X)\times I(Y)\), we have with the interpolation operator \(I_p\) from Lemma 15 and (1)
With the admissibility condition (5), we get
and hence
The combination of the above estimates concludes the proof. \(\square \)
Rights and permissions
About this article
Cite this article
Feischl, M., Kuo, F.Y. & Sloan, I.H. Fast random field generation with H-matrices. Numer. Math. 140, 639–676 (2018). https://doi.org/10.1007/s00211-018-0974-2
Received:
Revised:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00211-018-0974-2