New Lower Bounds for the Integration of Periodic Functions

We study the integration problem on Hilbert spaces of (multivariate) periodic functions.The standard technique to prove lower bounds for the error of quadrature rules uses bump functions and the pigeon hole principle. Recently, several new lower bounds have been obtained using a different technique which exploits the Hilbert space structure and a variant of the Schur product theorem. The purpose of this paper is to (a) survey the new proof technique, (b) show that it is indeed superior to the bump-function technique, and (c) sharpen and extend the results from the previous papers.


Introduction 1.Overview
We study the integration problem on reproducing kernel Hilbert spaces of multivariate periodic functions.That is, we consider Hilbert spaces H of complex-valued and continuous functions on [0, 1] d which admit a complete orthogonal system of trigonometric monomials.This scale of functions contains the classical Sobolev Hilbert spaces of isotropic, mixed or anisotropic smoothness (see, e.g., [7,32]) as well as the weighted Korobov spaces which are common in tractability studies (see, e.g., [8] and the references therein).
We are interested in the error e n (H), which is the smallest possible difference (in the operator norm) between the integration functional and a quadrature rule using up to n function values.This corresponds to the worst case setting of information-based complexity, see, e.g., [29,30,34].It is standard to prove lower bounds for e n (H) using the bump-function technique: If we construct 2n bump functions in H with disjoint support, then n of the supports do not contain a quadrature point and therefore, for the fooling function given by the sum of the corresponding bumps divided by its norm, the quadrature rule outputs zero and the error of the quadrature rule is bounded below by the integral of the fooling function.This simple technique goes back at least to [2] and leads to optimal lower bounds for many classical spaces from the above family.Interestingly, however, it is not sufficient to prove optimal lower bounds in the extreme cases when the functions in the Hilbert space are either very smooth (analytic) or barely continuous.In these cases, optimal lower bounds were recently obtained using a different technique which exploits the Hilbert space structure and a variant of the Schur product theorem, see [12,13,38].In this paper, we want to (a) survey the new proof technique, (b) show that it is indeed superior to the bump-function technique, and (c) sharpen and extend the results from the previous papers.The structure of this paper is as follows: • The precise setting is introduced in Section 1.2.An overview of results is given in Section 1.3.
• In Section 2, we first show that the bump function technique fails in certain ranges of smoothness.
We then describe the new technique, which we will refer to as the Schur technique.We do this for a more general family of reproducing kernel Hilbert spaces than the above, see Theorem 2.
• In Section 3, we present some general lower bounds on the error e n (H) for the spaces of periodic functions.As a corollary, we also obtain a new result on the largest possible gap between the error of sampling algorithms and general algorithms for L 2 -approximation on reproducing kernel Hilbert spaces with finite trace, see Section 3.3.We improve a lower bound from [13], further narrowing the gap to the best known upper bounds from [5,19,25].
• In Section 4, we derive new lower bounds for function spaces which combine the fractional isotropic smoothness s = d/2 with an additional logarithmic term.Since this ensures that function values are well-defined and that the functions in the space are continuous, while isotropic smoothness s = d/2 without any logarithmic perturbation is not enough, we refer to this setting as spaces of small smoothness.We consider also function spaces of mixed smoothness, where the borderline is given by s = 1/2 independently on the dimension d.Spaces combining fractional and logarithmic smoothness have been studied, e.g., in [6,11,17,24].
• In Section 5, we obtain lower bounds for the case of large smoothness, namely, for analytic periodic functions.We present a result on the curse of dimensionality, which is in sharp contrast to a recent result from [10], and an asymptotic lower bound that complements recent error bounds for classes of analytic functions on R d from [16,23].

Preliminaries
Let H be a reproducing kernel Hilbert space (RKHS) on a non-empty set D. That is, H shall be a Hilbert space of functions f : D → C such that point evaluation is continuous for each x ∈ D. This means that for every x ∈ D, there is a function is called the reproducing kernel of H.We refer to [1] and [4] for basics on RKHSs.We are interested in the computation of a continuous functional on H, that is, an operator of the form with some h ∈ H.The main interest lies within integration functionals of the form with a probability measure µ on D. Clearly, if the functional INT µ is continuous, then it can be written in the form (1).In this case, the corresponding representer h satisfies We want to compute S h (f ) using a quadrature rule with nodes x k ∈ D and weights a k ∈ C. The worst-case error of the quadrature rule is defined by We are interested in the error of the best possible quadrature rule, which we denote by In this paper, we will mainly consider Hilbert spaces of multivariate periodic functions, defined as follows.Here, e k = exp(2πi k, • ) denotes the trigonometric monomial with frequency k ∈ Z d .The Fourier coefficients of an integrable function f : The condition (3) implies that the Fourier series of f ∈ H λ converges point-wise (and even uniformly) for all and that f is continuous as a 1-periodic function on R d .Moreover, and thus, the space H λ is a reproducing kernel Hilbert space.The trigonometric monomials (e k ) k∈Z d are a complete orthogonal system in H λ and the reproducing kernel of H λ is given by We note that the space H λ can be defined analogously for bounded sequences λ ∈ ℓ 1 (Z d ) as the space of all f ∈ L 2 ([0, 1] d ) satisfying (3).In this case, however, function evaluation is not well defined and H λ it is not a reproducing kernel Hilbert space.The most common choices (see, e.g., [32] or [22]) of the sequence λ include In that case, H λ is the Sobolev space of periodic functions with isotropic smoothness.If s is a positive integer, then an equivalent norm on H λ is given by In this way, we obtain the spaces of periodic functions with dominating mixed smoothness.If s ∈ N is a positive integer, then also these spaces allow for an equivalent norm given in terms of derivatives, namely On the space H λ , we consider the integration problem with respect to the Lebesgue measure λ d on [0, 1] d , that is, we take µ = λ d .In this case, the representer of the functional INT µ is given by the constant function h ≡ λ 0 .For this standard integration problem on the unit cube, we leave out the S h in the notation of the error, writing e n (H λ ) instead of e n (H λ , S h ), e(Q n , H) instead of e(Q n , H, S h ) and we simply write INT instead of INT µ .The initial error in this case is given by In analogy to the numbers e n (H λ ), we also define the sampling numbers and the approximation numbers which reflect the error of the best possible (linear) algorithm for L 2 -approximation using at most n function values or n arbitrary linear measurements, respectively.Further notation.For sequences (a n ) and (b n ), we write a n b n if there is a constant c > 0 such that a n ≤ c b n for all but finitely many n.We write a n b n if b n a n and a n ≍ b n if both relations are satisfied.If H is a Hilbert space, H ′ denotes the space of all continuous linear functionals on H.

Results
As mentioned earlier, the main purpose of the paper is to survey a new proof technique for lower bounds on the error e n (H λ ).Here, we gather some of the new results obtained in this paper using the new technique.
The first category of results concerns the asymptotic behavior of the errors e n (H λ ) and g n (H λ ) in the case of small smoothness, i.e., in the case that λ decays barely fast enough to ensure that H λ is a reproducing kernel Hilbert space.For function spaces H λ with fractional isotropic smoothness d/2 and logarithmic smoothness β > 1/2, i.e., if λ is given by (19), we obtain see Corollary 4. Of special interest is the logarithmic gap between the sampling and the approximation numbers, which has been shown in [13] for the univariate case, and is now extended to the multivariate case.The lower bound on the integration error follows from a result for more general sequences λ, see Theorem 4, and cannot be proven by the standard technique of bump functions, see Section 2.1.We also prove the same relation (5) for the (multivariate) spaces of fractional mixed smoothness 1/2 and logarithmic smoothness β > 1/2, where λ is given by (22), see Section 4.2.
Second, we obtain new results on the numbers e * n (σ) and g * n (σ), which reflect the worst possible behavior of the nth minimal integration error and sampling number, respectively, on reproducing kernel Hilbert spaces for any given sequence (σ n ) n∈N of singular values, see Section 3.3 for a precise definition.We obtain that, up to universal constants, both these numbers are equivalent to see Corollary 3. The new thing here is again the lower bound, the upper bound is known from [5].
Thirdly, we also derive a result on the curse of dimensionality.We obtain that the curse is present for numerical integration and L 2 -approximation on classes of the form for any weight function g : N 0 → (0, ∞) and p = 2, see Theorem 7.This is in sharp contrast to recent tractability results for analogous classes with p = 1, see [10] for numerical integration and [18] for L 2 approximation.

Proof techniques for lower bounds
In this section, we discuss techniques to prove lower bounds for the integration error e n (H λ ).We show the limits of the bump-function technique, which is probably the most common technique, and describe a new method used in this paper.We note that there are other proof techniques which we do not discuss here like the technique of decomposable kernels from [28] and a method from [31].We only compare our new approach with the most standard technique.

Limits of the bump-function technique
The most well-known technique to prove lower bounds for the integration problem is the use of bump functions.It is based on the observation (cf.[29,Chapter 4]) that a lower bound for e(Q n , H) for a fixed quadrature rule Q n with nodes x 1 , . . ., x n is equivalent to the construction of a function f , which vanishes at the sampling points, i.e., with f (x 1 ) = . . .= f (x n ) = 0, and which has simultaneously large integral INT(f ) and small norm f H .This leads to Q n (f ) = 0 and the lower bound A function like this is called a fooling function.One way how to construct such a fooling function to a given Q n is to start with a bump function, i.e., with a smooth periodic function ϕ ∈ C([0, 1]) with supp ϕ ⊂ [0, 1/(2n)] and consider its translates ϕ(x − j/(2n)) for j = 0, . . ., 2n − 1 (and similarly for the multivariate case).By the pigeonhole principle, there are at least n of these dilates, which vanish at all the sampling points x 1 , . . ., x n .Adding them, we then obtain the fooling function f .The use of this (rather intuitive) technique can be traced back at least to [2].It is probably the most widely used method to prove lower bounds for numerical integration and other approximation problems, where one is limited to the use of function values, cf.[26,27,34].The main reason for its wide use is surely that it often leads to optimal results.Note however that for example in [37] it was necessary to combine the bump-function technique with the Khintchine inequality to obtain optimal lower bounds in certain limiting cases.
The aim of this section is to show that the technique of bump functions can only provide suboptimal results for the integration problem on H λ in the cases that • the sequence λ decays very slowly.This means that H λ contains functions with low smoothness, which are just about continuous.
• the sequence λ decays very fast.This means that the functions in H λ are very smooth or even analytic.
In the second case, it is rather obvious that the bump function technique does not work.If λ decays fast enough (for example if |λ k | ≤ c 1 exp(−c 2 |k|) for some c 1 , c 2 > 0 and all k ∈ Z, c.f. [3, §25]), all functions in H λ are analytic and since there is no analytic function with compact support (except the zero function), any fooling function obtained by the bump-function technique will not be contained in the space H λ .
We now show the sub-optimality of the bump-function technique in the case of small smoothness.To ease the presentation, we consider only the case d = 1, but we note that similar results can be obtained in the multivariate case.We consider sequences of the form which are just about summable.It was shown in [13,Theorem 4] and it also follows from Corollary 4 below that This lower bound is sharp, see [13, Proposition 1] and [5].Here, we are going to prove the following.
Theorem 1.Let λ ∈ ℓ 1 (Z) be given by (6).There exists an absolute constant C > 0 such that for every even n ≥ 2 and every ϕ ∈ C([0, 1]) with supp ϕ ⊂ [0, 1/(2n)] there exists {z 1 , . . ., z n } ⊂ {j/(2n) : j = 0, . . ., 2n − 1} such that the function Theorem 1 shows that no matter how we choose the bump function ϕ ∈ C([0, 1]) with support in [0, 1/(2n)], it is possible to choose the set of sampling points X = {x 1 , . . ., x n } ⊂ [0, 1] in such a way that leaving from the sum the terms which do not vanish on X does not provide a fooling function sufficient to show (7).In other words, the classical bump function technique cannot yield the sharp lower bound on e n (H λ ) in (5).Before we come to the proof of Theorem 1, we need a characterization of the norm of f in the space H λ in terms of first order differences, which are defined for every 0 < h < 1 simply by Here, we interpret f as a periodic function defined on all R, which makes ∆ h f (x) indeed well-defined for all The following proposition is in principle a special case of [6, Theorem 10.9], where a characterization by first and higher order differences is given for Besov spaces of generalized smoothness on R d .We provide a short and direct proof for reader's convenience.Proposition 1.Let λ ∈ ℓ 1 (Z) be given by (6).Then f ∈ C([0, 1]) belongs to H λ if, and only if, for all k ∈ Z and, by orthonormality of (e k ) k∈Z in L 2 ([0, 1]), To simplify the notation, we denote and obtain For j ∈ Z we put and we will show that γ j ≍ 1/λ j for all j ∈ Z with universal constants of equivalence which do not depend on j.In view of (3), this will finish the proof.
The estimate of γ j for j = 0 can be always achieved by changing the constants of equivalence.Let j = 0. Then The estimate of γ j from above is then obtained by which finishes the proof.
Remark 1.We point out also another reason, why Theorem 1 is rather counter-intuitive.Motivated by an explicit formula for the optimal fooling function for equi-distributed sampling points we consider (for fixed n ∈ N) the bump function Note, that Φ n (t) indeed vanishes in the points t = j/(2n), j = 0, . . ., 2n.
Finally, an easy calculation reveals that with constants of equivalence independent of n.Theorem 1 therefore shows, that removing some of the bumps from the sum in ( 12) can actually increase the norm of such a function.Indeed, if ϕ (n) is the function constructed in Theorem 1 using the ϕ n from ( 12), then the integrals of Φ n and ϕ (n) are comparable and (13) together with (9) shows that ϕ (n) H λ / Φ n H λ grows (at least) as log(n) if n tends to infinity.

The Schur technique
So how can we prove lower bounds for the integration problem in the cases where the bump-function technique does not work?The recent results for small smoothness and for analytic functions have been obtained using a certain modification of the classical Schur product theorem on the entry-wise product of positive semi-definite matrices.We will describe this technique now in the general setting of Section 1.2.That is, we are given a RKHS H with kernel K on a domain D and a functional S h represented by h ∈ H.
The first ingredient of this technique is a characterization of lower bounds on numerical integration via the positive definiteness of certain matrices involving the kernel K and the representer h of the integral from [13].
Proposition 2. Let H be a RKHS on a domain D with the kernel K : D × D → C and let h ∈ H.Then, for every α > 0, if, and only if, the matrix is positive semi-definite for all {x 1 , . . ., x n } ⊂ D.
Proof.Let Q n be given by ( 2).Then we denote a = (a 1 , . . ., a n ) * , h = (h(x 1 ), . . ., h(x n )) * and K = (K(x j , x k )) n j,k=1 and obtain Let us assume now that K − αhh * is positive-semidefinite.If a * Ka = 0, then also a * hh * a = |a * h| 2 = 0 and (15) implies that e(Q n , H, S h ) 2 ≥ h 2 H .If a * Ka is positive, then we continue (15) by We use that a * Ka−αa * hh * a = a * Ka−α|a * h| 2 ≥ 0 and take the infimum over all quadrature formulas Q n and obtain that e n (H, S h On the other hand, assume that ( 14) holds.Then e(Q n , H, S h ) 2 ≥ h 2 H −α −1 for every quadrature formula Q n with arbitrary nodes {x 1 , . . ., x n } ⊂ D and arbitrary weights a 1 , . . ., a n ∈ C. If a * Ka = 0, then it follows from (15) that 2Re (a * h) ≤ α −1 holds for a and all its complex multiples.Hence, a * h = 0 and a * Ka − αa * hh * a = 0.If a * Ka is positive, then we can assume (possibly after rescaling a with a non-zero t ∈ C) that a * h = a * Ka, in which case ( 16) becomes an identity.Hence, a * Ka ≥ α|a * h| 2 and the result follows.
The second ingredient is a lower bound on the entry-wise square of a positive semi-definite matrix related to the Schur product theorem, which was proven in [38].If M = (M i,j ) n i,j=1 ∈ C n×n , then we denote by M = (M i,j ) n i,j=1 the matrix with complex conjugated entries and by M • M the matrix with entries |M i,j | 2 .Furthermore, diag M is the column vector of the diagonal entries of M .
is also positive semi-definite.
Proposition 2 and Proposition 3 can be easily combined to obtain lower bounds for numerical integration.We state it under the assumption that the kernel K : D × D → C can be written as a sum of squares of reproducing kernels with constant diagonal.To be more specific, we assume that K can be written as where M i : D × D → C are positive semi-definite functions on D, which are constant on the diagonal, i.e., M i (x, x) = c i ≥ 0 for all x ∈ D. Then K(x, x) = κ := m i=1 c 2 i for every x ∈ D and also K is constant on the diagonal.Theorem 2. Let H be a RKHS on a domain D with the kernel K : D × D → C, which can be written as a sum of squares of reproducing kernels with constant diagonal, and let 1 ∈ H. Consider the integration problem S = S h with the constant representer h = 1.Then where κ is the value of K on the diagonal.
Proof.Let K be written as in (17) and let M i (x, x) = c i .Further, let x 1 , . . ., x n ∈ D. Then (M i (x j , x k )) n j,k=1 is positive semi-definite and by Proposition 3, so is the matrix .
Therefore, also the sum of these matrices over all i ≤ m, i.e., the matrix is positive semi-definite.By Proposition 2, together with h = 1, this implies the error bound.
In particular, one needs at least 1 2 1 2 H • K(x, x) quadrature points in order to reduce the initial error by a factor of two.This insight is often already enough to prove the curse of dimensionality, see [12] and Section 5. Surprisingly, Theorem 2 can also be used to prove new results on the order of convergence of the integration error.This path is described in the next section.
Open Problem 1.The Conjecture 2 of [15] suggests that, if f : R d → R is non-negative and has a non-negative Fourier transform, then the matrix is positive semi-definite for every n ∈ N and every choice of {x 1 , . . ., x n } ⊂ R d .Let us note that Proposition 3 together with the classical Bochner theorem (cf.[38,Theorem 5]) gives an affirmative answer to this conjecture if f = g 2 , where g : R d → R has a non-negative Fourier transform.But in its full generality, the Conjecture 2 of [15] seems to be still open.

Some general lower bounds for periodic functions
We now transfer Theorem 2 to the setting of periodic function spaces H λ on D = [0, 1] d .We start with a result for sequences λ ∈ ℓ 1 (Z d ) which are given as a sum of convolution squares in Section 3.1.
We extend this result to the more general class of sequences which can be written as a non-increasing function of a norm in Section 3.2.As this covers all non-increasing sequences in the univariate case, we obtain as a corollary a new and sharp result on the largest possible error e n (H) for any fixed sequence of approximation numbers in Section 3.3.

Sums of squares
Recall, that the reproducing kernel K λ of H λ for a non-negative and summable sequence λ = (λ k ) k∈Z d is given by (4).The square of its absolute value is then given by Therefore, we define the convolution of two non-negative sequences λ, θ ∈ ℓ 1 (Z d ) by A straightforward calculation shows that λ * θ ∈ ℓ 1 (Z d ) and that λ * θ 1 = λ 1 • θ 1 .This notation allows us to reformulate (18) as We say that λ is a sum of convolution squares if there are . Theorem 2 then takes the following form.
Proof.Since both sides of the inequality are homogeneous in λ, we may assume that λ 0 = 1.In this case, the representer of the integral on H λ is given by h = 1, where h 2 H λ = 1.By (18), we obtain Therefore, we may apply Theorem 2 and simply have to note that This is a generalization of [13, Theorem 1] which covers the case d = 1 and m = 1.

Norm-decreasing sequences
Our next step is to save Corollary 1 for more general sequences λ, which are not given as a sum of convolution squares.Namely, we consider sequences of the form λ k = g( k ), where • is a norm on R d and g : [0, ∞) → [0, ∞) is monotonically decreasing.We call such sequences • -decreasing.
Clearly, λ is • -decreasing if and only if it satisfies λ k ≤ λ ℓ for all k, ℓ ∈ Z d with k ≥ ℓ .
Proof.Again, since both sides of the stated inequality are homogeneous with respect to λ, we may assume that k∈Z d λ 2k = 1.We set µ ℓ = 2 −1/2 λ 2ℓ .By the triangle inequality, one of the two relations 2 k ≥ ℓ or 2 k + ℓ ≥ ℓ must hold for each pair k, ℓ ∈ Z d , and therefore λ 2k ≤ λ ℓ or else λ 2k+2ℓ ≤ λ ℓ .Thus we have for all ℓ ∈ Z d that Moreover, (µ * µ) 0 ≤ λ 0 /2.We put ν = µ * µ + tδ 0 and choose t ≥ λ 0 /2 such that ν 0 = λ 0 .Then ν is a sum of convolution squares.It follows from Corollary 1 and ν ≤ λ that Theorem 3 is well-suited to prove the curse of dimensionality, see Section 5.However, we can also tune it to be used for results on the asymptotic behavior of the nth minimal error.Theorem 4. Let λ ∈ ℓ 1 (Z d ) be • -decreasing.For n ∈ N, we let r n be the norm of the (4n − 1)th element in a • -increasing rearrangement of (2Z) d , i.e., Proof.Choose m ∈ Z d with m = r n .We define τ ∈ ℓ 1 (Z d ) by setting τ k = λ k for k > r n and τ k = λ m for 0 < k ≤ r n as well as Then τ is • -decreasing and bounded above by λ.Moreover, and thus Theorem 3 gives e n (H λ ) 2 ≥ e n (H τ ) 2 ≥ τ 0 /2, which leads to the stated lower bound.
In the univariate case, the previous result looks as follows.
Corollary 2. Let λ ∈ ℓ 1 (Z) be non-negative, symmetric and monotonically decreasing on N 0 .Then Proof.We apply Theorem 4 for d = 1 and • = | • |.Therefore, r n = 4n − 2 and because of monotonicity and symmetry, we get This improves upon our previous result [13,Theorem 4], where we obtained a similar lower bound but only under an additional regularity assumption on the sequence λ.

Detour: The power of function values for L 2 approximation
Recently, there has been an increased interest in the comparison of standard information given by function values and general linear information for the problem of L 2 approximation.We refer to [5,19,20,25,33] for recent upper bounds and to [12,13] for lower bounds.Let us denote by Ω the set of all pairs (H, µ) consisting of a separable RKHS H on an arbitrary set D and a measure µ on D such that H is embedded into L 2 (D, µ).For (H, µ) ∈ Ω, we define the sampling numbers and the approximation numbers .
One is interested in the largest possible gap between the two concepts, that is, given a sequence σ 0 ≥ σ 1 ≥ . . .and an integer n ≥ 0, one considers It is known from [14] that g * n (σ) = σ 0 , whenever σ ∈ ℓ 2 .On the other hand, it was proven in [5] that there is a universal constant c ∈ N such that, whenever σ ∈ ℓ 2 , We obtain a matching lower bound as a consequence of Corollary 2. For the spaces H λ , the sequence of the squared approximation numbers equals the non-increasing rearrangement of the sequence λ.Here, we use that approximation on H λ is harder than integration, namely, e n (H λ ) ≤ g n (H λ ).Indeed, if S n = n i=1 g i δ x i is a sampling operator, then we consider the quadrature formula We apply Corollary 2 for the Hilbert spaces H λ with λ k = σ 2 |2k| and obtain This improves upon the currently best known lower bound from [13,Theorem 2] in the sense that our lower bound holds for all and not just infinitely many n ∈ N 0 .Moreover, due to e n (H) ≤ g n (H) and the fact that our lower bounds are proven for the integration problem, an analogous result holds with g * n (σ) replaced by where Ω 0 is the set of pairs (H, µ) ∈ Ω such that µ is a probability measure.Thus, we have the following corollary.
Corollary 3.There are universal constants 0 < c < 1 < C such that, for any sequence σ 0 ≥ σ 1 ≥ . . .and any integer n ≥ 0, we have In this sense, the worst possible behavior of the sampling numbers (or the minimal integration error) for a given sequence of approximation numbers σ is always described by the sequence σ * .If σ is regularly decreasing in the sense that σ n ≍ σ 2n , we obtain that Let us consider the case of polynomial decay, that is, σ n ≍ n −α log −β n.This sequence is squaresummable if and only if α > 1/2 or α = 1/2 and β > 1/2.In the case α > 1/2 it follows that g * n (σ) ≍ σ * n ≍ σ n .For the reproducing kernel Hilbert spaces H λ of multivariate periodic functions the sequence of approximation numbers is always square-summable.Thus there can only be a gap between the concepts of sampling and approximation numbers in the case α = 1/2.This corresponds to function spaces of small smoothness which are discussed in the next section.

Lower bounds for small smoothness
In this section, we consider various function spaces of small smoothness.Spaces of that type appeared already in [24] to characterize path regularity of the Wiener process.We refer also to [17] to recent results on this subject and to [6] for an extensive treatment of function spaces with logarithmic smoothness.These spaces are of particular interest to us, since • this is the only case where an asymptotic gap between the approximation numbers and the sampling numbers is possible, see Section 3.3.
• the standard technique of bump functions does not yield optimal lower bounds, see Section 2.1.
In [13], we obtained lower bounds for the univariate Sobolev spaces which merge fractional smoothness 1/2 and logarithmic smoothness.We want to extend these results to the multivariate case.In the multivariate regime, there are different smoothness scales that generalize the univariate smoothness scale.We consider spaces of isotropic smoothness and spaces of mixed smoothness.

Isotropic smoothness
In the case of isotropic smoothness, we consider sequences of the form where | • | denotes the Euclidean norm on Z d .This sequence is | • |-decreasing and we may therefore apply Theorem 4 to obtain the following result.Recall that the approximation numbers and sampling numbers are defined in Section 3.3.
Corollary 4. Let λ be given by (19).Then Proof.Recall that a n (H λ ) 2 is the (n + 1)st largest entry of λ.We have for ε → 0 + and this easily implies the asymptotic behavior of a n (H λ ).Recalling that e n (H λ ) ≤ g n (H λ ), the upper bounds on e n (H λ ) and g n (H λ ) follow from [5, Theorem 1] and The lower bounds follow from Theorem 4 and (20) as the condition |2k| > r n excludes only O(n) approximation numbers from the sum in the lower bound of Theorem 4.
We remark that, without much additional work, the norm characterization from Proposition 1 can be generalized to H λ with λ given by (19).In this case we need differences of higher order, which are defined for h, x ∈ R d inductively by Using this notation, the multivariate counterpart of Proposition 1 then reads as follows.
Proposition 4. Let λ ∈ ℓ 1 (Z d ) be given by (19) and let M > d/2 be an integer.Then f ∈ C([0, 1] d ) belongs to H λ if, and only if, The proof resembles very much the proof of Proposition 1 and we leave out the rather technical details, cf. also [6,Theorem 10.9], where one can find a more general characterization for function spaces defined on whole R d .Note that in contrast to the univariate case (10), where h was from the unit interval (0, 1), we now consider in (21) all directions h from the unit ball of R d .Similar to the univariate case, using Proposition 4 instead of Proposition 1, one can show that the bump function technique would not suffice to prove the lower bound on e n (H λ ) in Corollary 4.
Open Problem 2. A logarithmic gap between upper and lower bounds of the worst-case error for numerical integration was recently observed also in [11] for Sobolev spaces of functions on the unit sphere S d ⊂ R d+1 .As conjectured already in [11], we also believe that the existing lower bound can be improved.Unfortunately, our results can not be directly applied in this setting, because the norm of a function in these function spaces is defined in terms of its decomposition into the orthonormal basis of spherical harmonics instead of the trigonometric system.Still, it might be possible to transfer our results to the sphere by • showing that our lower bounds from Corollary 4 already hold for the subspace H • λ of functions with compact support in (0, 1) d and • establishing an equivalent characterization of the spaces of generalized smoothness on the sphere using a decomposition of unity and lifting of the spaces H • λ in analogy to [35,Section 27], or alternatively, by working with Theorem 2 directly and a closer examination of (sums of) squares of kernels of Sobolev spaces on the sphere.For the first approach, Proposition 4 might help.For the second approach, the paper [9] might be useful.

Mixed smoothness
We consider spaces of small mixed smoothness.The space is defined as the d-fold tensor product of the univariate space of mixed smoothness from Section 2.1.This results in the space H λ with Here, λ is not norm-decreasing and therefore we cannot use Theorem 4.However, it will turn out that already the lower bound from the univariate space is sharp in this case.The approximation numbers for d > 1 have the same asymptotic behavior as in the case d = 1.
Theorem 5. Let λ be given by (22).Then we have a n,d < r(n) −1/2 and the statement is proven.
Open Problem 3. It would be interesting to determine the so-called asymptotic constants in the case of mixed smoothness 1/2 with logarithmic perturbation.Opposed to Theorem 5, the asymptotic constants might reveal a dependence of the asymptotic decay upon the dimension d.The asymptotic constants for mixed smoothness s > 1/2 have been determined in [22].In contrast to larger smoothness, the asymptotic constants for mixed smoothness 1/2 with logarithmic perturbation cannot decay as a function of d.Now, we immediately obtain the behavior of the sampling numbers and the integration error on the spaces of mixed smoothness 1/2 with logarithmic perturbation.Theorem 6.Let λ be given by (22).Then Proof.The upper bound follows from Theorem 5 and [5, Corollary 2].The proof of the lower bound is essentially transferred from the one-dimensional case.For that sake, we denote by λ = (λ k ) k∈Z the sequence introduced in (6).Note that λ k = λ ke 1 , where e 1 = (1, 0, . . ., 0) is the first canonical unit vector.To prove now the lower bound for INT, let Q n be a quadrature formula on H λ with nodes n we denote the quadrature rule with nodes x n ∈ [0, 1] and the same weights as in Q n .By Corollary 4 or [13,Theorem 4] there is a function f (1) : [0, 1] → C in the unit ball of H λ such that Q (1)  n (f (1) ) − 1 0 f (1) (x) dx n −1/2 log −β+1/2 n.
The function f : [0, 1] d → C, f (x) = f (1) (x (1) ) is contained in the unit ball of H λ and satisfies n (f (1) ) and [0,1] d f (x) dx = 1 0 f (1) (x) dx.Thus, Open Problem 4. As [5,Corollary 2] is only an existence result, the upper bound on the integration error in Theorem 6 is not constructive and does not tell us how to choose the quadrature points optimally.The results from [19,36] show that O(n log n) i.i.d.uniformly distributed quadrature points are suitable with high probability to achieve an error of order e n (H λ ).In the case of isotropic smoothness s > d/2, it is known that O(n) such points suffice, see [21], and the question arises whether the same holds true for other spaces from the family H λ , λ ∈ ℓ 1 (Z d ).

Lower bounds for large smoothness
We discussed in Section 4 that the bump-function technique does not provide optimal lower bounds for numerical integration of functions from H λ as introduced in Definition 1 if the sequence λ = (λ k ) k∈Z d is large, i.e., if it is barely square-summable.Quite naturally, the technique fails also in the other extremal regime -namely when the sequence λ decays very rapidly.Then the space H λ consists only of analytic functions and, therefore, contains no bump functions with compact support at all.Function spaces of analytic functions have a long history.Nevertheless, it is an active research question whether or not their use can help to avoid the curse of dimension for numerical integration and approximation.Based on the technique of [31], the more recent papers [23] and [16] present lower bounds for numerical integration of such classes of functions which avoid any use of bump functions.Here, we prove the curse of dimension for a class of analytic functions that is in connection with a recent result of [10].In the paper [10], it was shown that the integration problem on is polynomially tractable already for very slowly increasing functions g : N 0 → (0, ∞), namely, for g(k) = max(1, log(k)).We contrast this very nice result by showing that the integration problem suffers from the curse of dimension for essentially any non-trivial function g if we replace | f (k)| by | f (k)| 2 in the definition of F 1 d .
Proof.We identify the class F 2 d with H λ , where λ k = 1/g( k ∞ ).Then we simply apply Theorem 3 and obtain .
We also add an asymptotic lower bound for analytic functions.We only write down the univariate case for simplicity.Recall that all the functions in H λ are analytic if the sequence λ decays geometrically and that a lower bound on e n (H λ ) in this case cannot possibly be proven with the bump function technique.We therefore write down the lower bound obtained with the Schur technique.We note, however, that a similar lower bound might be proven in this case with the technique from [31].