Skip to main content
Log in

Mercer’s Theorem on General Domains: On the Interaction between Measures, Kernels, and RKHSs

  • Published:
Constructive Approximation Aims and scope

Abstract

Given a compact metric space X and a strictly positive Borel measure ν on X, Mercer’s classical theorem states that the spectral decomposition of a positive self-adjoint integral operator T k :L 2(ν)→L 2(ν) of a continuous k yields a series representation of k in terms of the eigenvalues and -functions of T k . An immediate consequence of this representation is that k is a (reproducing) kernel and that its reproducing kernel Hilbert space can also be described by these eigenvalues and -functions. It is well known that Mercer’s theorem has found important applications in various branches of mathematics, including probability theory and statistics. In particular, for some applications in the latter areas, however, it would be highly convenient to have a form of Mercer’s theorem for more general spaces X and kernels k. Unfortunately, all extensions of Mercer’s theorem in this direction either stick too closely to the original topological structure of X and k, or replace the absolute and uniform convergence by weaker notions of convergence that are not strong enough for many statistical applications. In this work, we fill this gap by establishing several Mercer type series representations for k that, on the one hand, make only very mild assumptions on X and k, and, on the other hand, provide convergence results that are strong enough for interesting applications in, e.g., statistical learning theory. To illustrate the latter, we first use these series representations to describe ranges of fractional powers of T k in terms of interpolation spaces and investigate under which conditions these interpolation spaces are contained in L (ν). For these two results, we then discuss applications related to the analysis of so-called least squares support vector machines, which are a state-of-the-art learning algorithm. Besides these results, we further use the obtained Mercer representations to show that every self-adjoint nuclear operator L 2(ν)→L 2(ν) is an integral operator whose representing function k is the difference of two (reproducing) kernels.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

Notes

  1. We usually omit a symbol for the corresponding σ-algebra, since, in general, we do not use it.

  2. For the sake of simplicity, we restrict our considerations to σ-finite measures, since otherwise we would have to deal with local ν-zero sets and, later, when dealing with liftings, with technically involved assumptions on ν. Since for the applications we are most interested in we typically have a probability measure, the σ-finiteness is no restriction.

References

  1. Agler, J., McCarthy, J.E.: Pick Interpolation and Hilbert Function Spaces. Am. Math. Soc., Providence (2002)

    MATH  Google Scholar 

  2. Ali, S.T., Antoine, J.-P., Gazeau, J.-P.: Coherent States, Wavelets and Their Generalizations. Springer, New York (2000)

    Book  MATH  Google Scholar 

  3. Alpay, D. (ed.): Reproducing Kernel Spaces and Applications. Birkhäuser Verlag, Basel (2003)

    MATH  Google Scholar 

  4. Aronszajn, N.: Theory of reproducing kernels. Trans. Am. Math. Soc. 68, 337–404 (1950)

    Article  MathSciNet  MATH  Google Scholar 

  5. Bauer, H.: Measure and Integration Theory. De Gruyter, Berlin (2001)

    Book  MATH  Google Scholar 

  6. Bennett, C., Sharpley, R.: Interpolation of Operators. Academic Press, Boston (1988)

    MATH  Google Scholar 

  7. Berlinet, A., Thomas-Agnan, C.: Reproducing Kernel Hilbert Spaces in Probability and Statistics. Kluwer, Boston (2004)

    Book  MATH  Google Scholar 

  8. Caponnetto, A., De Vito, E.: Fast rates for regularized least-squares algorithm. Technical Report CBCL Paper #248, AI Memo #2005-013, MIT, Cambridge, MA (2005)

  9. Caponnetto, A., De Vito, E.: Optimal rates for regularized least squares algorithm. Found. Comput. Math. 7, 331–368 (2007)

    Article  MathSciNet  MATH  Google Scholar 

  10. Cesa-Bianchi, N., Lugosi, G.: Prediction, Learning, and Games. Cambridge University Press, Cambridge (2006)

    Book  MATH  Google Scholar 

  11. Conway, J.B.: A Course in Functional Analysis, 2nd edn. Springer, New York (1990)

    MATH  Google Scholar 

  12. Cristianini, N., Shawe-Taylor, J.: An Introduction to Support Vector Machines. Cambridge University Press, Cambridge (2000)

    Google Scholar 

  13. Cucker, F., Smale, S.: On the mathematical foundations of learning. Bull. Am. Math. Soc. 39, 1–49 (2002)

    Article  MathSciNet  MATH  Google Scholar 

  14. Cucker, F., Zhou, D.X.: Learning Theory: An Approximation Theory Viewpoint. Cambridge University Press, Cambridge (2007)

    Book  MATH  Google Scholar 

  15. De Vito, E., Caponnetto, A., Rosasco, L.: Model selection for regularized least-squares algorithm in learning theory. Found. Comput. Math. 5, 59–85 (2005)

    Article  MathSciNet  Google Scholar 

  16. Hein, M., Bousquet, O.: Kernels, associated structures and generalizations. Technical Report 127, Max-Planck-Institute for Biological Cybernetics (2004)

  17. Hille, E.: Introduction to general theory of reproducing kernels. Rocky Mt. J. Math. 2, 321–368 (1972)

    Article  MathSciNet  MATH  Google Scholar 

  18. Kato, T.: Perturbation Theory for Linear Operators, 2nd edn. Springer, Berlin–New York (1976)

    Book  MATH  Google Scholar 

  19. König, H.: Eigenvalue Distribution of Compact Operators. Birkhäuser, Basel (1986)

    MATH  Google Scholar 

  20. Mendelson, S., Neeman, J.: Regularization in kernel learning. Ann. Stat. 38, 526–565 (2010)

    Article  MathSciNet  MATH  Google Scholar 

  21. Meschkowski, H.: Hilbertsche Räume mit Kernfunktion. Springer, Berlin (1962)

    MATH  Google Scholar 

  22. Novak, E., Woźniakowski, H.: Tractability of Multivariate Problems. Linear Information, vol. 1. European Mathematical Society (EMS), Zürich (2008)

    Book  MATH  Google Scholar 

  23. Pietsch, A.: Eigenvalues and s-Numbers. Geest & Portig K.-G., Leipzig (1987)

    MATH  Google Scholar 

  24. Rao, M.M.: Measure Theory and Integration, 2nd edn. Dekker, New York (2004)

    MATH  Google Scholar 

  25. Riesz, F., Nagy, B.Sz.: Functional Analysis, 2nd edn. Dover, New York (1990)

    MATH  Google Scholar 

  26. Ritter, K.: Average-Case Analysis of Numerical Problems. Lecture Notes in Math., vol. 1733. Springer, Berlin (2000)

    Book  MATH  Google Scholar 

  27. Rudin, W.: Functional Analysis, 2nd edn. McGraw-Hill, New York (1991)

    MATH  Google Scholar 

  28. Saitoh, S.: Theory of Reproducing Kernels and Applications. Longman Scientific & Technical, Harlow (1988)

    MATH  Google Scholar 

  29. Saitoh, S.: Integral Transforms, Reproducing Kernels and Their Applications. Longman Scientific & Technical, Harlow (1997)

    MATH  Google Scholar 

  30. Saitoh, S., Alpay, D., Ball, J.A., Ohsawa, T. (eds.): Reproducing Kernels and Their Applications, Kluwer Academic, Dordrecht (1999)

    MATH  Google Scholar 

  31. Schölkopf, B., Smola, A.J.: Learning with Kernels. MIT Press, Cambridge (2002)

    Google Scholar 

  32. Schölkopf, B., Smola, A.J., Müller, K.-R.: Nonlinear component analysis as a kernel eigenvalue problem. Neural Comput. 10, 1299–1319 (1998)

    Article  Google Scholar 

  33. Shawe-Taylor, J., Williams, C.K.I., Cristianini, N., Kandola, J.: On the eigenspectrum of the Gram matrix and the generalization error of kernel-PCA. IEEE Trans. Inf. Theory 51, 2510–2522 (2005)

    Article  MathSciNet  Google Scholar 

  34. Smale, S., Zhou, D.-X.: Estimating the approximation error in learning theory. Anal. Appl. 1, 17–41 (2003)

    Article  MathSciNet  MATH  Google Scholar 

  35. Smale, S., Zhou, D.-X.: Learning theory estimates via integral operators and their approximations. Constr. Approx. 26, 153–172 (2007)

    Article  MathSciNet  MATH  Google Scholar 

  36. Steinwart, I., Christmann, A.: Support Vector Machines. Springer, New York (2008)

    MATH  Google Scholar 

  37. Steinwart, I., Hush, D., Scovel, C.: Optimal rates for regularized least squares regression. In: Dasgupta, S., Klivans, A. (eds.) Proceedings of the 22nd Annual Conference on Learning Theory, pp. 79–93 (2009)

    Google Scholar 

  38. Steinwart, I., Hush, D., Scovel, C.: Training SVMs without offset. J. Mach. Learn. Res. 12, 141–202 (2011)

    MathSciNet  Google Scholar 

  39. Strauss, W., Macheras, N.D., Musiał, K.: Liftings. In: Pap, E. (ed.) Handbook of Measure Theory, vol. II, pp. 1131–1184. Elsevier, Amsterdam (2002)

    Chapter  Google Scholar 

  40. Sun, H.: Mercer theorem for RKHS on noncompact sets. J. Complex. 21, 337–349 (2005)

    Article  MATH  Google Scholar 

  41. Tartar, L.: An Introduction to Sobolev Spaces and Interpolation Spaces. Springer, Berlin (2007)

    MATH  Google Scholar 

  42. Ionescu Tulcea, A., Ionescu Tulcea, C.: Topics in the Theory of Lifting. Springer, New York (1969)

    MATH  Google Scholar 

  43. Wahba, G.: Spline Models for Observational Data. Series in Applied Mathematics, vol. 59. SIAM, Philadelphia (1990)

    Book  MATH  Google Scholar 

  44. Wendland, H.: Scattered Data Approximation. Cambridge University Press, Cambridge (2005)

    MATH  Google Scholar 

  45. Werner, D.: Funktionalanalysis. Springer, Berlin (1995)

    MATH  Google Scholar 

  46. Yao, Y., Rosasco, L., Caponnetto, A.: On early stopping in gradient descent learning. Constr. Approx. 26, 289–315 (2007)

    Article  MathSciNet  MATH  Google Scholar 

  47. Zhou, D.-X.: The covering number in learning theory. J. Complex. 18, 739–767 (2002)

    Article  MATH  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Ingo Steinwart.

Additional information

Communicated by G. Kerkyacharian.

Appendices

Appendix A: Related Operators and the Spectral Theorem

This appendix recalls some facts from the spectral theory of compact, self-adjoint operators acting between Hilbert spaces. We begin with the classical spectral theorem, see, e.g., [18, Theorem V.2.10 on p. 260] or [45, Theorem VI.3.2].

Theorem A.1

(Spectral Theorem)

Let H be a Hilbert space and A:HH be a compact, positive, and self-adjoint operator. Then there exist an at most countable ONS (e i ) iI of H and a family (μ i ) iI converging to 0 such that μ 1μ 2≥…>0 and

(53)
(54)

Moreover, (μ i ) iI is the family of nonzero eigenvalues of A (including geometric multiplicities), and, for all iI, e i is an eigenvector for μ i . Finally, both (53) and (54) actually hold for all ONSs \((\tilde{e}_{i})_{i\in I}\) of H for which, for all iI, the vector \(\tilde{e}_{i}\) is an eigenvector of μ i .

The next well known theorem, see, e.g., [27, Theorem 12.10], relates the image of an operator B to the null-space of its adjoint B .

Theorem A.2

Let H 1 and H 2 be Hilbert spaces and B:H 1H 2 be a bounded linear operator. Then we have

$$\overline{ \operatorname {ran}B} = \bigl(\ker B^*\bigr)^\perp.$$

In particular, B is injective if and only if B has a dense image.

The following theorem lists other well-known facts about the eigenvalues and the spectral representations of certain operators. These facts are widely known, but since we are unaware of a reference for the particular formulation we need, we decided to include its relatively straightforward proof for the sake of completeness. The name for this theorem was borrowed from [23, Sect. 3.3.4].

Theorem A.3

(Principle of Related Operators)

Let H 1 and H 2 be Hilbert spaces and B:H 1H 2 be a bounded linear operator. We define the self-adjoint and positive operators A 1:H 1H 1 and A 2:H 2H 2 by A 1:=B B and A 2:=BB , respectively. Then the following statements are true:

  1. (i)

    Given a μ>0, we denote the eigenspaces of A 1 and A 2 that correspond to μ by

    Then the map

    is well defined, bijective, and, in addition, we have \(B^{*}_{\mu}B_{\mu}=\mu \operatorname {id}_{E_{1}(\mu)}\).

  2. (ii)

    We have kerA 1=kerB and kerA 2=kerB .

  3. (iii)

    Assume that A 1 is compact, and let (e i ) iI be an at most countable ONS of H 1 and (μ i ) iI be a family converging to 0 such that μ 1μ 2≥…>0 and

    $$A_1x = \sum_{i \in I} \mu_i\langle x,e_i\rangle_{H_1} e_i,\quad x\in H_1,$$

    where we note that there exist such (e i ) iI and (μ i ) iI by Theorem A.1. For iI, we define \(f_{i} := \mu_{i}^{-1/2} B e_{i}\). Then (f i ) iI is an ONS of H 2, and A 2 has the spectral representation

    $$ A_2 y = \sum_{i \in I}\mu_i \langle y,f_i\rangle_{H_2}f_i, \quad y\in H_2.$$
    (55)

    Finally, we have \(e_{i} = \mu_{i}^{-1/2} B^{*} f_{i}\) for all iI.

  4. (iv)

    A 1 is compact if and only if A 2 is compact.

  5. (v)

    If A 1 is compact, we have \(\overline{\operatorname {ran}B} = \overline {\operatorname {span}\{f_{i}:i\in I \}} = \overline{\operatorname {ran}A_{2}}\) and \(\overline{\operatorname {ran}B^{*}} = \overline{\operatorname {span}\{e_{i}:i\in I \}} =\overline{\operatorname {ran}A_{1}}\).

Proof

(i) It is easy to check that A 1 and A 2 are indeed self-adjoint and positive. Let us first show that B μ is well defined, that is, that BxE 2(μ) for all xE 1(μ). To this end, we pick an xE 1(μ); i.e., we have A 1 x=μx. From this we conclude A 2 Bx=BB Bx=BA 1 x=μBx, and thus we have BxE 2(μ). Similarly, for xE 1(μ) with Bx=0, we obtain μx=Ax=B Bx=0, and since μ>0, we conclude x=0; i.e., B μ is injective. By interchanging the role of B and B , we analogously see that

is well defined and injective. To show that B μ is surjective, we now pick an yE 2(μ) and define x:=μ −1 B y. Our previous consideration then gives xE 1(μ), and since we further have Bx=μ −1 BB y=μ −1 A 2 y=y, we obtain the surjectivity of B μ . The last assertion is a consequence of B B=A 1.

(ii) By symmetry, it suffices to show kerA 1=kerB. Moreover, the inclusion kerA 1⊃kerB immediately follows from A 1=B B. To show the converse inclusion, we fix an x∈kerA 1. Then the definition of A 1 yields \(0 = \langle A_{1}x,x\rangle_{H_{1}} =\langle Bx,Bx\rangle_{H_{2}}\), which implies x∈kerB.

(iii) Let us show that (f i ) iI is an ONS of H 2. To this end, we fix i,jI and find

$$\langle f_i, f_j\rangle_{H_2} = \frac{1}{\sqrt{ \mu_i \mu_j}} \langle Be_i,Be_j\rangle_{H_2} = \frac{1}{\sqrt{ \mu_i\mu_j}}\langle e_i, A_1 e_j\rangle_{H_1} = \sqrt{\frac{\mu_j}{\mu_i}} \langle e_i,e_j\rangle_{H_1},$$

and since (e i ) iI is an ONS, we then conclude that (f i ) iI is an ONS. Moreover, by (i) we have

$$ A_2f_i = BB^* f_i= \mu_i^{-1/2}BB^*B e_i = \mu^{1/2}_iBe_i = \mu_i f_i, \quad i\in I.$$
(56)

By Theorem A.1 and (ii), the compactness of A 1 further implies

Using Theorem A.2 and (ii), we then find

$$H_2 = \ker B^* \oplus\overline{ \operatorname {ran}B} = \ker A_2\oplus\overline{{\operatorname {span}\{f_i:i\in I\}}},$$

and combining this with (56), we obtain (55) by Theorem A.1. The last assertion follows from \(\mu_{i}^{-1/2} B^{*} f_{i} = \mu_{i}^{-1} B^{*}B e_{i} = e_{i}\), where in the last step we used (i).

(iv) If A 1 is compact, we obtain the spectral representation (55) for A 2, and hence A 2 is compact. The inverse implication follows by symmetry.

(v) The proof of (iii) has already shown \(\overline {\operatorname {ran}B} = \overline{\operatorname {span}\{f_{i}:i\in I \}}= \overline{\operatorname {ran}A_{2}}\), and the second equality follows by symmetry and (iv). □

Appendix B: Liftings

In this appendix, we briefly recall some results related to liftings on the space of bounded measurable functions. To this end, we assume that \((X,\mathcal {A})\) is a measurable space. We denote by

$$\mathcal {L}_{\infty}(X) := \{f:X\to \mathbb {R}\, | \, f \mbox{ bounded and measurable} \}$$

the space of all bounded measurable functions on X and equip this space with the usual supremum norm ∥⋅∥. Furthermore, if ν is a σ-finite measureFootnote 2 on \((X,\mathcal {A})\), we define, for a measurable f:X→ℝ,

$$\Vert f \Vert _{\mathcal {L}_{\infty}(\nu)} := \inf\bigl\{ a\geq0 : \bigl\{x\in X:\bigl|f(x)\bigr| > a\bigr \} \mbox{ is a $\nu$-zero set} \bigr\}.$$

This leads to the space

$$\mathcal {L}_{\infty}(\nu):= \bigl\{ f:X\to \mathbb {R}\, | \, f \mbox{ measurable and }\Vert f \Vert _{\mathcal {L}_{\infty}(\nu)} < \infty\bigr\}$$

of essentially bounded, measurable functions on X. By considering a sequence \(\alpha_{n}\searrow \Vert f \Vert _{\mathcal {L}_{\infty}(\nu)}\), it is straightforward to show that the infimum in the definition of \(\Vert \cdot \Vert _{\mathcal {L}_{\infty}(\nu)} \) is actually attained; that is,

$$ \nu\bigl( \bigl\{|f|> \Vert f \Vert _{\mathcal {L}_{\infty}(\nu)} \bigr\} \bigr )=0,\quad f\in \mathcal {L}_{\infty}(\nu).$$
(57)

Furthermore, we write \({L}_{\infty}(\nu):= \mathcal {L}_{\infty}(\nu)_{/\sim}\) for the quotient space with respect to the usual equivalence relation fg:⇔ν({fg})=0 on \(\mathcal {L}_{\infty}(\nu)\). From (57), we can then conclude that the linear map

is a metric surjection, and hence its quotient map \(\bar{I}_{\nu}:\mathcal {L}_{\infty}(X)_{/\sim} \to {L}_{\infty}(\nu)\), which is given by \(\bar{I}([f]_{\sim}) = [f]_{\sim}\), is an isometric isomorphism. Here we note that, given an \(f\in \mathcal {L}_{\infty}(X)\), the equivalence class [f] in L (ν) consists, in general, of more functions than the corresponding equivalence class [f] in \(\mathcal {L}_{\infty}(X)_{/\sim}\). Nevertheless, \(\bar{I}_{\nu}\) gives us a canonical tool to identify the spaces \(\mathcal {L}_{\infty}(X)_{/\sim}\) and L (ν). Our next goal is to find a continuous and linear right-inverse of I ν . To this end, we recall the definition of a lifting on \(\mathcal {L}_{\infty}(X)\).

Definition B.1

Let \((X,\mathcal {A})\) be a measurable space and ν be a σ-finite measure on \((X,\mathcal {A})\). We say that a map \(\rho:\mathcal {L}_{\infty}(X)\to \mathcal {L}_{\infty}(X)\) is a ν-lifting if the following conditions are satisfied:

  1. (i)

    ρ is an algebra homeomorphism; that is, ρ is linear and ρ(fg)=ρ(f)ρ(g) for all \(f,g\in \mathcal {L}_{\infty}(X)\).

  2. (ii)

    ρ(1 X )=1 X .

  3. (iii)

    [ρ(f)]=[f] for all \(f\in \mathcal {L}_{\infty}(X)\).

  4. (iv)

    ρ(f)=ρ(g) for all \(f,g\in \mathcal {L}_{\infty}(X)\) with [f]=[g].

If \(\rho:\mathcal {L}_{\infty}(X)\to \mathcal {L}_{\infty}(X)\) is a ν-lifting and \(f\in \mathcal {L}_{\infty}(X)\) satisfies f(x)≥0, then \(h:=\sqrt{f} \in \mathcal {L}_{\infty}(X)\), and hence we obtain

$$\rho(f) = \rho\bigl(h^2\bigr) = \rho^2(h) \geq0.$$

Consequently, ρ also respects the ordering on \(\mathcal {L}_{\infty}(X)\). From this we can conclude that ρ is also continuous with respect to ∥⋅∥. Indeed, if we have an \(f\in \mathcal {L}_{\infty}(X)\) with ∥f≤1, we find −1 X f1 X , and since ρ is linear and respects the ordering, we obtain

$$-\boldsymbol {1}_X = \rho(-\boldsymbol {1}_X ) \leq\rho(f)\leq\rho(\boldsymbol {1}_X) = \boldsymbol {1}_X;$$

that is, ∥ρ(f)∥≤1. In other words, we have shown ∥ρ∥≤1. Moreover, from conditions (iii) and (iv), we can conclude that

$$\ker\rho= \bigl\{ f\in \mathcal {L}_{\infty}(X): \nu\bigl(\{f\neq0\}\bigr) =0 \bigr\} =[0]_\sim,$$

where the equivalence class [0] is the one in \(\mathcal {L}_{\infty}(X)\). Consequently, the quotient map \(\bar{\rho}:\mathcal {L}_{\infty}(X)_{/\sim} \to \mathcal {L}_{\infty}(X)\), defined by [f]f, is a well-defined, bounded, linear, and injective operator with \(\Vert \bar{\rho}\Vert \leq1\). Moreover, from condition (ii), we obtain

$$\bigl[\bar{\rho}\bigl([f]_\sim\bigr)\bigr]_\sim= \bigl[\rho(f)\bigr]_\sim= [f]_\sim, \quad f\in \mathcal {L}_{\infty}(X);$$

that is, \([\,\cdot \,]_{\sim}\circ\bar{\rho}= \operatorname {id}_{\mathcal {L}_{\infty}(X)_{/\sim}}\). Considering the isometric isomorphism \(\bar{I}_{\nu}^{-1}: {L}_{\infty}(\nu)\to \mathcal {L}_{\infty}(X)_{/\sim}\), we can now define the bounded, linear, and injective operator

$$ \varphi:=\bar{\rho}\circ\bar{I}_\nu^{-1}:{L}_{\infty}(\nu)\to \mathcal {L}_{\infty}(X),$$
(58)

which, by construction and our previous considerations, satisfies ∥φ∥≤1 and

$$ \bigl[\varphi\bigl([f]_\sim\bigr)\bigr]_\sim=[f]_\sim, \quad f\in \mathcal {L}_{\infty}(\nu),$$
(59)

where the outer [ ⋅ ] on the left-hand side and [ ⋅ ] on the right-hand side both refer to equivalence classes in \(\mathcal {L}_{\infty}(X)_{/\sim}\). Applying \(\bar{I}_{\nu}\) to both sides, we conclude that \(I_{\nu}\circ\varphi= \operatorname {id}_{{L}_{\infty}(\nu)}\); i.e., φ is the desired right-inverse of I ν . In other words, φ picks from each equivalence class of L (ν) a bounded representative in a linear and continuous way. Moreover, it is not hard to check that φ is actually an algebra homeomorphism with φ([1 X ])=1 X , and, in addition, it also respects the ordering of L (ν) and \(\mathcal {L}_{\infty}(X)\). Consequently, the map φ has highly desirable properties. So far, however, we have not shown that there actually exists a ν-lifting, and hence we do not know whether a map φ with the above properties exists. This gap is filled by the following theorem:

Theorem B.2

(Existence of liftings)

Let \((X,\mathcal {A})\) be a measurable space and ν be a σ-finite measure on \((X,\mathcal {A})\) such that \(\mathcal {A}\) is ν-complete. Then there exists a ν-lifting \(\rho:\mathcal {L}_{\infty}(X)\to \mathcal {L}_{\infty}(X)\).

Proof

See [39, Theorem 3.2], where we note that σ-finite measures are strictly localizable and Carathéodory completeness of \(\mathcal {A}\) equals ν-completeness. □

As [39, Theorem 3.2] shows, the σ-finiteness of the measure is not necessary if one assumes, instead, some other, technically more involved conditions on ν. Finally, further information on liftings and their applications can be found in, e.g., [24, 42].

Rights and permissions

Reprints and permissions

About this article

Cite this article

Steinwart, I., Scovel, C. Mercer’s Theorem on General Domains: On the Interaction between Measures, Kernels, and RKHSs. Constr Approx 35, 363–417 (2012). https://doi.org/10.1007/s00365-012-9153-3

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00365-012-9153-3

Keywords

Mathematics Subject Classification (2000)

Navigation