Abstract
Let (X, Y) be a pair of random variables withsupp(X)⊆[0,1]l andEY 2<∞. Letm * be the best approximation of the regression function of (X, Y) by sums of functions of at mostd variables (1≤d≤l). Estimation ofm * from i.i.d. data is considered.
For the estimation interaction least squares splines, which are defined as sums of polynomial tensor product splines of at mostd variables, are used. The knot sequences of the tensor product splines are chosen equidistant. Complexity regularization is used to choose the number of the knots and the degree of the splines automatically using only the given data.
Without any additional condition on the distribution of (X, Y) the weak and strongL 2-consistency of the estimate is shown. Furthermore, for everyp≥1 and every distribution of (X, Y) withsupp(X)⊆[0,1]l,y bounded andm * p-smooth, the integrated squared error of the estimate achieves up to a logarithmic factor the (optimal) rate\(n^{ - \tfrac{{2p}}{{2p + d}}} \)
Similar content being viewed by others
References
Barron AR, Birgé L, Massart P (1995) Risk bounds for model selection via penalization. Technical Report No. 95.54, Université Paris Sud
Chen Z (1991) Interaction spline models and their convergence rates. Ann. Statist. 19:1855–1868
de Boor C (1978) A practical guide to splines. Springer, New York
Devroye L, Györfi L, Lugosi G (1996) A probabilistic theory of pattern recognition. Springer, Berlin
Dudley R (1978) Central limit theorems for empirical measures. Annals of Probability 6:899–929
Haussler D (1992) Decision theoretic generalizations of the PAC model for neural net and other learning applications. Information and Computation 100:78–150
Kohler M (1996) Universally consistent regression function estimation using hierarchical B-splines. Preprint 96-11, Mathematisches Institut A, Universität Stuttgart. Submitted to Journal of Multivariate Analysis
Kohler M (1997) On the universal consistency of a least squares spline regression estimator. Math. Methods of Statistics 6:349–364
Krzyzak A, Linder T (1996) Radial basis function networks and complexity regularization in function learning. To appear in IEEE Transaction on Neural Networks 9, 1998
Lee WS, Bartlett PL, Williamson RC (1996) Efficient agnostic learning of neural networks with bounded fan-in. IEEE Trans. Inform. Theory 42:2118–2132
Lugosi G, Zeger K (1995) Nonparametric estimation via empirical risk minimization. IEEE Trans. Inform. Theory 41:677–687
Pollard D (1984) Convergence of stochastic processes. Springer-Verlag, New York
Schumaker L (1981) Spline functions: Basic theory. Wiley, New York
Stone CJ (1982) Optimal global rates of convergence for nonparametric regression. Ann. Statist. 10:1040–1053
Stone CJ (1985) Additive regression and other nonparametric models. Ann. Statist. 13:689–705
Stone CJ (1986) The dimensionality reduction principle for generalized additive models. Ann. Statist. 14:590–606
Stone CJ (1994) The use of polynomial splines and their tensor products in multivariate function estimation. Ann. Statist. 22:118–184
Author information
Authors and Affiliations
Rights and permissions
About this article
Cite this article
Kohler, M. Nonparametric regression function estimation using interaction least squares splines and comlexity regularization. Metrika 47, 147–163 (1998). https://doi.org/10.1007/BF02742869
Received:
Issue Date:
DOI: https://doi.org/10.1007/BF02742869