Skip to main content
Log in

Learning Rates for Regularized Classifiers Using Trigonometric Polynomial Kernels

  • Published:
Neural Processing Letters Aims and scope Submit manuscript

Abstract

Regularized classifiers are known to be a kind of kernel-based classification methods generated from Tikhonov regularization schemes, and the trigonometric polynomial kernels are ones of the most important kernels and play key roles in signal processing. The main target of this paper is to provide convergence rates of classification algorithms generated by regularization schemes with trigonometric polynomial kernels. As a special case, an error analysis for the support vector machines (SVMs) soft margin classifier is presented. The norms of Fejér operator in reproducing kernel Hilbert space and properties of approximation of the operator in L 1 space with periodic function play key roles in the analysis of regularization error. Some new bounds on the learning rate of regularization algorithms based on the measure of covering number for normalized loss functions are established. Together with the analysis of sample error, the explicit learning rates for SVM are also derived.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

  1. Arouszajiu N (1950) Theory of reproducing kernels. Trans Math Soc 68: 337–404

    Article  Google Scholar 

  2. Bartlett PL (1998) The sample complexity of pattern classificayion with neural networks: the size of the weights is more import than the size of the network. IEEE Trans Inf Theory 44: 525–536

    Article  MathSciNet  MATH  Google Scholar 

  3. Bartlett PL (2008) Fast rates for estimation error and oracle inequalities for model selection. Econom Theory 24: 545–552

    Article  MathSciNet  MATH  Google Scholar 

  4. Bousquet O (2003) New approaches to statistical learning theory. Ann Inst Stat Math 55: 371–389

    MathSciNet  MATH  Google Scholar 

  5. Bousquet O, Elisseeff A (2002) Stability and generalization. J Mach Learn Res 2: 499–526

    MathSciNet  MATH  Google Scholar 

  6. Butzer PL, Nessel RJ (1971) Fourier analysis and approximation, vol I. Birkhäuser/Academic Press, Basel

    Google Scholar 

  7. Chen DR, Wu Q, Ying YM, Zhou DX (2004) Support vector machine soft margin classifiers: error analysis. J Mach Learn Res 5: 1143–1175

    MathSciNet  MATH  Google Scholar 

  8. Cucker F, Smale S (2001) On the mathematical foundations of learning theory. Bull Am Math Soc 39: 1–49

    Article  MathSciNet  Google Scholar 

  9. Cucker F, Smale S (2002) Best choices for regularization parameters in learning theory: on the biasvariance problem. Found Comput Math 1: 413–428

    Article  MathSciNet  Google Scholar 

  10. Cucker F, Zhou DX (2007) Learning theory: an approximation theory viewpoint. Cambridge University Press, Cambridge

    Book  MATH  Google Scholar 

  11. Devroye L, Gyorfi L, Lugosi G (1997) A probabilistic theory of pattern recognition. Springer, New York

    Google Scholar 

  12. Evgeniou T, Pontil M (1999) On the V-gamma dimension for regression in reproducing Kernel Hilbert spaces. In Proceedings of algorithmic learning theory. Lecture notes in computer science, vol 1720. Springer, London, pp 106–117

  13. Evgeniou T, Pontil M, Poggio T (2000) Regularization networks and support vector machines. Adv Comput Math 13: 1–50

    Article  MathSciNet  MATH  Google Scholar 

  14. Hastie T, Tibshirani R, Friedman J (2001) The elements of statistical learning: data mining, inference, and prediction. Springer, New York

    MATH  Google Scholar 

  15. Huang CM, Lee YJ, Lin DKJ, Huang SY (2007) Model selection for support vector machines via uniform design. Comput Stat Data Anal 52(1): 335–346

    Article  MathSciNet  MATH  Google Scholar 

  16. Kutin S, Niyogi P (2002) Almost-everywhere algorithmic stability and generalization error. Technical Report TR-2002-03, Department of Computer Science, The University of Chicago

  17. Koltchinskii V, Panchenko D (2000) Rademacher processes and bounding the risk of function learning. High Dimens Probab 47: 1902–1914

    MathSciNet  Google Scholar 

  18. Martinand A, Bartlett PL (1999) Neural network learning: theoretical foundations. Cambridge University Press, Cambridge

    Google Scholar 

  19. Régis V, Jean-Philippe V (2006) Consistency and convergence rates of one-class SVMs and related algorithms. J Mach Learn Res 7: 817–854

    MathSciNet  MATH  Google Scholar 

  20. Shim J, Hwang C (2009) Support vector censored quantize regression under random censoring. Comput Stat Data Anal 53(4): 912–919

    Article  MathSciNet  MATH  Google Scholar 

  21. Smale S, Zhou DX (2004) Shannon sampling and function reconstruction from point values. Bull Am Math Soc 41: 279–305

    Article  MathSciNet  MATH  Google Scholar 

  22. Steinwart I (2001) On the influence of the kernel on the consistency of support vector machines. J Mach Learn Res 2: 67–73

    MathSciNet  Google Scholar 

  23. Tong HZ, Chen DR, Li ZP (2008) Learning rates for regularized classifiers using multivariate polynomial kernels. J Complex 24: 619–631

    Article  MATH  Google Scholar 

  24. Vapnik V (1998) Statistical learning theory. Wiley, New York

    MATH  Google Scholar 

  25. Wu Q, Zhou DX (2005) SVM soft margin classifiers: linear programming versus quadratic programming. Neural Comput 17: 1160–1187

    Article  MathSciNet  MATH  Google Scholar 

  26. Wu Q, Zhou DX (2006) Analysis of support vector machine classification. J Comput Anal Appl 8: 108–134

    MathSciNet  Google Scholar 

  27. Ye GB, Zhou DX (2008) Learning and approximation by Gaussian on Riemannian manifolds. Adv Comput Math 3: 291–310

    Article  MathSciNet  Google Scholar 

  28. Wu Q, Ying YM, Zhou DX (2007) Multi-kernel regularized classifiers. J Complex 23: 108–134

    Article  MathSciNet  MATH  Google Scholar 

  29. Zhang T (2004) Statistical behavior and consistency of classification methods based on convex risk minimization. Ann Stat 32: 56–85

    Article  MATH  Google Scholar 

  30. Zhou DX (2003) Capacity of reproducing kernel spaces in learning theory. IEEE Trans Inf Theory 49: 1734–1752

    Google Scholar 

  31. Zhou DX, Jetter K (2006) Approximation with polynomial kernel and SVM classifiers. Adv Comput Math 25: 323–344

    Article  MathSciNet  MATH  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Feilong Cao.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Cao, F., Wu, D. & Lee, J. Learning Rates for Regularized Classifiers Using Trigonometric Polynomial Kernels. Neural Process Lett 35, 265–281 (2012). https://doi.org/10.1007/s11063-012-9217-1

Download citation

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11063-012-9217-1

Keywords

Navigation