Skip to main content
Log in

Statistical Learning Theory: A Primer

  • Published:
International Journal of Computer Vision Aims and scope Submit manuscript

Abstract

In this paper we first overview the main concepts of Statistical Learning Theory, a framework in which learning from examples can be studied in a principled way. We then briefly discuss well known as well as emerging learning techniques such as Regularization Networks and Support Vector Machines which can be justified in term of the same induction principle.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

  • Alon, N., Ben-David, S., Cesa-Bianchi, N., and Haussler, D. 1993. Scale-sensitive dimensions, uniform convergence, and learnability. Symposium on Foundations of Computer Science.

  • Cortes, C. and Vapnik, V. 1995. Support vector networks. Machine Learning, 20:1–25.

    Google Scholar 

  • Devroye, L., Györfi, L., and Lugosi, G. 1996. A Probabilistic Theory of Pattern Recognition, No. 31 in Applications of Mathematics. Springer: New York.

    Google Scholar 

  • Evgeniou, T., Pontil, M., Papageorgiou, C., and Poggio, T. 2000. Image representations for object detection using kernel classifiers. In Proceedings ACCV. Taiwan, p. To appear.

  • Evgeniou, T., Pontil, M., and Poggio, T. 1999. A unified framework for Regularization Networks and Support Vector Machines. A.I. Memo No. 1654, Artificial Intelligence Laboratory, Massachusetts Institute of Technology.

  • Ezzat, T. and Poggio, T. 1996. Facial analysis and synthesis using image-based models. In Face and Gesture Recognition. pp. 116–121.

  • Girosi, F., Jones, M., and Poggio, T. 1995. Regularization theory and neural networks architectures. Neural Computation, 7:219–269.

    Google Scholar 

  • Jaakkola, T. and Haussler, D. 1998. Probabilistic kernel regression models. In Proc. of Neural Information Processing Conference.

  • Kearns, M. and Shapire, R. 1994. Efficient distribution-free learning of probabilistic concepts. Journal of Computer and Systems Sciences, 48(3):464–497.

    Google Scholar 

  • Mohan, A. 1999. Robust object detection in images by components. Master's Thesis, Massachusetts Institute of Technology.

  • Osuna, E., Freund, R., and Girosi, F. 1997. An improved training algorithm for support vector machines. In IEEEWorkshop on Neural Networks and Signal Processing, Amelia Island, FL.

  • Papageorgiou, C., Oren, M., and Poggio, T. 1998. A general framework for object detection. In Proceedings of the International Conference on Computer Vision, Bombay, India.

  • Platt, J.C. 1998. Sequential minimal imization: A fast algorithm for training support vector machines. Technical Report MST-TR-98-14, Microsoft Research.

  • Tikhonov, A.N. and Arsenin, V.Y. 1977. Solutions of Ill-posed Problems. Washington, D.C.: W.H. Winston.

    Google Scholar 

  • Vapnik, V.N. 1998. Statistical Learning Theory. Wiley: New York.

    Google Scholar 

  • Vapnik, V.N. and Chervonenkis, A.Y. 1971. On the uniform convergence of relative frequences of events to their probabilities. Th. Prob. and its Applications, 17(2):264–280.

    Google Scholar 

  • Wahba, G. 1990. Splines Models for Observational Data. Vol. 59, Series in Applied Mathematics: Philadelphia.

Download references

Author information

Authors and Affiliations

Authors

Rights and permissions

Reprints and permissions

About this article

Cite this article

Evgeniou, T., Pontil, M. & Poggio, T. Statistical Learning Theory: A Primer. International Journal of Computer Vision 38, 9–13 (2000). https://doi.org/10.1023/A:1008110632619

Download citation

  • Issue Date:

  • DOI: https://doi.org/10.1023/A:1008110632619

Navigation