Skip to main content

Reducing SVR Support Vectors by Using Backward Deletion

  • Conference paper
Knowledge-Based Intelligent Information and Engineering Systems (KES 2008)

Abstract

Support Vector Regression (SVR) is one of the most famous sparse kernel machines which inherits many advantages of Support Vector Machines (SVM). However, since the number of support vectors grows rapidly with the increase of training samples, sparseness of the SVR is sometimes insufficient. In this paper, we propose two methods which reduce the SVR support vectors using backward deletion. Experiments show our method can dramatically reduce the number of support vectors without sacrificing the generalization performance.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Vapnik, V.N.: The Nature of Statistical Learning Theory. Springer, Heidelberg (1995)

    MATH  Google Scholar 

  2. Cristianini, N., Shawe-Taylor, J.: An Introduction to Support Vector Machines and other kernel-based learning methods. Cambridge University Press, Cambridge (2000)

    Google Scholar 

  3. Chang, C.C., Lin, C.J.: Libsvm - a library for support vector machines, http://www.csie.ntu.edu.tw/~cjlin/libsvm/

  4. Wang, L., Bo, L., Jiao, L.: Sparse kernel ridge regression using backward deletion. In: Yang, Q., Webb, G. (eds.) PRICAI 2006. LNCS (LNAI), vol. 4099, pp. 365–374. Springer, Heidelberg (2006)

    Google Scholar 

  5. Asuncion, A., Newman, D.: UCI machine learning repository (2007), http://www.ics.uci.edu/~mlearn/MLRepository.html

  6. Saunders, G., Gammerman, A., Vovk, V.: Ridge regression learning algorithm in dual variables. In: Proc. 15th International Conf. on Machine Learning, pp. 515–521. Morgan Kaufmann, San Francisco (1998)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Ignac Lovrek Robert J. Howlett Lakhmi C. Jain

Rights and permissions

Reprints and permissions

Copyright information

© 2008 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Karasuyama, M., Takeuchi, I., Nakano, R. (2008). Reducing SVR Support Vectors by Using Backward Deletion. In: Lovrek, I., Howlett, R.J., Jain, L.C. (eds) Knowledge-Based Intelligent Information and Engineering Systems. KES 2008. Lecture Notes in Computer Science(), vol 5179. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-85567-5_10

Download citation

  • DOI: https://doi.org/10.1007/978-3-540-85567-5_10

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-85566-8

  • Online ISBN: 978-3-540-85567-5

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics