Skip to main content

Choosing the Kernel Parameters for the Directed Acyclic Graph Support Vector Machines

  • Conference paper
  • 3649 Accesses

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 4571))

Abstract

The directed acyclic graph support vector machines (DAGSVMs) have been shown to be able to provide classification accuracy comparable to the standard multiclass SVM extensions such as Max Wins methods. The algorithm arranges binary SVM classifiers as the internal nodes of a directed acyclic graph (DAG). Each node represents a classifier trained for the data of a pair of classes with the specific kernel. The most popular method to decide the kernel parameters is the grid search method. In the training process, classifiers are trained with different kernel parameters, and only one of the classifiers is required for the testing process. This makes the training process time-consuming. In this paper we propose using separation indexes to estimate the generalization ability of the classifiers. These indexes are derived from the inter-cluster distances in the feature spaces. Calculating such indexes costs much less computation time than training the corresponding SVM classifiers; thus the proper kernel parameters can be chosen much faster. Experiment results show that the testing accuracy of the resulted DAGSVMs is competitive to the standard ones, and the training time can be significantly shortened.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   84.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Vapnik, V.: Statistical Learning Theory. Wiley, New York (1998)

    MATH  Google Scholar 

  2. Kreß el, U.: Pairwise classification and support vector machines. In: Schölkopf, B., Burges, C.J.C., Smola, A.J. (eds.) Advances in Kernel Methods – Support Vector Learning, MIT Press, Cambridge, MA (1999)

    Google Scholar 

  3. Bottu, L., Cortes, C., Denker, J., Drucker, H., Guyon, I., Jackel, L., LeCun, Y., Muller, U., Sackinger, E., Simard, P., Vapnik, V.: Comparison of classifier methods: A case study in handwritten digit recognition. In: Proc. Int. Conf. Pattern Recognition, pp. 77–87 (1994)

    Google Scholar 

  4. Platt, J.C., Cristianini, N., Show-Tayler, J.: Large margin DAG’s for multiclass classification. In: Advances in Neural Information Processing Systems, vol. 12, pp. 547–553. MIT Press, Cambridge, MA (2000)

    Google Scholar 

  5. Hsu, C.-W., Lin, C.-J.: A comparison of methods for multiclass support vector machines. IEEE trans. Neural Networks 13(2), 415–425 (2002)

    Article  Google Scholar 

  6. Keerthi, S.S., Lin, C.-J.: Asymptotic behaviors of support vector machines with Gaussian kernel. Neural Computation 15(7), 1667–1689 (2003)

    Article  MATH  Google Scholar 

  7. Lin, H.-T., Lin, C.-J.: A study on sigmoid kernels for SVM and the training of non-PSD kernels by SMO-type methods, Technical report, Department of Computer Science and Information Engineering, National Taiwan University

    Google Scholar 

  8. Rojas, S.A., Fernandez-Reyes, D.: Adapting Multiple Kernel Parameters for Support Vector Machines using Genetic Algorithms. The 2005 IEEE Congress on Evolutionary Computation 1, 626–631 (2005)

    Article  Google Scholar 

  9. Liang, X., Liu, F.: Choosing multiple parameters for SVM based on genetic algorithm. In: 6th International Conference on Signal Processing, vol. 1, pp. 117–119 (August 26-30, 2002)

    Google Scholar 

  10. Liu, H.-J., Wang, Y.-N., Lu, X.-F.: A Method to Choose Kernel Function and its Parameters for Support Vector Machines. In: Proceedings of 2005 International Conference on Machine Learning and Cybernetics, vol. 7, pp. 4277–4280 (August 18–21, 2005)

    Google Scholar 

  11. Hush, D., Scovel, C.: Polynomial-time decomposition algorithms for support vector machines. Machine Learning 51, 51–71 (2003)

    Article  MATH  Google Scholar 

  12. Platt, J.C.: Sequential minimal optimization: A fast algorithm for training support vector machines, Microsoft Research, Technical Report MST-TR-98-14 (1998)

    Google Scholar 

  13. Maruyama, K.-I., Maruyama, M., Miyao, H., Nakano, Y.: A method to make multiple hypotheses with high cumulative recognition rate using SVMs. Pattern Recognition 37, 241–251 (2004)

    Article  MATH  Google Scholar 

  14. Debnath, R., Takahashi, H.: An efficient method for tuning kernel parameter of the support vector machine. In: IEEE International Symposium on Communications and Information Technology, vol. 2, pp. 1023–1028 (October 26–29, 2004)

    Google Scholar 

  15. Bi, L.-P., Huang, H., Zheng, Z.-Y., Song, H.-T.: New Heuristic for Determination Gaussian Kernel’s Parameter. In: Proceedings of 2005 International Conference on Machine Learning and Cybernetics, vol. 7, pp. 4299–4304 (August 18–21, 2005)

    Google Scholar 

  16. Takahashi, F., Abe, S.: Optimizing directed acyclic graph support vector machines. In: ANNPR 2003. Proc. Artificial Neural Networks in Pattern Recognition, pp. 166–170 (September 2003)

    Google Scholar 

  17. Phetkaew, T., Kijsirikul, B., Rivepiboon, W.: Reordering adaptive directed acyclic graphs: An improved algorithm for multiclass support vector machines. In: IJCNN 2003. Proc. Internat. Joint Conf. on Neural Networks, vol. 2, pp. 1605–1610 (2003)

    Google Scholar 

  18. Wu, K.-P., Wang, S.-D.: Choosing the Kernel parameters of Support Vector Machines According to the Inter-cluster Distance. In: IJCNN 2006. Proc. Internat. Joint Conf. on Neural Networks (2006)

    Google Scholar 

  19. Bezdek, C., Pal, N.R.: Some New Indexes of Cluster Validity. IEEE Transaction on Systems, Man, And Cybernetics-Part B: Cybernetics 28(3), 301–315 (1998)

    Article  Google Scholar 

  20. Michie, D., Spiegelhalter, D.J., Taylor, C.C.: Machine learning, neural and statistical classification (1994), [Online available]: http://www.amsta.leeds.ac.uk/~charles/statlog/

  21. Chang, C.-C., Lin, C.-J.: LIBSVM: a library for support vector machines, [Online] available from World Wide Web: http://www.csie.ntu.edu.tw/~cjlin/libsvmtools/datasets/

Download references

Author information

Authors and Affiliations

Authors

Editor information

Petra Perner

Rights and permissions

Reprints and permissions

Copyright information

© 2007 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Wu, KP., Wang, SD. (2007). Choosing the Kernel Parameters for the Directed Acyclic Graph Support Vector Machines. In: Perner, P. (eds) Machine Learning and Data Mining in Pattern Recognition. MLDM 2007. Lecture Notes in Computer Science(), vol 4571. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-73499-4_21

Download citation

  • DOI: https://doi.org/10.1007/978-3-540-73499-4_21

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-73498-7

  • Online ISBN: 978-3-540-73499-4

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics