Skip to main content

Appropriate Kernel Functions for Support Vector Machine Learning with Sequences of Symbolic Data

  • Conference paper
Deterministic and Statistical Methods in Machine Learning (DSMML 2004)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 3635))

Abstract

In classification problems, machine learning algorithms often make use of the assumption that (dis)similar inputs lead to (dis)similar outputs. In this case, two questions naturally arise: what does it mean for two inputs to be similar and how can this be used in a learning algorithm? In support vector machines, similarity between input examples is implicitly expressed by a kernel function that calculates inner products in the feature space. For numerical input examples the concept of an inner product is easy to define, for discrete structures like sequences of symbolic data however these concepts are less obvious. This article describes an approach to SVM learning for symbolic data that can serve as an alternative to the bag-of-words approach under certain circumstances. This latter approach first transforms symbolic data to vectors of numerical data which are then used as arguments for one of the standard kernel functions. In contrast, we will propose kernels that operate on the symbolic data directly.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Schölkopf, B.: The kernel trick for distances. Technical report, Microsoft Research (2000)

    Google Scholar 

  2. Joachims, T.: Learning to Classify Text Using Support Vector Machines. Kluwer Academic Publishers, Dordrecht (2002)

    Google Scholar 

  3. Vanschoenwinkel, B., Manderick, B.: A weighted polynomial information gain kernel for resolving pp attachment ambiguities with support vector machines. In: The Proceedings of the Eighteenth International Joint Conferences on Artificial Intelligence (IJCAI 2003), pp.133–138 (2003)

    Google Scholar 

  4. Vapnik, V.: The Nature of Statistical Learning Theory. Springer, Heidelberg (1998)

    Google Scholar 

  5. Cristianini, N., Shawe-Taylor, J.: An Introduction to Support Vector Machines and other Kernel-based Learning Methods. Cambridge University Press, Cambridge (2000)

    Google Scholar 

  6. Haussler, D.: Convolution kernels on discrete structures (1999)

    Google Scholar 

  7. Hein, M., Bousquet, O.: Maximal margin classification for metric spaces. In: Proceedings of Learning Theory and Kernel Machines: 16th Annual Conference on Learning Theory and 7th Kernel Workshop, Colt/Kernel 2003, Washington, DC, USA, pp. 72–86. Springer, Heidelberg (2003)

    Google Scholar 

  8. Rost, B., Sander, C.: Improved prediction of protein secondary structure by use of sequence profiles and neural networks. In: Proceedings of the National Academy of Science USA, vol. 90, pp. 7558–7562 (1993)

    Google Scholar 

  9. Chih-Chung, C., Chi-Jen, L.: LIBSVM: A Library for Support Vector Machines (2004)

    Google Scholar 

  10. McNamee, P., Mayfield, J.: Entity extraction without language-specific resources. In: Proceedings of CoNLL 2002, pp. 183–186. Taipei, Taiwan (2002)

    Google Scholar 

  11. Kudoh, T., Matsumoto, Y.: Use of support vector learning for chunk identification. In: Proceedings of CoNLL 2000 and LLL 2000, Lisbon, Portugal (2000)

    Google Scholar 

  12. Takeuchi, K., Collier, N.: Use of support vector machines in extended named entity. In: Proceedings of CoNLL 2002, pp. 119–125. Taipei, Taiwan (2002)

    Google Scholar 

  13. Hua, S., Sun, Z.: A novel method of protein secondary structure prediction with high segment overlap measure: Support vector machine approach. Journal of Molecular Biology 308 (2), 397–407 (2001)

    Article  Google Scholar 

  14. Degroeve, S.: Design and Evaluation of a Linear Classification Strategy for Gene Structural Element Recognition. PhD thesis, Universiteit Gent, Faculty of Sciences, Gent, Belgium (2004)

    Google Scholar 

  15. Levenshtein, V.: Binary codes capable of correcting deletions, insertions and reversals. Soviet Physics - Doklady 10, 10, 707–710 (1966)

    MathSciNet  Google Scholar 

  16. Nerbonne, J., Heeringa, W., Kleiweg, P.: Comparison and classification of dialects. In: Preceedings of EACL 1999 (1999)

    Google Scholar 

  17. Daelemans, W., Zavrel, J., van der Sloot, K., van den Bosch, A.: Timbl: Tilburg memory-based learner, version 4.3. Technical report, Tilburg University and University of Antwerp (2002)

    Google Scholar 

  18. Vanschoenwinkel, B.: A discrete kernel approach to support vector machine learning in language independent named entity recognition. In: Proceedings of BENELEARN 2004 (Annual Machine Learning Conference of Belgium and The Netherlands), pp. 154–161 (2004)

    Google Scholar 

  19. Zavrel, J., Daelemans, W., Veenstra, J.: Resolving pp attachment ambiguities with memory-based learning. In: Proceedings CoNNL, Madrid, Computational Linguistics, Tilburg University, pp. 136–144 (1997)

    Google Scholar 

  20. Leslie, C., Eskin, E., Noble, W.S.: The spectrum kernel: A string kernel for svm protein classification. In: Proceedings of the Pacific Symposium on Biocomputing, pp. 564–575 (2002)

    Google Scholar 

  21. Kuang, R., Ie, E., Wang, K., Siddiqi, M., Freund, Y., Leslie, C.: Profile-based string kernels for remote homology detection and motif extraction. In: Computational Systems Biology Conference (2004)

    Google Scholar 

  22. Saunders, C., Tschach, H., Shawe-Taylor, J.: Syllables and other string kernel extensions. In: Proceedings of the 19th International Conference on Machine Learning, pp. 530–537. Morgan Kaufmann, San Francisco (2002)

    Google Scholar 

  23. Rost, B.: Rising accuracy of protein secondary structure prediction. Marcel Dekker, New York (2003)

    Google Scholar 

  24. Dayhoff, M., Schwartz, R.: Matrices for detecting distant relationships. Atlas of Protein Sequences, 353–358 (1979)

    Google Scholar 

  25. Henikoff, S., Henikoff, J.: Amino acid substitution matrices from protein blocks. Proceedings of the National Academy of Science USA 89(2), 10915–10919 (1992)

    Article  Google Scholar 

  26. Cuff, J., Barton, G.: Evaluation and improvement of multiple sequence methods for protein secondary structure prediction. Proteins: Struct., Funct., Genet. 35, 508–519 (1999)

    Google Scholar 

  27. Riis, S.K., Krogh, A.: Improving prediction of protein secondary structure using structured neural networks and multiple sequence alignments. Journal Of Computational Biology 3, 163–183 (1996)

    Article  Google Scholar 

  28. Noreen, E.W.: Computer-Intensive Methods for Testing Hypotheses. John Wiley & Sons, Chichester (1989)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2005 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Vanschoenwinkel, B., Manderick, B. (2005). Appropriate Kernel Functions for Support Vector Machine Learning with Sequences of Symbolic Data. In: Winkler, J., Niranjan, M., Lawrence, N. (eds) Deterministic and Statistical Methods in Machine Learning. DSMML 2004. Lecture Notes in Computer Science(), vol 3635. Springer, Berlin, Heidelberg. https://doi.org/10.1007/11559887_16

Download citation

  • DOI: https://doi.org/10.1007/11559887_16

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-29073-5

  • Online ISBN: 978-3-540-31728-9

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics