Skip to main content
Log in

Multicategory Classification by Support Vector Machines

  • Published:
Computational Optimization and Applications Aims and scope Submit manuscript

Abstract

We examine the problem of how to discriminate between objects of three or more classes. Specifically, we investigate how two-class discrimination methods can be extended to the multiclass case. We show how the linear programming (LP) approaches based on the work of Mangasarian and quadratic programming (QP) approaches based on Vapnik's Support Vector Machine (SVM) can be combined to yield two new approaches to the multiclass problem. In LP multiclass discrimination, a single linear program is used to construct a piecewise-linear classification function. In our proposed multiclass SVM method, a single quadratic program is used to construct a piecewise-nonlinear classification function. Each piece of this function can take the form of a polynomial, a radial basis function, or even a neural network. For the k > 2-class problems, the SVM method as originally proposed required the construction of a two-class SVM to separate each class from the remaining classes. Similarily, k two-class linear programs can be used for the multiclass problem. We performed an empirical study of the original LP method, the proposed k LP method, the proposed single QP method and the original k QP methods. We discuss the advantages and disadvantages of each approach.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

  1. S. Aeberhard, D. Coomans, and O. de Vel, “Comparison of classifiers in high dimensional settings,” Technical Report 92–02, Departments of Computer Science and of Mathematics and Statistics, James Cook University of North Queensland, 1992.

  2. K. P. Bennett, “Decision tree construction via linear programming,” in Proceedings of the 4th Midwest Artificial Intelligence and Cognitive Science Society Conference, Utica, Illinois, pages 97–101, 1992.

  3. K. P. Bennett and E. J. Bredensteiner, “Geometry in learning,” In C. Gorini, E. Hart, W. Meyer, and T. Phillips, editors, Geometry at Work, Washington, D.C., 1998. Mathematical Association of America, To appear.

  4. K. P. Bennett and O. L. Mangasarian, “Neural network training via linear programming,” in P. M. Pardalos, editor, Advances in Optimization and Parallel Computing, North Holland, pp. 56–67, 1992.

  5. K. P. Bennett and O. L. Mangasarian, “Multicategory discrimination via linear programming,” Optimization Methods and Software, vol. 3, pp. 27–39, 1994.

    Google Scholar 

  6. K. P. Bennett and O. L. Mangasarian, “Serial and parallel multicategory discrimination,” SIAM Journal on Optimization, Vol. 4, pp. 722–734, 1994.

    Google Scholar 

  7. K. P. Bennett, D. H. Wu, and L. Auslender, “On support vector decision trees for database marketing,” R.P.I. Math Report No. 98–100, Rensselaer Polytechnic Institute, Troy, NY, 1998.

    Google Scholar 

  8. V. Blanz, B. Schölkopf, H. Bülthoff, C. Burges, V. Vapnik, and T. Vetter, “Comparison of view-based object recognition algorithms using realistic 3D models,” in C. von der Malsburg, W. von Seelen, J. C. Vorbrüggen, and B. Sendhoff, editors, Artificial Neural Networks — ICANN'96, vol. 1112 of Lecture Notes in Computer Science, Springer: Berlin, 1996.

    Google Scholar 

  9. P. S. Bradley and O. L. Mangasarian, “Feature selection via concave minimization and support vector machines,” Machine Learning Proceedings of the Fifteenth International Conference (ICML '98) (J. Shavlik, editor), Morgan Kaufmann: San Francisco, CA, pp. 82–90, 1998.

    Google Scholar 

  10. E. J. Bredensteiner and K. P. Bennett, “Feature minimization within decision trees,” Computational Optimization and Applications, vol. 10, pp. 111–126, 1998.

    Google Scholar 

  11. C. Cortes and V. N. Vapnik, “Support vector networks,” Machine Learning, vol. 20, pp. 273–297, 1995.

    Google Scholar 

  12. R. Courant and D. Hilbert, Methods of Mathematical Physics, J. Wiley: New York, 1953.

    Google Scholar 

  13. Y. Le Cun, B. Boser, J. S. Denker, D. Henderson, R. E. Howard, W. Hubbard, and L. J. Jackel, “Backpropagation applied to handwritten zip code recognition,” Neural Computation, vol. 1. pp. 541–551, 1989.

    Google Scholar 

  14. G. B. Dantzig, Linear Programming and Extensions, Princeton University Press, Princeton, New Jersey, 1963.

    Google Scholar 

  15. I. W. Evett and E. J. Spiehler, “Rule induction in forensic science,” Technical report, Central Research Establishment, Home Office Forensic Science Service, Aldermaston, Reading, Berkshire RG7 4PN, 1987.

    Google Scholar 

  16. O. L. Mangasarian, “Linear and nonlinear separation of patterns by linear programming,” Operations Research, vol. 13, pp. 444–452, 1965.

    Google Scholar 

  17. O. L. Mangasarian, “Multi-surface method of pattern separation,” IEEE Transactions on Information Theory, vol. IT-14, pp. 801–807, 1968.

    Google Scholar 

  18. O. L. Mangasarian, Nonlinear Programming, McGraw-Hill: New York, 1969.

    Google Scholar 

  19. O. L. Mangasarian, “Mathematical programming in machine learning,” In G. DiPillo and F. Giannessi, editors, Proceedings of Nonlinear Optimization and Applications Workshop, pp. 283–295, Plenum Press: New York, 1996.

    Google Scholar 

  20. O. L. Mangasarian, “Arbitrary-norm separating plane,” Mathematical Programming Technical Report 97–07, University of Wisconsin-Madison, 1997.

  21. O. L. Mangasarian, W. N. Street, and W. H. Wolberg, “Breast cancer diagnosis and prognosis via linear programming,” Operations Research, vol. 43, pp. 570–577, 1995.

    Google Scholar 

  22. P. M. Murphy and D. W. Aha, “UCI repository of machine learning databases,” [http://www.ics.uci.edu/~mlearn/MLRepository.html], Department of Information and Computer Science, University of California, Irvine, California, 1994.

    Google Scholar 

  23. B. A. Murtagh and M. A. Saunders, “MINOS 5.4 user's guide,” Technical Report SOL 83.20, Stanford University, 1993.

  24. A. Roy, S. Govil, and R. Miranda, An algorithm to generate radial basis function (RBF)-like nets for classification problems, Neural Networks, vol. 8, pp. 179–202, 1995.

    Google Scholar 

  25. A. Roy, L. S. Kim, and S. Mukhopadhyay, “A polynomial time algorithm for the construction and training of a class of multilayer perceptrons,” Neural Networks, vol. 6, pp. 535–545, 1993.

    Google Scholar 

  26. A. Roy and S. Mukhopadhyay, “Pattern classification using linear programming,” ORSA Journal of Computing, vol. 3, pp. 66–80, 1990.

    Google Scholar 

  27. B. Schölkopf, C. Burges, and V. Vapnik, “Incorporating invariances in support vector machines,” in C. von der Malsburg, W. von Seelen, J. C. Vorbrüggen, and B. Sendhoff, editors, Artificial Neural Networks — ICANN'96, vol. 1112 of Lecture Notes in Computer Science, pp. 47–52, Springer: Berlin 1996.

    Google Scholar 

  28. B. Schölkopf, K. Sung, C. Burges, F. Girosi, P. Niyogi, T. Poggio, and V. Vapnik, “Comparing support vector machines with gaussian kernels to radial basis function classifiers,” AI Memo No. 1599; CBCL Paper No. 142, Massachusetts Institute of Technology, Cambridge, 1996.

    Google Scholar 

  29. V. Vapnik, The Nature of Statistical Learning Theory, Springer-Verlag, 1995.

  30. V. N. Vapnik and A. Ja. Chervonenkis, Theory of Pattern Recognition, Nauka, Moscow, 1974. In Russian.

    Google Scholar 

  31. W. H. Wolberg and O. L. Mangasarian, “Multisurface method of pattern separation for medical diagnosis applied to breast cytology,” Proceedings of the National Academy of Sciences U.S.A., vol. 87, pp. 9193–9196, 1990.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Rights and permissions

Reprints and permissions

About this article

Cite this article

Bredensteiner, E.J., Bennett, K.P. Multicategory Classification by Support Vector Machines. Computational Optimization and Applications 12, 53–79 (1999). https://doi.org/10.1023/A:1008663629662

Download citation

  • Issue Date:

  • DOI: https://doi.org/10.1023/A:1008663629662

Navigation