A Modified SMO Algorithm for SVM Regression and Its Application in Quality Prediction of HP-LDPE

  • Hengping Zhao
  • Jinshou Yu
Conference paper
Part of the Lecture Notes in Computer Science book series (LNCS, volume 3610)

Abstract

A modified sequential minimal optimization (SMO) algorithm for support vector machine (SVM) regression is proposed based on Shevade’s SMO-1 algorithm. The main improvement is that a modified heuristics method is used in this modified SMO algorithm to choose the first Lagrange multiplier when optimizing the Lagrange multipliers corresponding to the non-boundary examples. To illustrate the validity of the proposed modified SMO algorithm, a benchmark dataset and a practical application in predicting the melt index of high-pressure low-density polyethylene (HP-LDPE) are used; the results demonstrate that this modified SMO algorithm is faster in most cases with the same parameters setting and more likely to obtain the better generalization performance than Shevade’s SMO-1 algorithm.

Keywords

Support Vector Machine Mean Square Error Lagrange Multiplier Sequential Minimal Optimization Training Support Vector Machine 
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. 1.
    Vapnik, V.N.: The Nature of Statistical Learning Theory, 2nd edn. Springer, New York (2000)MATHGoogle Scholar
  2. 2.
    Du, S., Wu, T.: Support Vector Machines for Regression. Journal of System Simulation 15(11), 1580–1585 (2003)Google Scholar
  3. 3.
    Li, J., Zhang, B.: An Improvement Algorithm to Sequential Minimal Optimization. Journal of Software 14(3), 918–924 (2003)MATHGoogle Scholar
  4. 4.
    Osuna, F.R., Girosi, F.: An Improved Training Algorithm for Support Vector Machines. In: Proc. of IEEE NNSP 1997, pp. 276–285 (1997)Google Scholar
  5. 5.
    Platt, J.C.: Fast Training of Support Vector Machines Using Sequential Minimal Optimization. In: Schölkopf, B., Burges, C.J.C., Smola, A.J. (eds.) Advances in Kernel Methods - Support Vector Learning, pp. 185–208. MIT Press, Cambridge (1998)Google Scholar
  6. 6.
    Smola, A.J.: Learning with Kernels. Ph.D. dissertation, GMD, Birlinghoven, Germany (1998)Google Scholar
  7. 7.
    Smola, A.J., Schölkopf, B.: A Tutorial on Support Vector Regression. NeuroColt Technical Report, TR-1998-030, Royal Holloway College (1998)Google Scholar
  8. 8.
    Shevade, S.K., Keerthi, S.S., Bhattacharyya, C., Murthy, K.R.K.: Improvements to SMO Algorithm for SVM Regression. IEEE Transactions on Neural Networks 11(5), 1188–1194 (2000)CrossRefGoogle Scholar
  9. 9.
    Vapnik, V.N.: Statistical Learning Theory. Wiley, New York (1998)MATHGoogle Scholar
  10. 10.
    Flake, G.W., Lawrence, S.: Efficient SVM Regression Training with SMO. Machine Learning 46, 271–290 (2002)MATHCrossRefGoogle Scholar
  11. 11.
    Keerthi, S.S., Shevade, S.K., Bhattacharyya, C., Murthy, K.R.K.: A Fast Iterative Nearest Point Algorithm for Support Vector Machine Classifier Design. IEEE Transaction on Neural Networks 11(1), 124–136 (2000)CrossRefGoogle Scholar
  12. 12.
    Blake, C.L., Merz, C.J.: UCI Repository of Machine Learning Databases. Dept. Inform. Comput. Sci., Univ. California, Irvine, CA (1998), Online Available http://www.ics.uci.edu/~mlearn/MLRepository.html

Copyright information

© Springer-Verlag Berlin Heidelberg 2005

Authors and Affiliations

  • Hengping Zhao
    • 1
  • Jinshou Yu
    • 1
  1. 1.Research Institution of AutomationEast China University of Science and TechnologyShanghaiPRC

Personalised recommendations