Skip to main content

Efficient Sequential Minimal Optimisation of Support Vector Classifiers

  • Conference paper

Abstract

This paper describes a simple modification to the sequential minimal optimisation (SMO) training algorithm for support vector machine (SVM) classifiers, reducing training time at the expense of a small increase in memory used proportional to the number of training patterns. Results obtained on real-world pattern recognition tasks indicate that the proposed modification can more than halve the average training time.

This work was supported by the European Commission grant number IST-99-11764, as part of its Framework V IST programme.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. B. Boser, I. Guyon, and V. N. Vapnik, “A training algorithm for optimal margin classifiers,’ in Proceedings of the fifth annual workshop on computational learning theory, (Pittsburgh), pp. 144–152, ACM, 1992.

    Google Scholar 

  2. C. Cortes and V. Vapnik, “Support vector networks,” Machine Learning, vol. 20, pp. 1–25, 1995.

    Google Scholar 

  3. V. N. Vapnik, The Nature of Statistical Learning Theory. New York: Springer-Verlag, 1995.

    Book  MATH  Google Scholar 

  4. V. N. Vapnik, Statistical Learning Theory. Wiley Series on Adaptive and Learning Systems for Signal Processing, Communications and Control, New York: Wiley, 1998.

    Google Scholar 

  5. N. Cristianini and J. Shawe-Taylor, An Introduction to Support Vector Machines (and other kernel-based learning methods). Cambridge, U.K.: Cambridge University Press, 2000.

    Book  Google Scholar 

  6. J. C. Platt, “Fast training of support vector machines using sequential minimal optimization,” in Advances in Kernel Methods — Support Vector Learning (B. Schölkopf, C. J. C. Burges, and A. J. Smola, eds.), ch. 12, pp. 185–208, Cambridge, Massachusetts: MIT Press, 1999.

    Google Scholar 

  7. E. Osuna, R. Freund, and F. Girosi, “An improved training algorithm for support vector machines,” in Neural Networks for Signal Processing VII — Proceedings of the 1997 IEEE Workshop (J. Principe, L. Gile, N. Morgan, and E. Wilson, eds.), (New York), pp. 276–285, IEEE, 1997.

    Google Scholar 

  8. K. Arnold, J. Gosling, and D. Holmes, The Java Programming Language. Addison-Wesley, third ed., 2000.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2001 Springer-Verlag Wien

About this paper

Cite this paper

Cawley, G.C. (2001). Efficient Sequential Minimal Optimisation of Support Vector Classifiers. In: Kůrková, V., Neruda, R., Kárný, M., Steele, N.C. (eds) Artificial Neural Nets and Genetic Algorithms. Springer, Vienna. https://doi.org/10.1007/978-3-7091-6230-9_107

Download citation

  • DOI: https://doi.org/10.1007/978-3-7091-6230-9_107

  • Publisher Name: Springer, Vienna

  • Print ISBN: 978-3-211-83651-4

  • Online ISBN: 978-3-7091-6230-9

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics