Abstract
A Half-Against-Half (HAH) multi-class SVM is proposed in this paper. Unlike the commonly used One-Against-All (OVA) and One-Against-One (OVO) implementation methods, HAH is built via recursively dividing the training dataset of K classes into two subsets of classes. The structure of HAH is same as a decision tree with each node as a binary SVM classifier that tells a testing sample belongs to one group of classes or the other. The trained HAH classifier model consists of at most K binary SVMs. For each classification testing, HAH requires at most K binary SVM evaluations. Both theoretical estimation and experimental results show that HAH has advantages over OVA and OVO based methods in the evaluation speed as well as the size of the classifier model while maintaining comparable accuracy.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Boser, B., Guyon, I., Vapnik, V.: A training algorithm for optimal margin classifiers. In: Haussler, D. (ed.) 5th Annual ACM Workshop on COLT, pp. 144–152 (1992)
Bredensteiner, E., Bennet, K.: Multicategory classification by support vector machines. In: Computational Optimizations and Applications, vol. 12, pp. 53–79 (1999)
Crammer, K., Singer, Y.: On the algorithmic implementation of multiclass kernel-based vector machines. Journal of Machine Learning Research 2, 265–292 (2001)
Vapnik, V.: Statistical Learning Theory. Wiley, New York (1998)
Kreßel, U.: Pairwise classification and support vector machines. In: Advances in Kernel Methods: Support Vector Learnings, pp. 255–268. MIT Press, Cambridge (1999)
Platt, J., Cristianini, N., Shawe-Taylor, J.: Large margin DAGs for multiclass classification. In: Advances in Neural Information Processing Systems, vol. 12, pp. 547–553 (2000)
Hsu, C., Lin, C.: A comparison of methods for multi-class support vector machines. IEEE Transactions on Neural Networks 13, 415–425 (2002)
Rifin, R., Klautau, A.: In defense of one vs-all-classification. Journal of Machine Learning Research 5, 101–141 (2004)
Hettich, S., Bay, S.: The UCI KDD archive (1999)
Liu, X., Xing, H., Wang, X.: A multistage support vector machine. In: The 2nd International Conference on Machine Learning and Cybernetics, pp. 1305–1308 (2003)
Ben-Hur, A., Horn, D., Siegelmann, H., Vapnik, V.: Support vector clustering. Journal of Machine Learning Research 2, 125–137 (2001)
Platt, J.: Fast training of support vector machines using sequential minimal optimization. In: Advances in Kernel Methods - Support Vector Learning, pp. 185–208. MIT Press, Cambridge (1999)
Chang, C., Lin, C.: Libsvm: a library for support vector machines (2001)
Schölkopf, B., Smola, A., Williamson, R., Bartlett, P.: New support vector algorithms. Neural Computation 12, 1207–1245 (2000)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2005 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Lei, H., Govindaraju, V. (2005). Half-Against-Half Multi-class Support Vector Machines. In: Oza, N.C., Polikar, R., Kittler, J., Roli, F. (eds) Multiple Classifier Systems. MCS 2005. Lecture Notes in Computer Science, vol 3541. Springer, Berlin, Heidelberg. https://doi.org/10.1007/11494683_16
Download citation
DOI: https://doi.org/10.1007/11494683_16
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-26306-7
Online ISBN: 978-3-540-31578-0
eBook Packages: Computer ScienceComputer Science (R0)