Abstract
Multitask learning is a learning paradigm which seeks to improve the generalization performance of a task with the help of other tasks. Learning multiple related tasks simultaneously has been empirically as well as theoretically shown to improve performance relative to learning each task independently. In this paper, we propose a new classification method named multitask twin support vector machines based on the regularization principle and twin support vector machines. Our new approach is that we put twin support vector machines to multitask learning. Experimental results demonstrate that the proposed method dramatically improves the performance relative to learning each task independently.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Shawe-Taylor, J., Sun, S.: A Review of Optimization Methodologies in Support Vector Machines. Neurocomputing 74, 3609–3618 (2011)
Mangasarian, O.L., Wild, E.W.: MultisurFace Proximal Support Vector Machine Classification via Generalized Eigenvalues. IEEE Transactions on Pattern Analysis and Machine Intelligence 28, 69–74 (2006)
Jayadeva, Khemchandani, R., Chandra, S.: Twin Support Vector Machines for Pattern Classification. IEEE Transactions on Pattern Analysis and Machine Intelligence 29(5), 905–910 (2007)
Ghorai, S., Mukherjee, A., Dutta, P.K.: Nonparallel Plane Proximal Classifier. Signal Processing, 510–522 (2009)
Caruana, R.: Multitask Learning. Machine Learning 28, 41–75 (1997)
Sun, S.: Multitask Learning for EEG-based Biometrics. In: Proceedings of 19th International Conference on Pattern Recognition, pp. 1–4 (2008)
Baxter, J.: A Model for Inductive Bias Learning. Journal of Artificial Intelligence Research 12, 149–198 (2000)
Evgeniou, T., Micchelli, C.A., Pontil, M.: Learning Multiple Tasks with Kernel Methods. Journal of Machine Learning Research 6, 615–637 (2005)
Ji, Y., Sun, S.: Multitask Multiclass Support Vector Machines. In: Proceedings of IEEE 11th International Conference on Data Mining Workshops, pp. 512–518 (2011)
Ben-David, S., Schuller, R.: Exploiting Task Relatedness for Multiple Task Learning. In: Schölkopf, B., Warmuth, M.K. (eds.) COLT/Kernel 2003. LNCS (LNAI), vol. 2777, pp. 567–580. Springer, Heidelberg (2003)
Ando, R.K., Zhang, T.: A Framework for Learning Predictive Structures from Multiple Tasks and Unlabeled Data. Journal of Machine Learning Research 6, 1817–1853 (2005)
Bakker, B., Heskes, T.: Task Clustering and Gating for Bayesian Multitask Learning. Journal of Machine Learning Research 4, 83–99 (2003)
Evgeniou, T., Pontil, M.: Regularized Multi-Task Learning. In: Proceedings of Knowledge Discovery and Data Mining, pp. 109–117 (2004)
Asuncion, A., Newman, D.J.: UCI Machine Learning Repository, http://archive.ics.uci.edu/ml/datasets.html
Duck Electrical Computer Engineering, http://www.ece.duke.edu/lcarin/LandmineData.zip
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2012 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Xie, X., Sun, S. (2012). Multitask Twin Support Vector Machines. In: Huang, T., Zeng, Z., Li, C., Leung, C.S. (eds) Neural Information Processing. ICONIP 2012. Lecture Notes in Computer Science, vol 7664. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-34481-7_42
Download citation
DOI: https://doi.org/10.1007/978-3-642-34481-7_42
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-34480-0
Online ISBN: 978-3-642-34481-7
eBook Packages: Computer ScienceComputer Science (R0)