Abstract
The scarcity of labeled training data relative to the high-dimensionality multi-modal features is one of the major obstacles for semantic concept classification of images and videos. Semi-supervised learning leverages the large amount of unlabeled data in developing effective classifiers. Feature subspace learning finds optimal feature subspaces for representing data and helping classification. In this paper, we present a novel algorithm, Locality Preserving Semi-supervised Support Vector Machines (LPSSVM), to jointly learn an optimal feature subspace as well as a large margin SVM classifier. Over both labeled and unlabeled data, an optimal feature subspace is learned that can maintain the smoothness of local neighborhoods as well as being discriminative for classification. Simultaneously, an SVM classifier is optimized in the learned feature subspace to have large margin. The resulting classifier can be readily used to handle unseen test data. Additionally, we show that the LPSSVM algorithm can be used in a Reproducing Kernel Hilbert Space for nonlinear classification. We extensively evaluate the proposed algorithm over four types of data sets: a toy problem, two UCI data sets, the Caltech 101 data set for image classification, and the challenging Kodak’s consumer video data set for semantic concept detection. Promising results are obtained which clearly confirm the effectiveness of the proposed method.
Chapter PDF
Similar content being viewed by others
Keywords
- Unlabeled Data
- Semantic Concept
- Generalize Eigenvalue Problem
- Locality Preserve Projection
- Feature Subspace
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.
References
Joachims, T.: Transductive inference for text classification using support vector machines. In: ICML, pp. 200–209 (1999)
Chapelle, O., et al.: Semi-supervised learning. MIT Press, Cambridge (2006)
Vapnik, V.: Statistical learning theory. Wiley-Interscience, New York (1998)
Zhu, X.: Semi-supervised learning literature survey. Computer Sciences Technique Report 1530. University of Wisconsin-Madison (2005)
Bengio, Y., Delalleau, O., Roux, N.: Efficient non-parametric function induction in semi-supervised learning. Technique Report 1247, DIRO. Univ. of Montreal (2004)
Belkin, M., Niyogi, P.: Laplacian eigenmaps for dimensionality reduction and data representation. Neural Computation 15, 1373–1396 (2003)
Cai, D., et al.: Spectral regression: a unified subspace learning framework for content-based image retrieval. ACM Multimedia (2007)
Duda, R.O., et al.: Pattern classification, 2nd edn. John Wiley and Sons, Chichester (2001)
He, X., Niyogi, P.: Locality preserving projections. Advances in NIPS (2003)
Belkin, M., Niyogi, P., Sindhwani, V.: Manifold regularization: A geometric framework for learning from labeled and unlabeled examples. Journal of Machine Learning Research 7, 2399–2434 (2006)
Blake, C., Merz, C.: Uci repository of machine learning databases (1998), http://www.ics.uci.edu/~mlearn/MLRepository.html
Li, F., Fergus, R., Perona, P.: Learning generative visual models from few training examples: An incremental bayesian approach tested on 101 object categories. In: CVPR Workshop on Generative-Model Based Vision (2004)
Loui, A., et al.: Kodak’s consumer video benchmark data set: concept definition and annotation. In: ACM Int’l Workshop on Multimedia Information Retrieval (2007)
Schölkopf, B., Smola, A., Müller, K.: Nonlinear component analysis as a kernel eigenvalue problem. Neural Computation 10, 1299–1319 (1998)
Hsu, C., Chang, C., Lin, C.: A practical guide to support vector classification, http://www.csie.ntu.edu.tw/~cjlin/libsvm/
Lazebnik, S., Schmid, C., Ponce, J.: Beyond bags of features: Spatial pyramid matching for recognizing natural scene categories. In: CVPR, vol, 2, pp. 2169–2178
Cai, D., et al.: http://www.cs.uiuc.edu/homes/dengcai2/Data/data.html
Joachims, T.: Training linear svms in linear time. ACM KDD, 217–226 (2006)
Fergus, R., et al.: Object class recognition by unsupervised scale-invariant learning. In: CVPR, pp. 264–271 (2003)
Lowe, D.: Distinctive image features from scale-invariant keypoints. International Journal of Computer Vision 60, 91–110 (2004)
Chang, S., et al.: Large-scale multimodal semantic concept detection for consumer video. In: ACM Int’l Workshop on Multimedia Information Retrieval (2007)
NIST TRECVID (2001 – 2007), http://www-nlpir.nist.gov/projects/trecvid/
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2008 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Jiang, W., Chang, SF., Jebara, T., Loui, A.C. (2008). Semantic Concept Classification by Joint Semi-supervised Learning of Feature Subspaces and Support Vector Machines. In: Forsyth, D., Torr, P., Zisserman, A. (eds) Computer Vision – ECCV 2008. ECCV 2008. Lecture Notes in Computer Science, vol 5305. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-88693-8_20
Download citation
DOI: https://doi.org/10.1007/978-3-540-88693-8_20
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-88692-1
Online ISBN: 978-3-540-88693-8
eBook Packages: Computer ScienceComputer Science (R0)