Abstract
Dimensionality reduction plays an important role in various machine learning tasks. In this paper, we propose a novel method dubbed Sparse Multi-label bILinear Embedding (SMILE) on Stiefel manifolds for supervised dimensionality reduction on multi-label data. Unlike the traditional multi-label dimensionality reduction algorithms that work on the vectorized data, the proposed SMILE directly takes the second-order tensor data as the input, and thus characterizes the spatial structure of the tensor data in an efficient way. Differentiating from the existing tensor-based dimensionality reduction methods that perform the eigen-decomposition in each iteration, SMILE utilizes a gradient ascent strategy to optimize the objective function in each iteration, and thus is more efficient. Moreover, we introduce column-orthonormal constraints to transformation matrices to eliminate the redundancy between the projection directions of the learned subspace and add an \(L_1\)-norm regularization term to the objective function to enhance the interpretability of the learned subspace. Experiments on a standard image dataset validate the effectiveness of the proposed method.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
In Algorithm 1, \(\mathbf {W}_{1}(t)\), \(\mathbf {W}_{2}(t)\), \(\nabla _{\mathbf {W}_1} J(t)\), \(\nabla _{\mathbf {W}_2} J(t)\), and J(t) denote the values of \(\mathbf {W}_{1}\), \(\mathbf {W}_{2}\), \(\nabla _{\mathbf {W}_1} J\), \(\nabla _{\mathbf {W}_2} J\), and J after the t-th iteration, respectively.
References
Boothby, W.M.: An Introduction to Differentiable Manifolds and Riemannian Geometry. Academic Press, New York (2002)
Cunningham, J.P., Ghahramani, Z.: Linear dimensionality reduction: survey, insights, and generalizations. J. Mach. Learn. Res. 16, 2859–2900 (2015)
He, X., Yan, S., Hu, Y., Niyogi, P., Zhang, H.J.: Face recognition using laplacianfaces. IEEE Trans. Pattern Anal. Mach. Intell. 27(3), 328–340 (2005)
Hotelling, H.: Analysis of a complex of statistical variables into principal components. J. Edu. Psychol. 24(417–441), 498–520 (1933)
Huang, J., Nie, F., Huang, H., Ding, C.: Supervised and projected sparse coding for image classification. In: Proceedings of 27th AAAI, pp. 438–444 (2013)
Lee, J.M.: Introduction to Smooth Manifolds. Graduate Texts in Mathematics, 2nd edn. Springer, New York (2012). https://doi.org/10.1007/978-1-4419-9982-5
Liu, Y., Liu, Y., Chan, K.C.C., Hua, K.A.: Hybrid manifold embedding. IEEE Trans. Neural Netw. Learn. Syst. 25(12), 2295–2302 (2014)
Wang, C., et al.: What strikes the strings of your heart? - multi-label dimensionality reduction for music emotion analysis via brain imaging. IEEE Trans. Auton. Mental Develop. 7(3), 176–188 (2015)
Liu, Y., Liu, Y., Zhao, Y., Hua, K.A.: What strikes the strings of your heart? - feature mining for music emotion analysis. IEEE Trans. Affect. Comput. 6(3), 247–260 (2015)
Maron, O., Ratan, A.L.: Multiple-instance learning for natural scene classification. In: Proceedings of 15th ICML, pp. 341–349 (1998)
Panagakis, I., Benetos, E., Kotropoulos, C.: Music genre classification: a multilinear approach. In: Proceedings of 9th ISMIR, pp. 583–588 (2008)
Qi, G.J., Hua, X.S., Rui, Y., Tang, J., Zhang, H.J.: Two-dimensional multilabel active learning with an efficient online adaptation model for image classification. IEEE Trans. Pattern Anal. Mach. Intell. 31(10), 1880–1897 (2009)
Sun, L., Ji, S., Ye, J.: Multi-Label Dimensionality Reduction. Chapman and Hall/CRC Machine Learning & Pattern Recognition, Chapman & Hall/CRC (2013)
Symeonidis, P., Nanopoulos, A., Manolopoulos, Y.: Tag recommendations based on tensor dimensionality reduction. In: Proceedings of 2nd ACM RecSys, pp. 43–50 (2008)
Ueda, N., Saito, K.: Parametric mixture models for multi-labeled text. In: NIPS, vol. 15, pp. 737–744 (2003)
Venna, J., Peltonen, J., Nybo, K., Aidos, H., Kaski, S.: Information retrieval perspective to nonlinear dimensionality reduction for data visualization. J. Mach. Learn. Res. 11, 451–490 (2010)
Wang, H., Ding, C., Huang, H.: Multi-label linear discriminant analysis. In: Daniilidis, K., Maragos, P., Paragios, N. (eds.) ECCV 2010. LNCS, vol. 6316, pp. 126–139. Springer, Heidelberg (2010). https://doi.org/10.1007/978-3-642-15567-3_10
Yu, K., Yu, S., Tresp, V.: Multi-label informed latent semantic indexing. In: Proceedings of 28th ACM SIGIR, pp. 258–265 (2005)
Zeng, J., Liu, Y., Leng, B., Xiong, Z., Cheung, Y.: Dimensionality reduction in multiple ordinal regression. IEEE Trans. Neural Netw. Learn. Syst. 29(9), 4088–4101 (2018). https://doi.org/10.1109/TNNLS.2017.2752003
Zhang, M.L., Zhou, Z.H.: Ml-knn: A lazy learning approach to multi-label learning. Pattern Recogn. 40(7), 2038–2048 (2007)
Zhang, M.L., Zhou, Z.H.: A review on multi-label learning algorithms. IEEE Trans. Knowl. Data Eng. 26(8), 1819–1837 (2014)
Zhang, Y., Zhou, Z.H.: Multilabel dimensionality reduction via dependence maximization. ACM Trans. Knowl. Discov. Data 4(3), 14:1–14:21 (2010)
Zhou, Z.H., Zhang, M.L.: Multi-instance multi-label learning with application to scene classification. In: Schölkopf, B., Platt, J.C., Hoffman, T. (eds.) NIPS 19, pp. 1609–1616 (2007)
Zhu, X., Li, X., Zhang, S.: Block-row sparse multiview multilabel learning for image classification. IEEE Trans. Cybern. 46(2), 450–461 (2016)
Acknowledgment
This work was supported in part by the National Natural Science Foundation of China (NSFC) under Grant 61503317, in part by the General Research Fund (GRF) from the Research Grant Council (RGC) of Hong Kong SAR under Project HKBU12202417, and in part by the SZSTI Grant with the Project Code JCYJ20170307161544087.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2018 Springer Nature Switzerland AG
About this paper
Cite this paper
Liu, Y., Dong, G., Gu, Z. (2018). Sparse Multi-label Bilinear Embedding on Stiefel Manifolds. In: Ceci, M., Japkowicz, N., Liu, J., Papadopoulos, G., Raś, Z. (eds) Foundations of Intelligent Systems. ISMIS 2018. Lecture Notes in Computer Science(), vol 11177. Springer, Cham. https://doi.org/10.1007/978-3-030-01851-1_20
Download citation
DOI: https://doi.org/10.1007/978-3-030-01851-1_20
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-01850-4
Online ISBN: 978-3-030-01851-1
eBook Packages: Computer ScienceComputer Science (R0)