Supervised Group Sparse Representation via Intra-class Low-Rank Constraint

  • Peipei Kang
  • Xiaozhao FangEmail author
  • Wei Zhang
  • Shaohua Teng
  • Lunke Fei
  • Yong Xu
  • Yubao Zheng
Conference paper
Part of the Lecture Notes in Computer Science book series (LNCS, volume 10996)


Group sparse representation (GSR) which uses the group structure of training samples as the prior knowledge has achieved extensive attention. However, GSR represents a new test sample using the original input features for least reconstruction, which may not be able to obtain discriminative reconstruction coefficients since redundant or noisy features may exist in the original input features. To obtain more discriminative data representation, in this paper we propose a novel supervised group sparse representation via intra-class low-rank constraint (GSRILC). Instead of representing the target by the original input features, GSRILC attempts to use the compact projection features in a new subspace for data reconstruction. Concretely, GSRILC projects data sharing the same class to a new subspace, and imposes low-rank constraint on the intra-class projections, which ensures that samples within the same class have a low rank structure. In this way, small intra-class distances and large inter-class distances can be achieved. To illustrate the effectiveness of the proposal, we conduct experiments on the Extended Yale B and CMU PIE databases, and results show the superiority of GSRILC.


Group sparse representation Intra-class low-rank Discriminative data representation Subspace learning 



This work is supported in part by the Natural Science Foundation of China under Grants 61772141, 61702110, 61603100, Guangdong Provincial Natural Science Foundation under Grant 17ZK0422, Guangdong Provincial Science and Technology Project under Grants 2015B090901016, 2016B010108007, and Guangzhou Science and Technology Project under Grants 201804010347.


  1. 1.
    Fang, X.Z., et al.: Approximate low-rank projection learning for feature extraction. IEEE Trans. Neural Netw. Learn. Syst. 99, 1–14 (2018)CrossRefGoogle Scholar
  2. 2.
    Wright, J., Yang, A.Y., Ganesh, A., Sastry, S.S., Ma, Y.: Robust face recognition via sparse representation. IEEE Trans. Pattern Anal. Mach. Intell. 31(2), 210–227 (2009)CrossRefGoogle Scholar
  3. 3.
    Zhang, L., Yang, M., Feng, X.C.: Sparse representation or collaborative representation: which helps face recognition? In: 2011 IEEE International Conference on Computer Vision, pp. 471–478. IEEE (2011)Google Scholar
  4. 4.
    Zhang, Y.M.Z., Jiang, Z.L., Davis, L.S.: Learning structured low-rank representations for image classification. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 676–683. IEEE Computer Society (2013)Google Scholar
  5. 5.
    Tang, K.W., Liu, R.S., Su, Z.X., Zhang, J.: Structure-constrained low-rank representation. IEEE Trans. Neural Netw. Learn. Syst. 25(12), 2167–2179 (2014)CrossRefGoogle Scholar
  6. 6.
    Yuan, M., Lin, Y.: Model selection and estimation in regression with grouped variables. J. Royal Stat. Soc. 68(1), 49–67 (2006)MathSciNetCrossRefGoogle Scholar
  7. 7.
    Majumdar, A., Ward, R.K.: Classification via group sparsity promoting regularization. In: IEEE International Conference on Acoustics, Speech and Signal Processing, pp. 861–864. IEEE (2009)Google Scholar
  8. 8.
    Gao, L.W., Li, Y.Q., Huang, J.Z., Zhou, S.G.: Semi-supervised group sparse representation: model, algorithm and applications. In: European Conference on Artificial Intelligence, pp. 507–514. IOS Press (2016)Google Scholar
  9. 9.
    Li, Z.C., Liu, J., Tang, J.H., Lu, H.Q.: Robust structured subspace learning for data representation. IEEE Trans. Pattern Anal. Mach. Intell. 37(10), 2085–2098 (2015)CrossRefGoogle Scholar
  10. 10.
    Fang, X.Z., Teng, S.H., Lai, Z.H., He, Z.S., Xie, S.L., Wong, W.K.: Robust latent subspace learning for image classification. IEEE Trans. Neural Netw. Learn. Syst. 99, 1–14 (2017)Google Scholar
  11. 11.
    Ren, C.X., Dai, D.Q., Yan, H.: Robust classification using \(l_{2,1}\) -norm based regression model. Pattern Recogn. 45(7), 2708–2718 (2012)CrossRefGoogle Scholar
  12. 12.
    Boyd, S., Parikh, N., Chu, E., Peleato, B., Eckstein, J.: Distributed optimization and statistical learning via the alternating direction method of multipliers. Found. Trends Mach. Learn. 3(1), 1–122 (2010)CrossRefGoogle Scholar
  13. 13.
    Wen, Z.W., Yin, W.T.: A feasible method for optimization with orthogonality constraints. Math. Program. 142(1–2), 397–434 (2013)MathSciNetCrossRefGoogle Scholar
  14. 14.
    Turk, M., Pentland, A.: Eigenfaces for recognition. J. Cogn. Neurosci. 3(1), 71–86 (1991)CrossRefGoogle Scholar
  15. 15.
    Zhou, P., Lin, Z.C., Zhang, C.: Integrated low-rank-based discriminative feature learning for recognition. IEEE Trans. Neural Netw. Learn. Syst. 27(5), 1080–1093 (2016)MathSciNetCrossRefGoogle Scholar

Copyright information

© Springer Nature Switzerland AG 2018

Authors and Affiliations

  • Peipei Kang
    • 1
  • Xiaozhao Fang
    • 1
    Email author
  • Wei Zhang
    • 1
  • Shaohua Teng
    • 1
  • Lunke Fei
    • 1
  • Yong Xu
    • 2
  • Yubao Zheng
    • 3
  1. 1.School of Computer Science and TechnologyGuangdong University of TechnologyGuangzhouChina
  2. 2.Bio-Computing Research Center, Shenzhen Graduate SchoolHarbin Institute of TechnologyShenzhenChina
  3. 3.Department of Infectious DiseasesThe Third Affiliated Hospital of Sun Yat-Sen UniversityGuangzhouChina

Personalised recommendations