Skip to main content

Unsupervised and Semi-supervised Dimensionality Reduction with Self-Organizing Incremental Neural Network and Graph Similarity Constraints

  • Conference paper
  • First Online:

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 9651))

Abstract

The complexity of optimizations in semi-supervised dimensionality reduction methods has limited their usage. In this paper, an unsupervised and semi-supervised nonlinear dimensionality reduction method that aims at lower space complexity is proposed. First, a positive and negative competitive learning strategy is introduced to the single layered Self-Organizing Incremental Neural Network (SOINN) to process partially labeled datasets. Then, we formulate the dimensionality reduction of SOINN weight vectors as a quadratic programming problem with graph similarities calculated from previous step as constraints. Finally, an approximation of distances between newly arrived samples and the SOINN weight vectors is proposed to complete the dimensionality reduction task. Experiments are carried out on two artificial datasets and the NSL-KDD dataset comparing with Isomap, Transductive Support Vector Machine etc. The results show that the proposed method is effective in dimensionality reduction and an efficient alternate transductive learner.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

References

  1. Beyer, O., Cimiano, P.: Online semi-supervised growing neural gas. Int. J. Neural Syst. 22(05), 1250023 (2012)

    Article  Google Scholar 

  2. Cai, X., Wei, J., Wen, G., Yu, Z.: Local and global preserving semisupervised dimensionality reduction based on random subspace for cancer classification. IEEE J. Biomed. Health Inform. 18(2), 500–507 (2014)

    Article  Google Scholar 

  3. Cai, X., Wen, G., Wei, J., Yu, Z.: Relative manifold based semi-supervised dimensionality reduction. Frontiers Comput. Sci. 8(6), 923–932 (2014)

    Article  MathSciNet  Google Scholar 

  4. Edmonds, J., Karp, R.M.: Theoretical improvements in algorithmic efficiency for network flow problems. J. ACM (JACM) 19(2), 248–264 (1972)

    Article  MATH  Google Scholar 

  5. Fritzke, B., et al.: A growing neural gas network learns topologies. Adv. Neural Inf. Process. Syst. 7, 625–632 (1995)

    Google Scholar 

  6. Gomory, R.E., Hu, T.C.: Multi-terminal network flows. J. Soc. Ind. Appl. Math. 9(4), 551–570 (1961)

    Article  MathSciNet  MATH  Google Scholar 

  7. Graepel, T., Herbrich, R., Bollmann-Sdorra, P., Obermayer, K.: Classification on pairwise proximity data. Adv. Neural Inf. Process. Syst. 438–444 (1999)

    Google Scholar 

  8. Kohonen, T.: The self-organizing map. Neurocomputing 21(1), 1–6 (1998)

    Article  MATH  Google Scholar 

  9. Maximo, V.R., Quiles, M.G., Nascimento, M.C.: A consensus-based semi-supervised growing neural gas. In: 2014 International Joint Conference on Neural Networks (IJCNN), pp. 2019–2026. IEEE (2014)

    Google Scholar 

  10. Pan, S.J., Kwok, J.T., Yang, Q.: Transfer learning via dimensionality reduction. In: AAAI, vol. 8, pp. 677–682 (2008)

    Google Scholar 

  11. Pan, S.J., Yang, Q.: A survey on transfer learning. IEEE Trans. Knowl. Data Eng. 22(10), 1345–1359 (2010)

    Article  Google Scholar 

  12. Raina, R., Battle, A., Lee, H., Packer, B., Ng, A.Y.: Self-taught learning: transfer learning from unlabeled data. In: Proceedings of the 24th International Conference on Machine Learning, pp. 759–766. ACM (2007)

    Google Scholar 

  13. Roweis, S.T., Saul, L.K.: Nonlinear dimensionality reduction by locally linear embedding. Science 290(5500), 2323–2326 (2000)

    Article  Google Scholar 

  14. Shaw, B., Jebara, T.: Structure preserving embedding. In: Proceedings of the 26th Annual International Conference on Machine Learning, pp. 937–944. ACM (2009)

    Google Scholar 

  15. Shen, F., Yu, H., Sakurai, K., Hasegawa, O.: An incremental online semi-supervised active learning algorithm based on self-organizing incremental neural network. Neural Comput. Appl. 20(7), 1061–1074 (2011)

    Article  Google Scholar 

  16. Tavallaee, M., Bagheri, E., Lu, W., Ghorbani, A.A.: A detailed analysis of the KDD CUP 99 data set. In: 2009 Proceedings of the Second IEEE Symposium on Computational Intelligence for Security and Defence Applications (2009)

    Google Scholar 

  17. Tenenbaum, J.B., de Silva, V., Langford, J.C.: A global geometric framework for nonlinear dimensionality reduction. Science 290(5500), 2319–2323 (2000)

    Article  Google Scholar 

  18. Torgerson, W.S.: Theory and methods of scaling (1958)

    Google Scholar 

  19. Wang, J.: Maximum variance unfolding. In: Wang, J. (ed.) Geometric Structure of High-Dimensional Data and Dimensionality Reduction, pp. 181–202. Springer, Heidelberg (2011)

    Chapter  Google Scholar 

  20. Weinberger, K.Q., Packer, B.D., Saul, L.K.: Nonlinear dimensionality reduction by semidefinite programming and kernel matrix factorization. In: Proceedings of the Tenth International Workshop on Artificial Intelligence and Statistics, pp. 381–388. Citeseer (2005)

    Google Scholar 

  21. Yang, X., Fu, H., Zha, H., Barlow, J.: Semi-supervised nonlinear dimensionality reduction. In: Proceedings of the 23rd International Conference on Machine Learning, pp. 1065–1072. ACM(2006)

    Google Scholar 

Download references

Acknowledgment

This work was partly supported by National Natural Science Foundations of China (No. 61301148, No. 61272061 and No. 71403299), the fundamental research funds for the central universities of China (No. 531107040263, 531107040276), the Research Funds for the Doctoral Program of Higher Education of China (No. 20120161120019 and No. 20130161110002), Hunan Natural Science Foundation of China (No. 14JJ7023).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Dong Wang .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2016 Springer International Publishing Switzerland

About this paper

Cite this paper

Xiang, Z., Xiao, Z., Huang, Y., Wang, D., Fu, B., Chen, W. (2016). Unsupervised and Semi-supervised Dimensionality Reduction with Self-Organizing Incremental Neural Network and Graph Similarity Constraints. In: Bailey, J., Khan, L., Washio, T., Dobbie, G., Huang, J., Wang, R. (eds) Advances in Knowledge Discovery and Data Mining. PAKDD 2016. Lecture Notes in Computer Science(), vol 9651. Springer, Cham. https://doi.org/10.1007/978-3-319-31753-3_16

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-31753-3_16

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-31752-6

  • Online ISBN: 978-3-319-31753-3

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics