Skip to main content

Incremental ELMVIS for Unsupervised Learning

  • Conference paper
  • First Online:
Proceedings of ELM-2016

Abstract

An incremental version of the ELMVIS+ method is proposed in this paper. It iteratively selects a few best fitting data samples from a large pool, and adds them to the model. The method keeps high speed of ELMVIS+ while allowing for much larger possible sample pools due to lower memory requirements. The extension is useful for reaching a better local optimum with greedy optimization of ELMVIS, and the data structure can be specified in semi-supervised optimization. The major new application of incremental ELMVIS is not to visualization, but to a general dataset processing. The method is capable of learning dependencies from non-organized unsupervised data—either reconstructing a shuffled dataset, or learning dependencies in complex high-dimensional space. The results are interesting and promising, although there is space for improvements.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 129.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 169.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 169.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Cambria, E., et al.: Extreme learning machines [Trends & Controversies]. IEEE Intell. Syst. 28(6) 30–59 (2013)

    Google Scholar 

  2. Huang, G.B., Zhu, Q.Y., Siew, C.K.: Extreme learning machine: theory and applications. In: Neural Networks Selected Papers from the 7th Brazilian Symposium on Neural Networks (SBRN ’04)7th Brazilian Symposium on Neural Networks, vol. 70(1–3), pp. 489–501 (2006)

    Google Scholar 

  3. Huang, G.B., Zhou, H., Ding, X., Zhang, R.: Extreme learning machine for regression and multiclass classification. IEEE Trans. Syst. Man Cybern. Part B: Cybern. 42(2), 513–529 (2012). April

    Article  Google Scholar 

  4. Huang, G.B.: What are extreme learning machines? Filling the gap between Frank Rosenblatt’s dream and John von Neumann’s puzzle. Cogn. Comput. 7(3), 263–278 (2015)

    Article  Google Scholar 

  5. Huang, G.B., Chen, L., Siew, C.K.: Universal approximation using incremental constructive feedforward networks with random hidden nodes. IEEE Trans. Neural Netw. 17(4), 879–892 (2006). July

    Article  Google Scholar 

  6. Miche, Y., Sorjamaa, A., Bas, P., Simula, O., Jutten, C., Lendasse, A.: OP-ELM: optimally-pruned extreme learning machine. IEEE Trans. Neural Netw. 21(1), 158–162 (2010). January

    Article  Google Scholar 

  7. Gutin, G., Punnen, A.P. (eds.): The Traveling Salesman Problem and its Variations. Combinatorial Optimization. Kluwer Academic, Dordrecht, London (2002)

    MATH  Google Scholar 

  8. Alpaydin, E.: Introduction to Machine Learning. MIT press (2014)

    Google Scholar 

  9. Akusok, A., Miche, Y., Björk, K.M., Nian, R., Lauren, P., Lendasse, A.: ELMVIS+: improved nonlinear visualization technique using cosine distance and extreme learning machines. In: Cao, J., Mao, K., Wu, J., Lendasse, A. (eds.) Proceedings of ELM-2015 Volume 2: Theory, Algorithms and Applications (II), pp. 357–369. Springer, Cham (2016)

    Google Scholar 

  10. Gritsenko, A., Akusok, A., Miche, Y., Björk, K.M., Baek, Stephen, Lendasse, A.: Combined Nonlinear Visualization and Classification: ELMVIS++C. Vancouver, Canada (2016)

    Google Scholar 

  11. Haykin, S.: Neural Networks: A Comprehensive Foundation, 2nd edn. Prentice Hall (1998)

    Google Scholar 

  12. Huang, G.B.: An insight into extreme learning machines: random neurons, random features and kernels. Cogn. Comput. 6(3), 376–390 (2014)

    Article  Google Scholar 

  13. Akusok, A., Björk, K.M., Miche, Y., Lendasse, A.: High-performance extreme learning machines: a complete toolbox for big data applications. IEEE Access 3, 1011–1025 (2015). July

    Article  Google Scholar 

  14. Rao, C.R., Mitra, S.K.: Generalized inverse of a matrix and its applications. In: Proceedings of the Sixth Berkeley Symposium on Mathematical Statistics and Probability, Volume 1: Theory of Statistics, pp. 601–620. University of California Press, Berkeley, CA (1972)

    Google Scholar 

  15. Akusok, A., Miche, Y., Björk, K.M., Nian, R., Lauren, P., Lendasse, A.: ELMVIS+: Fast nonlinear visualization technique based on cosine distance and extreme learning machines. Neurocomputing (2016)

    Google Scholar 

  16. Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition. Proc. IEEE 86(11), 2278–2324 (1998). November

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Anton Akusok .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2018 Springer International Publishing AG

About this paper

Cite this paper

Akusok, A. et al. (2018). Incremental ELMVIS for Unsupervised Learning. In: Cao, J., Cambria, E., Lendasse, A., Miche, Y., Vong, C. (eds) Proceedings of ELM-2016. Proceedings in Adaptation, Learning and Optimization, vol 9. Springer, Cham. https://doi.org/10.1007/978-3-319-57421-9_15

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-57421-9_15

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-57420-2

  • Online ISBN: 978-3-319-57421-9

  • eBook Packages: EngineeringEngineering (R0)

Publish with us

Policies and ethics