Skip to main content

Online Unsupervised Neural-Gas Learning Method for Infinite Data Streams

  • Conference paper
  • First Online:
Pattern Recognition Applications and Methods

Abstract

We propose an unsupervised online learning method based on the “growing neural gas” algorithm (GNG), for a data-stream configuration where each incoming data is visited only once and used to incrementally update the learned model as soon as it is available. The method maintains a model as a dynamically evolving graph topology of data-representatives that we call neurons. Unlike usual incremental learning methods, it avoids the sensitivity to initialization parameters by using an adaptive parameter-free distance threshold to produce new neurons. Moreover, the proposed method performs a merging process which uses a distance-based probabilistic criterion to eventually merge neurons. This allows the algorithm to preserve a good computational efficiency over infinite time. Experiments on different real datasets, show that the proposed method is competitive with existing algorithms of the same family, while being independent of sensitive parameters and being able to maintain fewer neurons, which makes it convenient for learning from infinite data-streams.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 129.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 169.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    The distance is weighted by the number of data-points associated to the neighbouring neuron.

  2. 2.

    We will refer to AING without the merging process by AING1, and to AING with the merging process by AING2.

References

  1. Fritzke, B.: A growing neural gas network learns topologies. In: Neural Information Processing Systems, pp. 625–632 (1995)

    Google Scholar 

  2. Martinetz, T.: Competitive Hebbian learning rule forms perfectly topology preserving maps. In: Proceedings of the International Conference on Artificial Neural Networks, pp. 427–434. Amsterdam, Netherlands (1993)

    Google Scholar 

  3. Prudent, Y., Ennaji, A.: An incremental growing neural gas learns topologies. In: International Joint Conference on Neural Networks, pp. 1211–1216 (2005)

    Google Scholar 

  4. Hamza, H., Belaid, Y., Belaid, A., Chaudhuri, B.: Incremental classification of invoice documents. In: Proceedings of International Conference on Pattern Recognition, pp. 1–4 (2008)

    Google Scholar 

  5. Shen, F., Ogura, T., Hasegawa, O.: An enhanced self-organizing incremental neural network for online unsupervised learning. Neural Netw. 20(8), 893–903 (2007)

    Article  MATH  Google Scholar 

  6. Keogh, E., Lonardi, S., Ratanamahatana, C.A.: Towards parameter-free data mining. In: Proceedings of the 10th International Conference on Knowledge Discovery and Data Mining, pp. 206–215 (2004)

    Google Scholar 

  7. O’Callaghan, L., Meyerson, A., Motwani, R., Mishra, N., Guha, S.: Streaming-data algorithms for high-quality clustering. In: Proceedings of International Conference on Data Engineering, pp. 685–696. San Francisco (2002)

    Google Scholar 

  8. Aggarwal, C.C., Han, J., Wang, J., Yu, P.S.: A framework for clustering evolving data streams. In: Proceedings of the 29th International Conference on Very Large Data Bases, pp. 81–92. Berlin, Germany (2003)

    Google Scholar 

  9. Chen, Y., Tu, L.: Density-based clustering for real-time stream data. In: Proceedings of International Conference on Knowledge Discovery and Data Mining, pp. 133–142 (2007)

    Google Scholar 

  10. Shindler, M., Wong, A., Meyerson, A.: Fast and accurate k-means for large datasets. In: Neural Information Processing Systems, pp. 2375–2383 (2011)

    Google Scholar 

  11. Frank, A., Asuncion, A.: The UCI machine learning repository. http://archive.ics.uci.edu/ml/. (2010)

  12. Yann, L., Corinna, C.: MNIST handwritten digit database. http://yann.lecun.com/exdb/mnist/. (2010)

  13. Rosenberg, A., Hirschberg, J.: V-measure: a conditional entropy-based external cluster evaluation measure. In: Neural Information Processing Systems, pp. 410–420 (2007)

    Google Scholar 

  14. Bouguelia, M-R., Belaid, Y., Belaid, A.: A stream-based semi-supervised active learning approach for document classification. In: International Conference on Document Analysis and Recognition (2013)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Mohamed-Rafik Bouguelia .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2015 Springer International Publishing Switzerland

About this paper

Cite this paper

Bouguelia, MR., Belaïd, Y., Belaïd, A. (2015). Online Unsupervised Neural-Gas Learning Method for Infinite Data Streams. In: Fred, A., De Marsico, M. (eds) Pattern Recognition Applications and Methods. Advances in Intelligent Systems and Computing, vol 318. Springer, Cham. https://doi.org/10.1007/978-3-319-12610-4_4

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-12610-4_4

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-12609-8

  • Online ISBN: 978-3-319-12610-4

  • eBook Packages: EngineeringEngineering (R0)

Publish with us

Policies and ethics