A Self-Organising Hybrid Model for Dynamic Text Clustering

  • Chihli Hung
  • Stefan Wermter
Conference paper


A text clustering neural model, traditionally, is assumed to cluster static text information and represent its inner structure on a flat map. However, the quantity of text information is continuously growing and the relationships between them are usually complicated. Therefore, the information is not static and a flat map may be not enough to describe the relationships of input data. In this paper, for a real-world text clustering task we propose a new competitive Self-Organising Map (SOM) model, namely the Dynamic Adaptive Self-Organising Hybrid model (DASH). The features of DASH are a dynamic structure, hierarchical clustering, non-stationary data learning and parameter self-adjustment. All features are data-oriented: DASH adjusts its behaviour not only by modifying its parameters but also by an adaptive structure . We test the performance of our model using the larger new Reuters news corpus based on the criteria of classification accuracy and mean quantization error.


Stop Criterion Best Match Unit Trained Unit Grow Cell Structure Reuter Corpus 
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.


Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.


  1. 1.
    Honkela, T., Kaski, S., Lagus, K., and Kohonen, T. Newsgroup exploration with WEBSOM method and browsing interface. Report A32, Helsinki University of Technology, 1996Google Scholar
  2. 2.
    Kohonen, T. Self-organization and associative memory. Springer-Verlag, Berlin, 1984MATHGoogle Scholar
  3. 3.
    Grossberg, S. Adaptive pattern classification and universal recoding: I. Parallel development and coding of neural feature detectors. Biological Cybernetics, 1976, 23:121–131MathSciNetMATHCrossRefGoogle Scholar
  4. 4.
    Martinetz, T. and Schulten, K. A ‘Neural-Gas’ network learns topologies. Artificial Neural Network, 1991, 1:397–402Google Scholar
  5. 5.
    Fritzke, B. Growing grid-a self-organizing network with constant neighborhood range and adaptation strength. Neural Processing Letters, 1995, 2(5):9–13CrossRefGoogle Scholar
  6. 6.
    Fritzke, B. Growing cell structures-a self-organizing network for unsupervised and supervised learning. Neural Networks, 1994,7(9):1441–1460CrossRefGoogle Scholar
  7. 7.
    Fritzke, B. A growing neural gas network learns topologies. Advances in Neural Information Processing Systems 7, Tesauro, G., Touretzky, D.S. and Leen, T.K. (Eds), MIT Press, Cambridge MA, 1995: 625-632Google Scholar
  8. 8.
    Blackmore, J. and Mükkulainen, R. Incremental grid growing: encoding highdimensional structure into a two-dimensional feature map. Proceedings of the IEEE International Conference on Neural Networks (ICNN’93), 1993Google Scholar
  9. 9.
    Alahakoon, D., Halgamuge, S.K., and Srinivasan, B. Dynamic self-organizing maps with controlled growth for knowledge discovery. IEEE Tractions on Neural Networks, 2000, 11(3):601–614CrossRefGoogle Scholar
  10. 10.
    Hodge, V. and Austin, J. Hierarchical growing cell structures: TreeGCS. Proceedings of the Fourth International Conference on Knowledge-Based Intelligent Engineering Systems, 2000Google Scholar
  11. 11.
    Chen, H., Schutfels, C. and Orwig, R. Internet categorization and search: a selforganizing approach. Journal of Visual Communication and Image Representation, 1996, 7(1):88–102CrossRefGoogle Scholar
  12. 12.
    Rauber, A., Merkl, D. and Dittenbach, M. The growing hierarchical self-organizing maps: exploratory analysis of high-dimensional data. IEEE Transactions on Neural Networks, 2002,13(6):1331-1341Google Scholar
  13. 13.
    Fritzke, B. A self-organizing network that can follow non-stationary distributions. Proceedings of ICANN-97, International Conference on Artificial Neural Networks, Springer, 1997:613–618Google Scholar
  14. 14.
    Lang, R. and Warwick, K. The plastic self organising map. IEEE World Congress on Computational Intelligence, 2002Google Scholar
  15. 15.
    Marsland, S., Shapiro, J. and Nehrnzow, U. A self-organising network that grows when required. Neural Networks, 2002, 15:1041–1058CrossRefGoogle Scholar
  16. 16.
    Kohonen, T. Self-organizing maps. Springer-Verlag, 2001Google Scholar
  17. 17.
    Martinetz, T.M. Competitive Hebbian learning rule forms perfectly topology preserving maps. International Conference on Artificial Neural Networks, ICANN’93, Amsterdam, 1993:427–434Google Scholar
  18. 18.
    Wermter, S. and Hung, C. Selforganizing Classification on the Reuters News Corpus. COLING2002, 19th International Conference on Computational Linguistics, Taipei, Taiwan, 2002:1086–1092Google Scholar
  19. 19.
    Salton, G. Automatic Text Processing: the Transformation, Analysis, and Retrieval of Information by Computer. Addison-Wesley, USA, 1989Google Scholar
  20. 20.
    Miller, G.A. WordNet: a dictionary browser. Proceedings of the First International Conference on Information in Data, 1985Google Scholar
  21. 21.
    Chakrabarti, S. Data mining for hypertext: a tutorial survey. ACM SIGKDD Explorations, 2000, 1(2):1–11CrossRefGoogle Scholar
  22. 22.
    Kohonen, T., Kaski, S., Lagus, K., Salojarvi, J., Honkela, J., Paatero, V. and Saarela, A. Self organization of a massive document collection. IEEE Transactions on Neural Networks, 2000, 11(3):574–585CrossRefGoogle Scholar
  23. 23.
    Salton, G. and Buckley, C. Term-weighting approaches in automatic text retrieval. Information Processing & Management, 1988, 24(5):513–523CrossRefGoogle Scholar

Copyright information

© Springer-Verlag London 2004

Authors and Affiliations

  • Chihli Hung
    • 1
  • Stefan Wermter
    • 1
  1. 1.Centre for Hybrid Intelligent SystemsThe University of SunderlandUK

Personalised recommendations