Skip to main content

A Comparative Study of Several Cluster Number Selection Criteria

  • Conference paper
Intelligent Data Engineering and Automated Learning (IDEAL 2003)

Part of the book series: Lecture Notes in Computer Science ((LNCS,volume 2690))

Abstract

The selection of the number of clusters is an important and challenging issue in cluster analysis. In this paper we perform an experimental comparison of several criteria for determining the number of clusters based on Gaussian mixture model. The criteria that we consider include Akaike’s information criterion (AIC), the consistent Akaike’s information criterion (CAIC), the minimum description length (MDL) criterion which formally coincides with the Bayesian inference criterion (BIC), and two model selection methods driven from Bayesian Ying-Yang (BYY) harmony learning: harmony empirical learning criterion (BYY-HEC) and harmony data smoothing criterion (BYY-HDS). We investigate these methods on synthetic data sets of different sample size and the iris data set. The results of experiments illustrate that BYY-HDS has the best overall success rate and obviously outperforms other methods for small sample size. CAIC and MDL tend to underestimate the number of clusters, while AIC and BYY-HEC tend to overestimate the number of clusters especially in the case of small sample size.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Jain, A.K., Duin, R.P.W., Mao, J.: Statistical pattern recognition: A review. IEEE Transactions on Pattern Analysis and Machine Intelligence 22, 4–37 (2000)

    Article  Google Scholar 

  2. Bozdogan, H.: Mixture-model cluster analysis using model selection criteria and a new informational measure of complexity. In: Bozdogan, H. (ed.) Proceedings of the First US/Japan Conference on the Frontiers of Statistical Modeling: An Informational Approach, Dordrecht, the Netherlands, vol. 2, pp. 69–113. Kluwer Academic Publishers, Dordrecht (1994)

    Google Scholar 

  3. Xu, L.: Byy harmony learning, structural rpcl, and topological self-organizing on mixture models. Neural Networks 15, 1125–1151 (2002)

    Article  Google Scholar 

  4. Akaik, H.: A new look at statistical model identification. IEEE Transactions on Automatic Control 19, 716–723 (1974)

    Article  Google Scholar 

  5. Bozdogan, H.: Model selection and akaike’s information criterion (aic): the general theory and its analytical extensions. Psychometrika 52, 345–370 (1987)

    Article  MATH  MathSciNet  Google Scholar 

  6. Rissanen, J.: Modeling by shortest data description. Automatica 14, 465–471 (1978)

    Article  MATH  Google Scholar 

  7. Barron, A., Rissanen, J.: The minimum description length principle in coding and modeling. IEEE Trans. Information Theory 44, 2743–2760 (1998)

    Article  MATH  MathSciNet  Google Scholar 

  8. Schwarz, G.: Estimating the dimension of a model. The Annals of Statistics 6, 461–464 (1978)

    Article  MATH  MathSciNet  Google Scholar 

  9. Dempster, A., Laird, N.: Rubin: Maximum likelihood estimation from incomplete data via the em algorithm. J. Royal Statistical Soc. B 39, 1–38 (1977)

    MATH  MathSciNet  Google Scholar 

  10. Xu, L.: Bayesian ying-yang machine, clustering and number of clusters. Pattern Recognition Letters 18, 1167–1178 (1997)

    Article  Google Scholar 

  11. Xu, L.: Bayesian ying-yang system and theory as a unified statistical learning approach (i) unsupervised and semi-unsupervised learning. In: Amari, S., Kassabov, N. (eds.) Brain-like Computing and Learning, pp. 241–274. Springer, Heidelberg (1997)

    Google Scholar 

  12. Xu, L.: Byy harmony learning, independent state space, and generalized apt financial analyses. IEEE Tansactions on Neural Networks 12, 822–849 (2001)

    Article  Google Scholar 

  13. Xu, L.: Data smoothing regularization, multi-sets-learning, and problem solving stategies. Neural Networks (2003) (in press)

    Google Scholar 

  14. Sclove, S.L.: Some aspects of model-selection criteria. In: Bozdogan, H. (ed.) Proceedings of the First US/Japan Conference on the Frontiers of Statistical Modeling: An Informational Approach, Dordrecht, the Netherlands, vol. 2, pp. 37–67. Kluwer Academic Publishers, Dordrecht (1994)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2003 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Hu, X., Xu, L. (2003). A Comparative Study of Several Cluster Number Selection Criteria. In: Liu, J., Cheung, Ym., Yin, H. (eds) Intelligent Data Engineering and Automated Learning. IDEAL 2003. Lecture Notes in Computer Science, vol 2690. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-45080-1_27

Download citation

  • DOI: https://doi.org/10.1007/978-3-540-45080-1_27

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-40550-4

  • Online ISBN: 978-3-540-45080-1

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics