Advertisement

A Comparative Study of Several Cluster Number Selection Criteria

  • Xuelei Hu
  • Lei Xu
Part of the Lecture Notes in Computer Science book series (LNCS, volume 2690)

Abstract

The selection of the number of clusters is an important and challenging issue in cluster analysis. In this paper we perform an experimental comparison of several criteria for determining the number of clusters based on Gaussian mixture model. The criteria that we consider include Akaike’s information criterion (AIC), the consistent Akaike’s information criterion (CAIC), the minimum description length (MDL) criterion which formally coincides with the Bayesian inference criterion (BIC), and two model selection methods driven from Bayesian Ying-Yang (BYY) harmony learning: harmony empirical learning criterion (BYY-HEC) and harmony data smoothing criterion (BYY-HDS). We investigate these methods on synthetic data sets of different sample size and the iris data set. The results of experiments illustrate that BYY-HDS has the best overall success rate and obviously outperforms other methods for small sample size. CAIC and MDL tend to underestimate the number of clusters, while AIC and BYY-HEC tend to overestimate the number of clusters especially in the case of small sample size.

Keywords

Gaussian Mixture Model Minimum Description Length Model Selection Criterion Mixture Parameter Model Selection Method 
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. 1.
    Jain, A.K., Duin, R.P.W., Mao, J.: Statistical pattern recognition: A review. IEEE Transactions on Pattern Analysis and Machine Intelligence 22, 4–37 (2000)CrossRefGoogle Scholar
  2. 2.
    Bozdogan, H.: Mixture-model cluster analysis using model selection criteria and a new informational measure of complexity. In: Bozdogan, H. (ed.) Proceedings of the First US/Japan Conference on the Frontiers of Statistical Modeling: An Informational Approach, Dordrecht, the Netherlands, vol. 2, pp. 69–113. Kluwer Academic Publishers, Dordrecht (1994)Google Scholar
  3. 3.
    Xu, L.: Byy harmony learning, structural rpcl, and topological self-organizing on mixture models. Neural Networks 15, 1125–1151 (2002)CrossRefGoogle Scholar
  4. 4.
    Akaik, H.: A new look at statistical model identification. IEEE Transactions on Automatic Control 19, 716–723 (1974)CrossRefGoogle Scholar
  5. 5.
    Bozdogan, H.: Model selection and akaike’s information criterion (aic): the general theory and its analytical extensions. Psychometrika 52, 345–370 (1987)zbMATHCrossRefMathSciNetGoogle Scholar
  6. 6.
    Rissanen, J.: Modeling by shortest data description. Automatica 14, 465–471 (1978)zbMATHCrossRefGoogle Scholar
  7. 7.
    Barron, A., Rissanen, J.: The minimum description length principle in coding and modeling. IEEE Trans. Information Theory 44, 2743–2760 (1998)zbMATHCrossRefMathSciNetGoogle Scholar
  8. 8.
    Schwarz, G.: Estimating the dimension of a model. The Annals of Statistics 6, 461–464 (1978)zbMATHCrossRefMathSciNetGoogle Scholar
  9. 9.
    Dempster, A., Laird, N.: Rubin: Maximum likelihood estimation from incomplete data via the em algorithm. J. Royal Statistical Soc. B 39, 1–38 (1977)zbMATHMathSciNetGoogle Scholar
  10. 10.
    Xu, L.: Bayesian ying-yang machine, clustering and number of clusters. Pattern Recognition Letters 18, 1167–1178 (1997)CrossRefGoogle Scholar
  11. 11.
    Xu, L.: Bayesian ying-yang system and theory as a unified statistical learning approach (i) unsupervised and semi-unsupervised learning. In: Amari, S., Kassabov, N. (eds.) Brain-like Computing and Learning, pp. 241–274. Springer, Heidelberg (1997)Google Scholar
  12. 12.
    Xu, L.: Byy harmony learning, independent state space, and generalized apt financial analyses. IEEE Tansactions on Neural Networks 12, 822–849 (2001)CrossRefGoogle Scholar
  13. 13.
    Xu, L.: Data smoothing regularization, multi-sets-learning, and problem solving stategies. Neural Networks (2003) (in press)Google Scholar
  14. 14.
    Sclove, S.L.: Some aspects of model-selection criteria. In: Bozdogan, H. (ed.) Proceedings of the First US/Japan Conference on the Frontiers of Statistical Modeling: An Informational Approach, Dordrecht, the Netherlands, vol. 2, pp. 37–67. Kluwer Academic Publishers, Dordrecht (1994)Google Scholar

Copyright information

© Springer-Verlag Berlin Heidelberg 2003

Authors and Affiliations

  • Xuelei Hu
    • 1
  • Lei Xu
    • 1
  1. 1.Department of Computer Science and EngineeringThe Chinese University of Hong KongShatin, NT, Hong Kong

Personalised recommendations