Abstract
When partitioning the data is the main concern, it is implicitly assumed that each cluster can be approximately regarded as a sample from one component of a mixture model. Thus, the clustering problem can be viewed as an estimation problem of the parameters of the mixture. Setting this problem under the Maximum likelihood and Classification likelihood approaches, we first study the clustering of objects described by categorical attributes using the latent class model and we concentrate our attention on the problem of the number of components. To this end, we use three criteria derived within a Bayesian framework to tackle this problem. These criteria based on approximations of integrated likelihood and of integrated classification likelihood have been recently compared in Gaussian mixture. In this work, we propose to extend these comparisons to the latent class model.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
BANFIELD, J.D. and RAFTERY, A.E. (1993): Model-based Gaussian and non-Gaussian Clustering. Biometrics, 49, 803–821.
BIERNACKI, C. CELEUX, G. and GOVAERT, G. (2000): Assessing a Mixture Model for Clustering with the Integrated Completed Likelihood. IEEE Transactions on Pattern Analysis and Machine Intelligence, 22, 7, 719–725.
CELEUX, G. and GOVAERT, G. (1992): A Classification EM Algorithm for Clustering and two Stochastic Versions. Computational Statistics ê? Data Analysis, 14, 315–332.
CELEUX, G. and GOVAERT, G. (1995): Gaussian Parsimonious Clustering Models. Pattern Recognition, 28, 781–793.
CHEESEMAN, P. and STUTZ, J. (1996): Bayesian Classification (AutoClass): Theory and Results. In Fayyad, U. and Pitesky-Shapiro, G. and Uthurusamy, R. (Eds.): Advances in Knowledge Discovery and Data Mining. AAAI Press, 6183.
CHICKERING, D.M and HECKERMAN, D. (1997): Maximum Approximations for the Marginal Likelihood of Bayesian Networks with Hidden Variables. Machine Learning, 29, 181–212.
DAY, N.E. (1969): Estimating the Components of a Mixture of Normal Distributions. Biometrika, 56, 464–474.
DEMPSTER, A. P., LAIRD, N. M. and RUBIN, D. B. (1997): Maximum Likelihood for Incomplete Data via the EM Algorithm. Journal of the Royal Statistical Society, 39, B, 1–38.
DOMINGOS, P. and PAZZANI, M. (1997): Beyond Independence: Conditions for the Optimality of the Simple Bayesian Classifier. Machine Learning, 29, 103130.
FRALEY, C. and RAFTERY, A.E. (1999): MCLUST: Software for Model-Based Cluster and Discriminant Analysis. Technical Report, 342, University of Washington.
KASS, R. E. and RAFTERY, A. E. (1995): Bayes Factors. Journal of the American Statistical Association, 90, 773–795.
LAZARFELD, P. F. and HENRY, N. W. (1968): Latent Structure Analysis. Houghton Mifflin, Boston.
MCLACHLAN, G. J. and BASFORD, K. E. (1988): Mixture Models: Inference and Applications to Clustering. Marcel Dekker, New York.
MCLACHLAN, G. and PEEL, D. (1998): User’s Guide to EMMIX-Version 1.0. Technical Report, University of Queensland.
SCHWARTZ, G. (1978): Estimating the Dimension of a Model. Annals of Statistics, 6, 461–464.
SYMONS, M. J. (1981): Clustering Criteria and Multivariate Normal Mixture. Biometrics, 27, 387–397.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2002 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Jollois, FX., Nadif, M., Govaert, G. (2002). Assessing the Number of Clusters of the Latent Class Model. In: Jajuga, K., Sokołowski, A., Bock, HH. (eds) Classification, Clustering, and Data Analysis. Studies in Classification, Data Analysis, and Knowledge Organization. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-56181-8_15
Download citation
DOI: https://doi.org/10.1007/978-3-642-56181-8_15
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-43691-1
Online ISBN: 978-3-642-56181-8
eBook Packages: Springer Book Archive