Abstract
This work deals with the classification problem in the case that groups are known and both labeled and unlabeled data are available. The classification rule is derived using Gaussian mixtures where covariance matrices are given according to a multiple testing procedure which asesses a pattern among heteroscedasticity, homometroscedasticity, homotroposcedasticity, and homoscedasticity. The mixture models are then fitted using all available data (labeled and unlabeled) and adopting the EM and the CEM algorithms. The performance of the proposed procedure is evaluated by a simulation study.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsReferences
Bensmail, H., & Celeux, G. (1996). Regularized Gaussian discriminant analysis through eigenvalue decomposition. Journal of the American statistical Association, 91(436), 1743–1748.
Campbell, N. A., & Mahon, R. J. (1974). A multivariate study of variation in two species of rock crab of genus Leptograpsus. Australian Journal of Zoology, 22(3), 417–425.
Celeux, G., & Govaert, G. (1992). A classification em algorithm for clustering and two stochastic versions. Computational Statistics & Data Analysis, 14(3), 315–332.
Dean, N., Murphy, T., & Downey, G. (2006). Using unlabelled data to update classification rules with applications in food authenticity studies. Journal of the Royal Statistical Society: Series C (Applied Statistics), 55(1), 1–14.
Dempster, A., Laird, N., & Rubin, D. (1977). Maximum likelihood from incomplete data via the em algorithm. Journal of the Royal Statistical Society, Series B (Methodological), 1–38.
Flury, B. N. (1984). Common principal components in k groups. Journal of the American Statistical Association, 79(388), 892–898.
Flury, B. N. (1986). Proportionality of k covariance matrices. Statistics & Probability Letters, 4, 29–33.
Flury, B. N., & Gautschi, W. (1986). An algorithm for simultaneous orthogonal transformation of several positive definite matrices to nearly diagonal form. SIAM Journal on Scientific and Statistical Computing, 7, 169–184.
Flury, B. N., & Riedwyl, H. (1983). Angewandte multivariate statistik. Gustav Fischer, Stuttgart, 112–124.
Friedman, J. (1989). Regularized discriminant analysis. Journal of the American Statistical Association, 84(405), 165–175.
Greselin, F., Ingrassia, S., & Punzo, A. (2011). Assessing the pattern of covariance matrices via an augmentation multiple testing procedure. Statistical Methods & Applications, 20(2), 141–170.
Hawkins, D. M. (1981). A new test for multivariate normality and homoscedasticity. Technometrics, 23(1), 105–110.
Manly, R. F. J., & Rayner, J. C. W. (1987). The comparison of sample covariance matrices using likelihood ratio tests. Biometrika, 74(4), 841–847.
McLachlan, G. (1992). Discriminant analysis and statistical pattern recognition (2nd printing). Hoboken, NJ: Wiley.
Peel, D., & McLachlan, G. J. (2000). Robust mixture modelling using the t distribution. Statistics & Computing, 10(4), 339–348.
Ripley, B. (1996). Pattern recognition and neural network. Cambridge: Cambridge University Press.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2013 Springer International Publishing Switzerland
About this paper
Cite this paper
Bagnato, L. (2013). Model-Based Classification Via Patterned Covariance Analysis. In: Giudici, P., Ingrassia, S., Vichi, M. (eds) Statistical Models for Data Analysis. Studies in Classification, Data Analysis, and Knowledge Organization. Springer, Heidelberg. https://doi.org/10.1007/978-3-319-00032-9_3
Download citation
DOI: https://doi.org/10.1007/978-3-319-00032-9_3
Published:
Publisher Name: Springer, Heidelberg
Print ISBN: 978-3-319-00031-2
Online ISBN: 978-3-319-00032-9
eBook Packages: Mathematics and StatisticsMathematics and Statistics (R0)