Skip to main content

Mixture Initialization Based on Prior Data Visual Analysis

  • Chapter
  • First Online:
  • 212 Accesses

Part of the book series: Studies in Computational Intelligence ((SCI,volume 757))

Abstract

The initialization is known to be a critical task for running a mixture estimation algorithm. A majority of approaches existing in the literature are related to initialization of the expectation-maximization algorithm widely used in this area. This study focuses on the initialization of the recursive mixture estimation for the case of normal components, where the mentioned methods are not applicable. Its key part is a choice of the initial statistics of normal components. Several initialization techniques based on visual analysis of prior data are discussed. Validation experiments are presented.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   99.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   129.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD   129.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

References

  1. Park, B.-J., Zhang, Y., Lord, D. (2010). Bayesian mixture modeling approach to account for heterogeneity in speed data, Transportation Research Part B: Methodological, vol. 44, 5, pp. 662–673.

    Google Scholar 

  2. Yoshigoe, K., Dai, W., Abramson, M., Jacobs, A. (2015). Overcoming invasion of privacy in smart home environment with synthetic packet injection, In: TRON Symposium (TRONSHOW), Tokyo, Japan, 2014, pp. 1–7.

    Google Scholar 

  3. Yu, J. (2011). Fault detection using principal components-based Gaussian mixture model for semiconductor manufacturing processes, IEEE Transactions on Semiconductor Manufacturing, vol. 24, 3, pp. 432–444.

    Google Scholar 

  4. Kárný, M., Böhm, J., Guy, T. V., Jirsa, L., Nagy, I., Nedoma, P., Tesař, L. (2006). Optimized Bayesian dynamic advising: theory and algorithms, Springer-Verlag London.

    Google Scholar 

  5. Kárný, M., Kadlec, J., Sutanto, E. L. (1998). Quasi-Bayes estimation applied to normal mixture, in: Preprints of the 3rd European IEEE Workshop on Computer-Intensive Methods in Control and Data Processing (eds. J. Rojíček, M. Valečková, M. Kárný, K. Warwick), CMP’98 /3./, Prague, CZ, pp. 77–82.

    Google Scholar 

  6. Gupta, M. R., Chen, Y. (2011). Theory and use of the EM method, in: Foundations and Trends in Signal Processing, vol. 4, 3, pp. 223–296.

    Google Scholar 

  7. Boldea, O., Magnus, J. R. (2009). Maximum likelihood estimation of the multivariate normal mixture model, Journal Of The American Statistical Association, vol. 104, 488, pp. 1539–1549.

    Google Scholar 

  8. Wang, H. X., Luo, B., Zhang, Q. B., Wei, S. (2004). Estimation for the number of components in a mixture model using stepwise split-and-merge EM algorithm, Pattern Recognition Letters, vol. 25, 16, pp. 1799–1809.

    Google Scholar 

  9. McGrory, C. A., Titterington, D. M. (2009). Variational Bayesian analysis for hidden Markov models, Australian & New Zealand Journal of Statistics, vol. 51, pp. 227–244.

    Google Scholar 

  10. Šmídl, V., Quinn, A. (2006). The Variational Bayes method in signal processing, Springer-Verlag Berlin Heidelberg.

    Google Scholar 

  11. Frühwirth-Schnatter, S. (2006). Finite mixture and Markov switching models, Springer-Verlag New York, 2006.

    Google Scholar 

  12. Doucet, A., Andrieu, C. (2001). Iterative algorithms for state estimation of jump Markov linear systems, IEEE Transactions on Signal Processing, vol. 49, 6, pp. 1216–1227.

    Google Scholar 

  13. Chen, R., Liu, J. S. (2000). Mixture Kalman filters, Journal of the Royal Statistical Society: Series B (Statistical Methodology), vol. 62, pp. 493–508.

    Google Scholar 

  14. Aggarwal, C. Ch. (2015). Outlier Analysis. In: Data Mining: The Textbook, Springer International Publishing, pp. 237–263.

    Google Scholar 

  15. Fukunaga, K. (2013). Introduction to Statistical Pattern Recognition, series Computer science and scientific computing, Elsevier Science.

    Google Scholar 

  16. Peterka, V. (1981). Bayesian system identification. In: Trends and Progress in System Identification (ed. P. Eykhoff), Oxford, Pergamon Press, 1981, pp. 239–304.

    Google Scholar 

  17. Nagy, I., Suzdaleva, E., Kárný, M., Mlynářová, T. (2011). Bayesian estimation of dynamic finite mixtures, Int. Journal of Adaptive Control and Signal Processing, vol. 25, 9, pp. 765–787.

    Google Scholar 

  18. Melnykov, V., Melnykov, I. (2012). Initializing the EM algorithm in Gaussian mixture models with an unknown number of components, Computational Statistics & Data Analysis, vol. 56, 6, pp. 1381–1395.

    Google Scholar 

  19. Kwedlo, W. (2013). A new method for random initialization of the EM algorithm for multivariate Gaussian mixture learning, in: Proceedings of the 8th International Conference on Computer Recognition Systems CORES 2013, (eds. R. Burduk, K. Jackowski, M. Kurzynski, M. Wozniak, A. Zolnierek), Springer International Publishing, Heidelberg, pp. 81–90.

    Google Scholar 

  20. Blömer, J., Bujna, K. (2013). Simple methods for initializing the EM algorithm for Gaussian mixture models, CoRR, vol. abs/1312.5946, arxiv.org/abs/1312.5946.

  21. Ning, H., Yuxiao Hu, Y., Huang, T. (2008). Efficient initialization of mixtures of experts for human pose estimation, in: Proceedings of the IEEE International Conference on Image Processing.

    Google Scholar 

  22. Paclik, P., Novovičová, J. (2001). Number of components and initialization in Gaussian mixture model for pattern recognition, in: Proceedings of the 5th International Conference on Artificial Neural Networks and Genetic Algorithms, ICANNGA 2001, Prague, Czech Republic, pp. 406–409.

    Google Scholar 

  23. Kárný, M., Nedoma, P., Khailova, N., Pavelková, L. (2003). Prior information in structure estimation, IEE Proceedings, Control Theory and Applications, vol. 150, 6, pp. 643–653.

    Google Scholar 

  24. Suzdaleva, E., Nagy, I., Mlynářová, T. Expert-based initialization of recursive mixture estimation. In Proceedings of 2016 IEEE 8th International Conference on Intelligent Systems IS’2016, p. 308–315, Sofia, Bulgaria, September 4–6, 2016.

    Google Scholar 

  25. Kerridge, D. (1961). Inaccuracy and inference, Journal of Royal Statistical Society B, vol. 23, pp. 284–294.

    Google Scholar 

Download references

Acknowledgements

The research was supported by project GAČR GA15-03564S.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Evgenia Suzdaleva .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2019 Springer International Publishing AG, part of Springer Nature

About this chapter

Check for updates. Verify currency and authenticity via CrossMark

Cite this chapter

Suzdaleva, E., Nagy, I. (2019). Mixture Initialization Based on Prior Data Visual Analysis. In: Hadjiski, M., Atanassov, K. (eds) Intuitionistic Fuzziness and Other Intelligent Theories and Their Applications. Studies in Computational Intelligence, vol 757. Springer, Cham. https://doi.org/10.1007/978-3-319-78931-6_3

Download citation

Publish with us

Policies and ethics