Skip to main content

Least-Mean-Square Training of Cluster-Weighted Modeling

  • Conference paper
Artificial Neural Networks – ICANN 2007 (ICANN 2007)

Part of the book series: Lecture Notes in Computer Science ((LNTCS,volume 4669))

Included in the following conference series:

Abstract

Aside from the Expectation-Maximization (EM) algorithm, Least-Mean-Square (LMS) is devised to further train the model parameters as a complementary training algorithm for Cluster-Weighted Modeling (CWM). Due to different objective functions of EM and LMS, the training result of LMS can be used to reinitialize CWM’s model parameters which provides an approach to mitigate local minimum problems.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Dempster, A.P., Laird, N.M, Rubin, D.B.: Maximum likelihood from incomplete data via the em algorithm. Journal of the Royal Statistical Society 39, 1–38 (1977)

    MATH  MathSciNet  Google Scholar 

  2. Gershenfeld, N., Schoner, B., Metois, E.: Cluster-weighted modeling for time-series analysis. Nature 397, 329–332 (1999)

    Article  Google Scholar 

  3. Hegger, R., Kantz, H., Schreiber, T.: Practical implementation of nonlinear time series methods. The tisean package. Chaos 9, 413 (1999)

    Article  MATH  Google Scholar 

  4. Jordan, M., Jacobs, R.: Hierarchical mixtures of experts and the em algorithm. Neural Computation 6, 181–214 (1994)

    Article  Google Scholar 

  5. Liou, C.Y., Chang, C.C.: The determination of modal damping ratios and natural frequencies from bispectrum modeling. OCEANS 2, 548–553 (1987)

    Article  Google Scholar 

  6. Liou, C.Y., Musicus, B.R.: Separable cross-entropy approach to power spectrum estimation. IEEE Transactions on Acoustics, Speech and Signal Processing 38, 105–113 (1990)

    Article  Google Scholar 

  7. Liou, C.Y., Musicus, B.R.: Cross entropy approximation of structured covariance matrices. arXiv (August 2006), http://arxiv.org/PS_cache/cs/pdf/0608/0608121v1.pdf

  8. Liou, C.Y., Wu, J.M.: Self-organization using Potts models. Neural Networks 9(4), 671–684 (1996)

    Article  Google Scholar 

  9. Mackey, M.C., Glass, L.: Oscillation and chaos in physiological control systems. Science 197, 716–723 (1997)

    Google Scholar 

  10. Poggio, T., Girosi, F.: Networks for approximation and learning. In: Proceedings of the IEEE, pp. 1481–1497. IEEE Computer Society Press, Los Alamitos (1990)

    Google Scholar 

  11. Prokhorov, D.V., Feldkamp, L.A., Feldkamp, T.M.: A new approach to cluster weighted modeling. In: Procedings of International Joint Conference Neural Networks, vol. 3, pp. 1669–1674 (2001)

    Google Scholar 

  12. Wettschereck, D., Dietterich, T.: Improving the performance of radial basis function networks by learning center locations. Advances in Neural Information Processing Systems 4, 1133–1140 (1992)

    Google Scholar 

  13. Wu, J.M., Chiu, S.J.: Independent component analysis using Potts models. IEEE Transactions on Neural Networks 12, 202–212 (2001)

    Article  Google Scholar 

  14. Wu, J.M., Lin, Z.H., Hsu, P.H.: Function approximation using generalized adalines. IEEE Transactions on Neural Networks 17, 541–558 (2006)

    Article  Google Scholar 

  15. Wu, J.M., Lu, C.Y., Liou, C.Y.: Independent component analysis of correlated neuronal responses in area MT. In: International Conference on Neural Information Processing, ICONIP, pp. 639–642 (2005)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Joaquim Marques de Sá Luís A. Alexandre Włodzisław Duch Danilo Mandic

Rights and permissions

Reprints and permissions

Copyright information

© 2007 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Lin, IC., Liou, CY. (2007). Least-Mean-Square Training of Cluster-Weighted Modeling. In: de Sá, J.M., Alexandre, L.A., Duch, W., Mandic, D. (eds) Artificial Neural Networks – ICANN 2007. ICANN 2007. Lecture Notes in Computer Science, vol 4669. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-74695-9_31

Download citation

  • DOI: https://doi.org/10.1007/978-3-540-74695-9_31

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-74693-5

  • Online ISBN: 978-3-540-74695-9

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics