Skip to main content

Weighted Mixture of Models For On-Line Learning

  • Chapter
Mathematics of Neural Networks

Part of the book series: Operations Research/Computer Science Interfaces Series ((ORCS,volume 8))

  • 2314 Accesses

Abstract

This paper proposes a weighted mixture of locally generalizing models in an attempt to resolve the trade-off between model mismatch and measurement noise given a sparse set of training samples so that the conditional mean estimation performance of the desired response can be made adequate over the input region of interest. In this architecture, each expert model has its corresponding variance model for estimating the expert’s modeling performance. Parameters associated with individual expert models are adapted in the usual least-mean-square sense, weighted by its variance model output. Whereas the variance models are adapted in such a way that expert models of higher-resolution (or greater modeling capability) are discouraged to contribute, except when the local modeling performance becomes inadequate.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 169.00
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 219.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 219.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. An P.E., Miller W.T., Parks P.C. Design Improvements in Associative Memories for Cerebellar Model Articulation Controllers (CMAC), Proc. Intl. Conf. on Artificial Neural Networks, Helsinki, North Holland, Vol.2 (1991), pp.1207–1210.

    Google Scholar 

  2. Bishop C. Mixture Density Networks, Neural Computing Research Group Report NCRG/4288 (1994), Department of Computer Science, Aston University, UK.

    Google Scholar 

  3. Brown M., Harris C.J. NeuroFuzzy Adaptive Modelling and Control (1994), Prentice-Hall.

    Google Scholar 

  4. Geman S., Bienenstock E. Neural Networks and the Bias/Variance Dilemma, Neural Computation, Vol. 4 (1992), pp.1–58.

    Article  Google Scholar 

  5. Jacobs R., Jordan M., Nowlan S., Hinton G. Adaptive Mixtures of Local Experts, Neural Computation, Vol. 3 (1991), pp.79–87.

    Article  Google Scholar 

  6. Sjoberg J., Ljung L. Overtraining, Regularization, and Searching for Minimum in Neural Networks, the 4th IFAC Symposium on Adaptive Systems in Control and Signal Processing, France, (1992) pp.669–674.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 1997 Springer Science+Business Media New York

About this chapter

Cite this chapter

An, P.E. (1997). Weighted Mixture of Models For On-Line Learning. In: Ellacott, S.W., Mason, J.C., Anderson, I.J. (eds) Mathematics of Neural Networks. Operations Research/Computer Science Interfaces Series, vol 8. Springer, Boston, MA. https://doi.org/10.1007/978-1-4615-6099-9_8

Download citation

  • DOI: https://doi.org/10.1007/978-1-4615-6099-9_8

  • Publisher Name: Springer, Boston, MA

  • Print ISBN: 978-1-4613-7794-8

  • Online ISBN: 978-1-4615-6099-9

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics