Skip to main content

Combination of Dynamic Reservoir and Feedforward Neural Network for Time Series Forecasting

  • Conference paper
  • 1582 Accesses

Part of the book series: Lecture Notes in Computer Science ((LNTCS,volume 5507))

Abstract

Echo State neural networks are a special case of recurrent neural networks. The most important part of Echo State neural networks is so called ”dynamic reservoir”. Echo State neural networks use dynamics of this massive and randomly initialized dynamic reservoir to extract interesting properties of incoming sequences. A standard training of these neural networks uses pseudo inverse matrix for one-step learning of weights from hidden to output neurons. In this approach, we have merged this dynamic reservoir with standard feedforward neural network, with a goal to achieve greater prediction ability. This approach was tested in laser fluctuations and Mackey-Glass time series prediction. The prediction error achieved by this approach was substantially smaller in comparison with prediction error achieved by standard algorithm or time delay neural network with backpropagation algorithm.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   129.00
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   169.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Haykin, S.: Neural networks - A comprehensive foundation. Macmillan Publishing, New York (1994)

    MATH  Google Scholar 

  2. Jaeger, H.: The Echo State Approach to Analysing and Training Recurrent Neural Networks. German National Research Center for Information Technology, GMD report 148 (2001)

    Google Scholar 

  3. Jaeger, H.: Short Term Memory in Echo State Networks. German National Research Center for Information Technology, GMD report 152 (2002)

    Google Scholar 

  4. Natschlager, T., Maass, W., Markram, H.: The “liquid computer”: A novel strategy for real-time computing on time series. Special Issue on Foundations of Information Processing of TELEMATIK 8(1), 39–43 (2002)

    MATH  Google Scholar 

  5. Babinec, Š., Pospíchal, J.: Improving the Prediction Accuracy of Echo State Neural Networks by Anti-Oja’s Learning. In: de Sá, J.M., Alexandre, L.A., Duch, W., Mandic, D.P. (eds.) ICANN 2007. LNCS, vol. 4668, pp. 19–28. Springer, Heidelberg (2007)

    Chapter  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2009 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Babinec, Š., Pospíchal, J. (2009). Combination of Dynamic Reservoir and Feedforward Neural Network for Time Series Forecasting. In: Köppen, M., Kasabov, N., Coghill, G. (eds) Advances in Neuro-Information Processing. ICONIP 2008. Lecture Notes in Computer Science, vol 5507. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-03040-6_35

Download citation

  • DOI: https://doi.org/10.1007/978-3-642-03040-6_35

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-642-03039-0

  • Online ISBN: 978-3-642-03040-6

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics