Skip to main content

Gradient based learning methods

  • Chapter
  • First Online:
Adaptive Processing of Sequences and Data Structures (NN 1997)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 1387))

Included in the following conference series:

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Albertini, F., Sontag, E. “For neural networks, function determines form”. Neural Networks. Vol 6, pp 975–990, 1993.

    Article  Google Scholar 

  2. Anderson, B., Moore, J. Optimal Filtering. Prentice Hall, Englewood Cliffs, N.J., 1979.

    MATH  Google Scholar 

  3. Back, A.D., Tsoi, A.C. “FIR and IIR synapses, a new neural network architecture for time series modelling”. Neural Computation. Vol. 3, No. 3, pp 375–385, 1991.

    Google Scholar 

  4. Back, A.D., Tsoi, A.C. “A low sensitivity recurrent neural network”. Neural Computation. To appear.

    Google Scholar 

  5. Baldi, P. “Gradient descent learning algorithm overview: A general dynamical systems perspective”. IEEE Trans on Neural Networks. Vol. 6, No 1, Jan. pp 182–195, 1995.

    Article  Google Scholar 

  6. Bryson, A., Ho, Y.C. Applied Optimal Control. Blaisdall Publishing Co., 1969.

    Google Scholar 

  7. Elman, J. “Finding structure in time”. Cognitive Science. Vol. 14, pp 179–211, 1990.

    Article  Google Scholar 

  8. Frasconi, P., Gori, M., Soda, G. “Local feedback multilayered networks”. Neural Computation. Vol. 4, pp 120–130, 1992.

    Google Scholar 

  9. Gill, P., Murray, W., Wright, M. Practical Optimization. Academic Press. New York, 1981.

    MATH  Google Scholar 

  10. Goodwin, G., Sin, K. S. Adaptive filtering, prediction and control. Prentice Hall, Englewood Cliffs, N.J. 1984.

    MATH  Google Scholar 

  11. Hassibi, B., Stork, D., Wolff, G.J., “Optimal Brain Surgeon and general network pruning”. IEEE International Conference on Neural Networks. San Francisco (Ed. E. H. Ruspini et al), pp 293–299, 1993.

    Google Scholar 

  12. Haykin, S. Neural Networks, A comprehensive foundation. MacMillan College Pub Co. 1994.

    Google Scholar 

  13. Hornik, K. “Approximation capabilities of multilayer feedforward neural networks”. Neural Networks. Vol. 4, pp 251–257, 1990.

    Article  Google Scholar 

  14. Hochreiter, S., Schmidhuber, J. “Long sort-term memory”. Neural Computation. Vol 9, pp 1735–1780, 1997.

    Article  Google Scholar 

  15. Kailath, T. Linear Systems. Prentice Hall, Englewood Cliffs, N.J., 1980.

    MATH  Google Scholar 

  16. Lawrence, S., Giles, L., Back, A., Tsoi, A. C. “The gamma MLP — multiple temporal resolutions, the curse of dimensionality, and gradient descent learning”. Neural Computation To appear.

    Google Scholar 

  17. Le Cun, Y., Denker, J., Solla, S. “Optimal brain damage”. Advances in Neural Information Processing Systems, 2. Morgan Kaufman, pp 598–605, 1990.

    Google Scholar 

  18. Leung, C.S., Wong, K. W., Sum, J., Chan, L.W. “Online training and pruning for RLS algorithms”. Electronics Letters. Vol. 32, pp 2152–2153, 1996.

    Article  Google Scholar 

  19. Müller, M. Efficient Training of Feedforward Neural Networks. PhD thesis, Aarhus University, Dec, 1993.

    Google Scholar 

  20. Nerrand, O., Roussel-Ragot, P., Personnaz, L., Dreyfus, G., Marcos, S. “Neural Networks and nonlinear adaptive filtering: Unifying concepts and new algorithms”. Neural Computation. Vol 5, pp 165–197, 1993.

    Google Scholar 

  21. Principe, J., de Vries, B., Oliveira, P. “The gamma filter — a new class of adaptive IIR filters with restricted feedback”. IEEE Trans Signal Processing. Vol. 41, pp 649–656, 1993.

    Article  MATH  Google Scholar 

  22. Puskorius, G. V., Feldkamp, L. A. “Neurocontrol of nonlinear dynamical systems with Kaiman filter trained recurrent networks”. IEEE Trans Neural Networks. Vol. 5, pp 279–297, 1994.

    Article  Google Scholar 

  23. Robinson, A.J. Dynamic error propagation networks. PhD thesis, University of Cambridge, Cambridge, U.K., 1989.

    Google Scholar 

  24. Scarselli, F. Tsoi, A.C. “Universal approximation using feedforward neural networks: A survey of some existing methods, and some results”. Neural Networks. To appear.

    Google Scholar 

  25. Tsoi, A.C., Back, A.D. “Locally recurrent globally feedforward networks: a critical review of architectures”. IEEE Trans on Neural Networks. Vol. 5, No. 2, pp 229–239, 1994.

    Article  Google Scholar 

  26. Tsoi, A.D., Back, A.D. “Discrete time recurrent neural network architectures: a unifying review”. Neurocomputing. Vol. 15, pp 183–224, 1997.

    Article  MATH  Google Scholar 

  27. Wan, E., Beaufays, F. “Diagrammatic methods for deriving and relating temporal neural network algorithms”. This volume.

    Google Scholar 

  28. Williams, R., Zipser, D. “A learning algorithm for continually running fully recurrent neural networks”. Neural Computation. Vol. 1, pp 270–280, 1989.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

C. Lee Giles Marco Gori

Rights and permissions

Reprints and permissions

Copyright information

© 1998 Springer-Verlag Berlin Heidelberg

About this chapter

Cite this chapter

Tsoi, A.C., Tsoi, A.C. (1998). Gradient based learning methods. In: Giles, C.L., Gori, M. (eds) Adaptive Processing of Sequences and Data Structures. NN 1997. Lecture Notes in Computer Science, vol 1387. Springer, Berlin, Heidelberg. https://doi.org/10.1007/BFb0053994

Download citation

  • DOI: https://doi.org/10.1007/BFb0053994

  • Published:

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-64341-8

  • Online ISBN: 978-3-540-69752-7

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics