Skip to main content

Learning Algorithms — theory and practice

  • Conference paper
Neural Network Applications

Part of the book series: Perspectives in Neural Computing ((PERSPECT.NEURAL))

  • 196 Accesses

Abstract

Computational Learning Theory is a recently-developed branch of mathematics which provides a framework for the discussion of experiments with learning machines, such as artificial neural networks. The basic ideas of the theory are described, and applied to an experiment involving the comparison of two learning machines. The experiment was devised so that the results could also be compared with those achieved by a human subject, and this comparison raises some interesting questions.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. M. Anthony and N.L. Biggs, Computational Learning Theory: an Introduction, Cambridge University Press, 1992.

    MATH  Google Scholar 

  2. P.L. Bartlett and R.C. Williamson, Investigating the distribution assumptions in the pac model, Proceedings of the 1991 Workshop on Computational Learning Theory, Morgan Kauffman, San Mateo, 1991.

    Google Scholar 

  3. E. B. Baum, The perceptron algorithm is fast for non-malicious distributions, Neural Computation, 2 (1990) 249–261.

    Google Scholar 

  4. A. Blumer, A. Ehrenfeucht, D. Haussler and M. Warmuth, Learnability and the Vapnik-Chervonenkis dimension, Journal of the ACM, 36 (1989) 929–965.

    Article  MATH  MathSciNet  Google Scholar 

  5. R. L. Gregory and J. G. Wallace, Recovery from early blindness: a case study, Exp. Psychol. Soc. Monographs No. 2, Cambridge 1963.

    Google Scholar 

  6. N. Littlestone, Learning quickly when irrelevant attributes abound, Machine Learning, 2 (1988) 245–318.

    Google Scholar 

  7. M. Minsky and S. Papert, Perceptrons, MIT Press, expanded edition 1988.

    MATH  Google Scholar 

  8. R. Rivest, Learning decision lists, Machine Learning, 2 (1987) 229–246.

    MathSciNet  Google Scholar 

  9. F. Rosenblatt, Principles of Neurodynamics, Spartan Books, 1962.

    Google Scholar 

  10. L. G. Valiant, A view of computational learning theory, NEC Research Symposium: Computation and Cognition (ed. C. W. Gear), SIAM, 1991.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 1992 Springer-Verlag London Limited

About this paper

Cite this paper

Biggs, N.L. (1992). Learning Algorithms — theory and practice. In: Taylor, J.G. (eds) Neural Network Applications. Perspectives in Neural Computing. Springer, London. https://doi.org/10.1007/978-1-4471-2003-2_1

Download citation

  • DOI: https://doi.org/10.1007/978-1-4471-2003-2_1

  • Publisher Name: Springer, London

  • Print ISBN: 978-3-540-19772-0

  • Online ISBN: 978-1-4471-2003-2

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics