Skip to main content

Entropy and Learning

  • Chapter
  • 122 Accesses

Part of the book series: NATO ASI Series ((NSSB,volume 260))

Abstract

Recently, neural networks have attracted a lot of attention from the scientific community, as is witnessed by the large number of conferences on the subject, and the introduction of a dedicated journal (neural networks). It has become a truely interdisciplinary field, involving artificial intelligence, neurophysiology, psychology and statistical physics. One of the reasons for this success is the relative high performance of neural networks in a number of search and learning problems (e.g. the construction of associative memories[[1]], optimization[[2]], and classification[[3]]). These tasks are performed rather easily by the human brain, despite its relatively low processing speed (the time for a neuron to fire is in the range of msec, while it takes a fraction of a second to recognize a face). The widely accepted explanation for this fact is the high level of parallel processing in the brain, each neuron being connected to thousands of other neurons. This high connectivity is also the characteristic feature of what are called neural networks.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   129.00
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   169.00
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD   169.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. T. Kohonen, Self-Organization and Associative Memory (Springer-Verlag, New York, 1984).

    Google Scholar 

  2. D. E. Rumelhart and J.L.Mc. Clelland, Parallel Distributed Processing (M.I.T. Press, Cambridge, MA, 1986).

    Google Scholar 

  3. E. Aarts and J. Korst, Simulated Annealing and Boltmann Machines (J. Wiley, New York, 1989).

    Google Scholar 

  4. T. J. Sejnowski and Ch. Rosenberg, Complex Systems 1, 145 (1987).

    Google Scholar 

  5. A. Lapedes and R. Farber, in the Proceedings of IEEE Denver Conference on Neural Nets.

    Google Scholar 

  6. S. Patarnello and P. Carnevali, Europhys. Lett. 4, 503 (1987), Europhys. Lett. 4, 1199 (1987).

    Google Scholar 

  7. more details will be given in the following paper: C. Van den Broeck and R. Kawai, Learning in Feedforward Boolean Networks, to be published.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 1991 Springer Science+Business Media New York

About this chapter

Cite this chapter

Van den Broeck, C. (1991). Entropy and Learning. In: Babloyantz, A. (eds) Self-Organization, Emerging Properties, and Learning. NATO ASI Series, vol 260. Springer, Boston, MA. https://doi.org/10.1007/978-1-4615-3778-6_16

Download citation

  • DOI: https://doi.org/10.1007/978-1-4615-3778-6_16

  • Publisher Name: Springer, Boston, MA

  • Print ISBN: 978-1-4613-6684-3

  • Online ISBN: 978-1-4615-3778-6

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics