Abstract
Proposed new neural net architectures and theoretical approaches reflect the multidisciplinary nature of the research community. The very diverse tools used for carrying out theoretical investigations in neural nets go far beyond those discussed in this book, and cover a very wide range of subjects. A brief review of examples of some of the research in the different areas may be of benefit to the reader interested in pursuing the theme. Predicate logic has played a significant part, especially in establishing the limitations of simpler neural nets in the famous book of Minsky and Papert [1]. Extensions of the capabilities of more complex neural nets have involved work in functional analysis of the mappings generated by multilayered non-linear networks. An example of work of this type in function approximation may be seen in Poggio [2]. The nonlinear dynamics of many physical systems has given birth to a number of optimisation techniques — simulated annealing for example [3] — and provides a basis for the study of dynamical response nets with feedback. Many nets are based on principles known from probability and statistics, and the performance of neural net classifiers in particular is related to that of optimal statistical classifiers [4]. Much of the original interest in neural nets was generated by new knowledge of the animal brain acquired by neuroscientists, and the reverse engineering approach to neural net design is based firmly in biology.
Keywords
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Minsky M. L. and Papert S. A.: ‘Perceptrons’ MIT Press, Cambridge, Ma (1969).
Poggio T. and Girosi F.: ‘Networks for approximation and learning’ Proc IEEE 78, No 9, pp 1481–1497 (Sept 1990).
Hinton G. E. and Sejnowski T. J.: ‘Learning and relearning in Boltzmann Machines’ in ‘Parallel Distributed Processing’, Rumelhart D. E. and McClelland J. L., eds, MIT Press, Cambridge, Mass (1986).
Wan E. A.: ‘Neural Net Classification: A Bayesian interpretation’, IEEE Trans on Neural Nets, 1, No 4, pp 303–305 (Dec 1990).
Anderson J. A. and Rosenfield E.: ‘Neurocomputing: foundations of research’, MIT Press, Cambridge, Ma (1988).
Rumelhart D. E. and McClelland J. L.: ‘Parallel Distributed Processing’ MIT Press, Cambridge, Ma (1986).
Grossberg S.: ‘Competitive learning: from interactive activation to adaptive resonance’, Cognitive Science 11, pp 22–23 (1987).
Aleksander I., Thomas W. V. and Bowden P. A.: ‘Wisard: a radical step forward in image recognition’, Sensor Review, pp 120–124 (July 1984).
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 1992 British Telecommunications plc
About this chapter
Cite this chapter
Nightingale, C. (1992). Architectures: An Introduction. In: Linggard, R., Myers, D.J., Nightingale, C. (eds) Neural Networks for Vision, Speech and Natural Language. BT Telecommunications Series, vol 1. Springer, Dordrecht. https://doi.org/10.1007/978-94-011-2360-0_22
Download citation
DOI: https://doi.org/10.1007/978-94-011-2360-0_22
Publisher Name: Springer, Dordrecht
Print ISBN: 978-94-010-5041-8
Online ISBN: 978-94-011-2360-0
eBook Packages: Springer Book Archive