Abstract
We investigate the distribution of the local fields in attractor neural networks using a simple model similar to that studied by Amit and Brunel (1993). This model consists of two layers of neurons: input layer and recurrent layer. The stimulus of the input layer, by means of feed-forward connections, creates receptive fields on the neurons of the recurrent layer. We discuss the dynamic properties and the basins of attraction of the network and propose a procedure to stabilize the system, driving it into a basin of attraction, without introducing explicitly inhibitory subnetwork. We also investigate the effect of the stimulus applied on the attractor network and propose a criterium of learnability of new patterns. The proposed techniques can be applied to more complex topologies of linked recurrent networks and could be used for designing hardware realizations with application to neurocomputers.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Amit A. and Brunel N. (1993): Adequate input for learning in attractor neural network, Network 4, 177.
Korutcheva E. and Koroutchev K. (1996): On the local-field distribution in attractor neural networks, Int. J. Mod. Phys. C 7, 463.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 1997 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Korutcheva, E., Koroutchev, K. (1997). Local Field Distribution in Attractor Neural Networks: Effect of the Stimulus. In: Garrido, P.L., Marro, J. (eds) Fourth Granada Lectures in Computational Physics. Lecture Notes in Physics, vol 493. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-662-14148-9_21
Download citation
DOI: https://doi.org/10.1007/978-3-662-14148-9_21
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-662-14150-2
Online ISBN: 978-3-662-14148-9
eBook Packages: Springer Book Archive