Abstract
This paper addresses the design and experimental characterization of a novel hybrid neural network, in which two distinct classical architectures interact: the Hopfield neural network and the Multi-Layer Perceptron. This hybrid neural system, named MLP+H (from MLP + Hopfield), presents a better performance than each one of the two classical architectures when considered in separate. In addition, it has the ability to deal with different classes of data than that normally allowed by the two conventional architectures. For example, while the Hopfield networks deal with binary patterns and the MLPs with information that always have some analog characteristics (due to the continuous nature of the MLP nodes), the MLP+H deals with analog inputs and purely digital outputs. Moreover, the MLP+H allows reduced training times when compared with the MLP architecture, also presenting compactness and flexibility in dealing with different applications, as implementation of anti-noise filters, an application currently under study.
Ackowledgments to the University of Sao Paulo (USP) and Foundation for the Support of Research in the State of Săo Paulo (FAPESP) for the financial support to this work.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Amit, D. J. Modeling Brain Function: the World of Attractor Neural Networks, New York: Cambridge University Press, 1989.
Haykin, Simon. Neural Networks: a comprehensive foundation. MacMillan Pub. Co., 1994.
Hopfield, J. J.. “Neural Networks and Physical Systems with Emergent Collective Computational Abilities”, in Proc. Nat. Acad. of Sciences, USA, Ap. 1982, V. 79, p. 2554-58.
Kosko, B., “Bidirectional Associative Memories”, IEEE Trans. on Systems, Man, and Cybernetics, Vol. 18, pp. 49–60. Jan-Feb, 1988.
McEliece, R., E. Posner, E., and S. Venkatesh, “The Capacity of the Hopfield Associative Memory” Information Theory, IEEE Trans. on, Vol.: 33 Issue: 4, Jul 1987 pp. 461–482
Oppenheim, Alan V., and R. W. Schafer, Discrete-Time Signal Processing, Englewood Cliffs, NJ, Prentice Hall, 2000.
Rumelhart, D. E. and J. L. McClelland, eds., Parallel Distributed Processing: Explorations in the Microstructure of Cognition, VI, Cambridge, MA: MIT Press, 1986.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2003 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Hernandez, C.S.O.e.E.D.M. (2003). MLP+H: a Hybrid Neural Architecture Formed by the Interaction of Hopfield and Multi-Layer Perceptron Neural Networks. In: Mira, J., Álvarez, J.R. (eds) Computational Methods in Neural Modeling. IWANN 2003. Lecture Notes in Computer Science, vol 2686. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-44868-3_22
Download citation
DOI: https://doi.org/10.1007/3-540-44868-3_22
Published:
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-40210-7
Online ISBN: 978-3-540-44868-6
eBook Packages: Springer Book Archive