Skip to main content

Abstract

Though artificial neural networks have been studied for five decades, they have experienced, in the past ten years, a very rapidly growing interest. Most of the applications in this domain are however either simulated on conventional machines or implemented on some specialized hardware dedicated to a given model. At the present day, there is no platform which is, at the same time, versatile enough to implement any neural-network model and learning rule, and fast enough to be used on large problems. This fact prompted many researchers to work toward the design of a generic neural-network computer (Treleaven et al 1989). As an intermediate step in this quest, a multi-model hardware implementation, whose architecture is influenced by biological considerations, is presented in this paper.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 129.00
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 169.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 169.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  • Blayo, F. and Hurat, P., “A VLSI Systolic Array Dedicated to Hopfield Neural Network,” in VLSI for Artificial Intelligence, J. G. Delgado-Frias and W. R. Moore (ed.), Norwell, MA: Kluwer Academic Publishers, pp. 255–264, 1989.

    Chapter  Google Scholar 

  • Blayo, F., Une Implantation Systolique des Algorithmes Connexionnistes, Ph.D. thesis N°904, EPFL, Lausanne, Switzerland, 1990.

    Google Scholar 

  • Chung, J.-H., Yoon, H. and Maeng, S. R., “A Systolic Array Exploiting the Inherent Parallelisms of Artificial Neural Networks,” Microprocessing and Microprogramming, vol. 33, no. 3, pp. 145–159, May 1992.

    Article  Google Scholar 

  • Gascuel, J.-D., Delaunay, E., Montoliu, L., Moobed, B. and Weinfeld, M., “A Software Reconfigurable Multi-Networks Simulator Using a Custom Associative Chip,” in Proc. of the Int. Joint Conference on Neural Networks, pp. 11–13–11–18, June 1992.

    Google Scholar 

  • Hopfield, J. J., “Neural Networks and Physical Systems with Emergent Collective Computational Abilities,” Proc. of the National Academy of Sciences, vol. 79, pp. 2254–2258, April 1982.

    Article  MathSciNet  Google Scholar 

  • Kohonen, T., “Analysis of a Simple Self-Organizing Process,” Biological cybernetics, vol. 44, pp. 135–140, 1982.

    Article  MATH  MathSciNet  Google Scholar 

  • Le Cun, Y., “A Learning Scheme for Asymmetric Threshold Network,” in Proc. of Cognitiva 85, June 1985.

    Google Scholar 

  • Lehmann, C. and Blayo, F., “A VLSI Implementation of a Generic Systolic Synaptic Building Block for Neural Networks,” in VLSI for Artificial Intelligence and Neural Networks, J. G. Delgado-Frias and W. R. Moore (ed.), New-York, NY: Plenum Press, pp. 325–334, 1991.

    Chapter  Google Scholar 

  • Niebur, D. and Germond, A. J., “Power System Static Security Assessment Using the Kohonen Neural Network Classifier,” in Proc. of the IEEE Power Industry Computer Application Conference, pp. 270–277, May 1991.

    Google Scholar 

  • Ramacher, U., Raab, W., Anlauf, J., Hachmann, U. and Wesseling, M., “SYNAPSE-X: A General-Purpose Neurocomputer,” in Proc. of the 2 nd Int. Conference on Microelectronics for Neural Networks, pp. 401–409, October 1991.

    Google Scholar 

  • Rosenblatt, F., “The Perceptron: A Probabilistic Model for Information Storage and Organization in the Brain,” Psychological Review, vol. 65, pp. 386–408, 1958.

    Article  MathSciNet  Google Scholar 

  • Treleaven, P., Pacheco, M. and Vellasco, M., “VLSI architectures for neural networks,” IEEE Micro, vol. 9, pp. 8–27, 1989.

    Article  Google Scholar 

  • Widrow, B. and Hoff, M. E., “Adaptive Switching Circuits,” in IRE-WESCON Convention Record, pp. 96–104, 1960.

    Google Scholar 

Download references

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 1994 Springer Science+Business Media New York

About this chapter

Cite this chapter

Viredaz, M.A., Lehmann, C., Blayo, F., Ienne, P. (1994). MANTRA: A Multi-Model Neural-Network Computer. In: Delgado-Frias, J.G., Moore, W.R. (eds) VLSI for Neural Networks and Artificial Intelligence. Springer, Boston, MA. https://doi.org/10.1007/978-1-4899-1331-9_9

Download citation

  • DOI: https://doi.org/10.1007/978-1-4899-1331-9_9

  • Publisher Name: Springer, Boston, MA

  • Print ISBN: 978-1-4899-1333-3

  • Online ISBN: 978-1-4899-1331-9

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics