Skip to main content

A generic formulation of neural nets as a model of parallel and self-programming computation

  • Formal Tools and Computational Models of Neurons and Neural Net Architectures
  • Conference paper
  • First Online:
Biological and Artificial Computation: From Neuroscience to Technology (IWANN 1997)

Part of the book series: Lecture Notes in Computer Science ((LNCS,volume 1240))

Included in the following conference series:

Abstract

In the same way the more conventional fields of computer science need some theory, including the mathematical foundations of the calculus and the establishment of formal models, neural computation also needs its own. The basic requirements of this model are modularity, “small grain”, high connectivity, parametric local computation and some capacity of self-programming by means of the adjustment of these parameters.

We present here a proposal in this line that allows the integration in a single frame of all current models (analogic, logic and inferential) and makes clear the natural way to bridge the symbolic and connectionistic perspectives of AI extending the model of local computation to hierarchic graphs, building networks by joining graphs and studying the set of operators we need for modifying local computation parameters values.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. J. Mira, et al. Aspectos básicos de la inteligencia artificial. (Sanz y Torres, 1995).

    Google Scholar 

  2. J. Mira, et al. Cooperative processes at the symbolic level in cerebral dynamics: reliability and fault tolerance. In Brain Processes Theories and Models, R. Moreno-Díaz and J. Mira (Eds.) (The MIT Press. Cambridge, MA, 1995).

    Google Scholar 

  3. J. Mira and A.E. Delgado. Computación neuronal, in Aspectos básicos de la inteligencia artificial, Cap. 11, 485–575. J. Mira, et al. (Eds.) (Sanz y Torres, 1995).

    Google Scholar 

  4. N. J. Nilsson. Principles of Artificial Intelligence. (Tioga Publishing, 1980).

    Google Scholar 

  5. D.E. Rumelhart, et al. Learning internal representations by error propagation, in Parallel Distributed Processing: Explorations in the Microestructures of Cognition, Vol. 1, 318–362. D.E Rumelhart and J.L. McClelland (Eds.) (Cambridge, MA. The MIT Press, 1986)

    Google Scholar 

  6. P. H. Winston. Artificial Intelligence, Third Edition. (Addison Wesley, 1992).

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

José Mira Roberto Moreno-Díaz Joan Cabestany

Rights and permissions

Reprints and permissions

Copyright information

© 1997 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Mira, J., Herrero, J.C., Delgado, A.E. (1997). A generic formulation of neural nets as a model of parallel and self-programming computation. In: Mira, J., Moreno-Díaz, R., Cabestany, J. (eds) Biological and Artificial Computation: From Neuroscience to Technology. IWANN 1997. Lecture Notes in Computer Science, vol 1240. Springer, Berlin, Heidelberg. https://doi.org/10.1007/BFb0032477

Download citation

  • DOI: https://doi.org/10.1007/BFb0032477

  • Published:

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-63047-0

  • Online ISBN: 978-3-540-69074-0

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics