Skip to main content

Part of the book series: Perspectives in Neural Computing ((PERSPECT.NEURAL))

Abstract

A novel connectionist neural network model based on a propagation rule which contains Multilayer Perceptron (MLP) and Radial Basis Function (RBF) parts is introduced in this paper. The network using this propagation rule is known as a Conic Section Function Network (CSFN). Two different strategies have been used for training the network. The contact lens fitting problem has been considered to demonstrate the performance of the training algorithms. The performances of a standard MLP trained by back propagation, a fast back propagation with adapted learning rates, a standard RBFN using Matlab Neural Network software toolbox, and the proposed algorithm are compared for this particular problem.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Warwick K, Irwin GW, and Hunt KJ, Neural networks for control and systems, Peter Peregrinus Ltd., London, UK, 1992.

    Google Scholar 

  2. Dorffner G, Unified framework for MLPs and RBFNs: Introducing conic section function networks, Cybernetics and Systems, 1994, Vol.25 pp.511554.

    Google Scholar 

  3. Chen S, Cowan CFN, Grant PM, Orthogonal least squares learning algorithm for radial basis function networks, IEEE Transactions on Neural Networks, March 1991, Vol. 2, No. 2, pp. 302–309.

    Article  Google Scholar 

  4. Rumelhart DE, Hinton GE, Williams RJ, Learning internal representations by error propagation, in Parallel Distributed Processing, ed. D.E.Rumelhart and J.L.McClelland, Cambridge, MA: MIT Press, 1987, Vol. 1, pp. 318–362.

    Google Scholar 

  5. Haykin S, Neural Networks: A comprehensive foundation, Macmillan College Publishing, New York, 1994.

    Google Scholar 

  6. Cendrowska J, PRISM: An algorithm for inaucing modular rules, International Journal of Man-Machine Studies, 1987, Vol. 27, pp. 349–370.

    Article  MATH  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 1998 Springer-Verlag London Limited

About this paper

Cite this paper

Yildirim, T., Marsland, J.S. (1998). A unified framework for connectionist models. In: Bullinaria, J.A., Glasspool, D.W., Houghton, G. (eds) 4th Neural Computation and Psychology Workshop, London, 9–11 April 1997. Perspectives in Neural Computing. Springer, London. https://doi.org/10.1007/978-1-4471-1546-5_3

Download citation

  • DOI: https://doi.org/10.1007/978-1-4471-1546-5_3

  • Publisher Name: Springer, London

  • Print ISBN: 978-3-540-76208-9

  • Online ISBN: 978-1-4471-1546-5

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics