Skip to main content

Solving the CLM Problem by Discrete-Time Linear Threshold Recurrent Neural Networks

  • Conference paper
Artificial Neural Networks – ICANN 2009 (ICANN 2009)

Part of the book series: Lecture Notes in Computer Science ((LNTCS,volume 5768))

Included in the following conference series:

  • 1994 Accesses

Abstract

The competitive layer model (CLM) can be described by the optimization problem that is formulated with the CLM energy function. The minimum points of CLM energy function can be achieved by running some proper recurrent neural networks. In other words, the CLM can be implemented by the recurrent neural networks. This paper proposes the discrete-time linear threshold recurrent networks to solve the CLM problem. The conditions for the stable attractors of the networks are obtained, which just correspond to the conditions of the minimum points of CLM energy function established in the literature before. Therefore, the proposed network can be used to implement the CLM.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 129.00
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 169.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Ritter, H.: A Spatial Approach for Feature Linking. In: Proc. Int. Neural Networks Conf., Paris, vol. 2, pp. 898–901 (1990)

    Google Scholar 

  2. Von der Malsburg, C.: Binding in Models of Perception and Brain Function. Current Opinion in Neurobiology 5, 520–526 (1995)

    Article  Google Scholar 

  3. Wersing, H., Steil, J.J., Ritter, H.: A Competitive Layer Model for Feature Binding and Sensory Segmentation. Neural Computation 13, 357–387 (2001)

    Article  MATH  Google Scholar 

  4. Yi, Z.: Foundations of Implementing the Competitive Layer Model by Lotka-Volterra Recurrent Neural Networks. Conditionally Accepted by IEEE Trans. Neural Networks (2009)

    Google Scholar 

  5. Zhang, L., Yi, Z., Yu, J.: Multiperiodicity and Attractivity of Delayed Recurrent Neural Networks with Unsaturating Piecewise Linear Transfer Functions. IEEE Trans. Neural Networks 19, 158–167 (2008)

    Article  Google Scholar 

  6. Yu, J., Yi, Z., Zhang, L.: Representations of Continuous Attractors of Recurrent Neural Networks. IEEE Trans. Neural Networks 20, 368–372 (2009)

    Article  Google Scholar 

  7. Yi, Z., Tan, K.K.: Multistability Analysis of Discrete Recurrent Neural Networks with Unsaturating Piecewise Linear Transfer Functions. IEEE Trans. Neural Networks 15, 329–336 (2004)

    Article  Google Scholar 

  8. Goldberg, J.L.: Matrix Theory with Applications. McGraw-Hill, New York (1992)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2009 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Zhang, L., Heng, P.A., Yi, Z. (2009). Solving the CLM Problem by Discrete-Time Linear Threshold Recurrent Neural Networks. In: Alippi, C., Polycarpou, M., Panayiotou, C., Ellinas, G. (eds) Artificial Neural Networks – ICANN 2009. ICANN 2009. Lecture Notes in Computer Science, vol 5768. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-04274-4_102

Download citation

  • DOI: https://doi.org/10.1007/978-3-642-04274-4_102

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-642-04273-7

  • Online ISBN: 978-3-642-04274-4

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics