Skip to main content

Recurrent Neural Networks: Some Systems-Theoretic Aspects

  • Chapter
Dealing with Complexity

Abstract

Recurrent nets have been introduced in control, computation, signal processing, optimization, and associate memory applications. Given matrices A ∈ ℝn ×n, B ∈ ℝn ×m, C ∈ ℝp ×n, as well as a fixed Lipschitz scalar function σ : ℝ → ℝ, the continuous time recurrent network Σ with activation function σ and weight matrices (A, B,C) is given by:

$$\frac{{dx}}{{dt}}(t) = {\overrightarrow \sigma ^{(n)}}\left( {Ax(t) + Bu(t)} \right) , y\left( t \right) = Cx\left( t \right)$$
(1)

where \({\overrightarrow \sigma ^{\left( n \right)}}\): ℝn → ℝn is the diagonal map

$${\overrightarrow \sigma ^{\left( n \right)}}:\left( {\begin{array}{*{20}{c}}{{x_1}} \\ \vdots \\ {{x_n}}\end{array}} \right) \mapsto \left({\begin{array}{*{20}{c}}{\sigma \left( {{x_1}} \right)} \\ \vdots \\ {\sigma \left( {{x_n}} \right)} \end{array}} \right)$$
(2)

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

eBook
USD 16.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Albertini, F., and P. Dai Pra, “Forward accessibility for recurrent neural networks,” IEEE Trans. Automat. Control 40 (1995): 1962–1968

    Article  MATH  MathSciNet  Google Scholar 

  2. Albertini, F., and E.D. Sontag, “For neural networks, function determines form,” Neural Networks 8 (1993): 975–990.

    Article  Google Scholar 

  3. Albertini, F., and E.D. Sontag, “Uniqueness of weights for recurrent nets,” Systems and Networks: Math Theory and Applics,Proc. MTNS ‘83, Vol. 2, Akademie Verlag, Regensburg, pp. 599–602. Extended version: http: //www.math.rutgers.edufsontag/FTP DIR/93mtns-nn-extended.ps.gz

    Google Scholar 

  4. Albertini, F., and E.D. Sontag, “State observability in recurrent neural networks,” Systems ei Control Letters 22 (1994): 235–244.

    Article  MATH  MathSciNet  Google Scholar 

  5. Hautus, M., “A set of IP-functions,” unpublished manuscript, Eindhoven University, August 1993.

    Google Scholar 

  6. Leshno, M., V.Ya. Lin, A. Pinkus, and S. Schocken, “Multilayer feedforward networks with a non-polynomial activation function can approximate any function,” Neural Networks 8 (1993): 861–867.

    Article  Google Scholar 

  7. Siegelmann, H.T., and E.D. Sontag, “Analog computation, neural networks, and circuits,” Theor. Comp. Sci. 131 (1994): 331–360.

    Article  MATH  MathSciNet  Google Scholar 

  8. Siegelmann, H.T., and E.D. Sontag, “On the computational power of neural nets,” J. Comp. Syst. Sci. 50 (1995): 132–150.

    Article  MATH  MathSciNet  Google Scholar 

  9. Sontag, E.D., “Neural nets as systems models and controllers,” in Proc. Seventh Yale Workshop on Adaptive and Learning Systems, pp. 73–79, Yale University, 1992.

    Google Scholar 

  10. Sontag, E.D., “Neural networks for control,” in Essays on Control: Perspectives in the Theory and its Applications ( H.L. Trentelman and J.C. Willems, eds.), Birkhauser, Boston, 1993, pp. 339–380.

    Google Scholar 

  11. Sontag, E.D., Mathematical Control Theory: Deterministic Finite Dimensional Systems, Springer, New York, 1990.

    MATH  Google Scholar 

  12. Sontag, E.D., and H.J. Sussmann, “Complete controllability of continuous-time recurrent neural networks,” Systems and Control Letters 30 (1997): 177–183.

    Article  MATH  MathSciNet  Google Scholar 

  13. Sussmann, H.J., “Existence and uniqueness of minimal realizations of nonlinear systems,” Math. Sys. Theory 10 (1977): 263–284.

    Article  MathSciNet  Google Scholar 

  14. Sussmann, H.J., “Uniqueness of the weights for minimal feedforward nets with a given input-output map,” Neural Networks 5 (1992): 589–593.

    Article  Google Scholar 

  15. Zbikowski, R., “Lie algebra of recurrent neural networks and identifiability,” Proc. Amer. Auto. Control Conf., San Francisco, 1993, pp. 2900–2901.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 1998 Springer-Verlag London Limited

About this chapter

Cite this chapter

Kárný, M., Warwick, K., Kůrková, V. (1998). Recurrent Neural Networks: Some Systems-Theoretic Aspects. In: Kárný, M., Warwick, K., Kůrková, V. (eds) Dealing with Complexity. Perspectives in Neural Computing. Springer, London. https://doi.org/10.1007/978-1-4471-1523-6_1

Download citation

  • DOI: https://doi.org/10.1007/978-1-4471-1523-6_1

  • Publisher Name: Springer, London

  • Print ISBN: 978-3-540-76160-0

  • Online ISBN: 978-1-4471-1523-6

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics