Skip to main content

Stochastic controllability and stochastic Lyapunov functions with applications to adaptive and nonlinear systems

  • Conference paper
  • First Online:
Stochastic Differential Systems

Part of the book series: Lecture Notes in Control and Information Sciences ((LNCIS,volume 126))

Abstract

Sufficient conditions are established under which the law of large numbers and related ergodic theorems hold for nonlinear stochastic systems operating under feedback. It is shown that these conditions hold whenever a moment condition is satisfied, which may be interpreted as a generalization of the martingale property.

If in addition a stochastic controllability condition holds, then it is shown that the underlying distributions governing the system converge to an invariant probability at a geometric rate.

These results are illustrated with general examples from linear, nonlinear, and adaptive control theory.

The key assumption used is that a Markov chain with stationary transition probabilities exists which serves as a state process for the closed loop system.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  • Athreya, K.B. and P. Ney (1980) “Some aspects of ergodic theory and laws of large numbers for Harris recurrent Markov chains” Colloquia Mathematica Societatis János Bolyai 32 Nonparametric statistical inference, Budapest, Hungary, pp. 41–56.

    Google Scholar 

  • Beneš, V.E. (1968) “Finite Regular Invariant Measures for Feller Processes”, J. Appl. Prob., 5, p.203.

    Google Scholar 

  • Bhatia, N.P. and G.P. Szegö (1970) Stability theory of dynamical systems, Springer-Verlag Berlin, Heidelberg, New York.

    Google Scholar 

  • Billingsley, P. (1968). Convergence of Probability Measures. John Wiley & Sons, N.Y.

    Google Scholar 

  • Brown, J.B. (1976) Ergodic theory and topological dynamics, Academic Press, N.Y.

    Google Scholar 

  • Caines, P.E. (1988) Linear stochastic systems, John Wiley & Sons, N.Y.

    Google Scholar 

  • Chan, K.S. (1986) Topics in Nonlinear Time Series Analysis Thesis, Dept. of Mathematics, Princeton University.

    Google Scholar 

  • Cogburn, R. (1975) “A uniform theory for sums of Markov chain transition probabilities”, Ann. Probab. 3 pp. 191–214.

    Google Scholar 

  • Doob, J.L. (1953). Stochastic Processes. John Wiley & Sons, N.Y.

    Google Scholar 

  • Foguel, S.R. (1969) “Positive operators on C(X)” Proceedings AMS 22 pp. 295–297.

    Google Scholar 

  • Foster, F.G. (1953) “On the stochastic matrices associated with certain queueing processess” Annals of Mathematical Statistics, 24 pp. 355–360.

    Google Scholar 

  • Goodwin, G.C., P.J. Ramadge and P.E. Caines, (1981). Discrete Time Stochastic Adaptive Control, SIAM J. Contr. Optimiz., 19, 829. Corrigendum, 20 No.6, 1982, p.893.

    Google Scholar 

  • Guo, L. and S.P. Meyn (1988) “Short memory adaptive control” to appear in the International J. of Adaptive Control and Signal Processing.

    Google Scholar 

  • Has'minskii, R.Z. (1980) Stochastic stability of differential equations sijthoff & Noordhoff Alphen an den Rijn, The Netherlands, Rockville, Maryland USA.

    Google Scholar 

  • Jakubczyk, B. and E.D. Sontag (1988) “Controllability of nonlinear discrete time systems: A liealgebraic approach” to appear Siam J. Contr. Optimiz.

    Google Scholar 

  • Kalman, R.E. and J.E. Bertram (1960) “Control system analysis and design by the second method of Lyapunov” Trans. ASME Ser. D: J. Basic Eng. 82 No. 2 pp. 371–400.

    Google Scholar 

  • Kushner, H.J. (1967). Stochastic stability and control, Academic Press, N.Y.

    Google Scholar 

  • Meyn S.P. (1988a) “Ergodic theorems for discrete time stochastic systems using a generalized stochastic Lyapunov function” submitted to Siam J. Control and Optimization.

    Google Scholar 

  • Meyn S.P. (1988b) “Ergodic theory and topological dynamics for discrete time controlled systems” in preperation.

    Google Scholar 

  • Meyn, S.P. and P.E. Caines, (1987) “A new approach to stochastic adaptive control”, IEEE Transactions on Automatic Control, AC-32, Number 3.

    Google Scholar 

  • Meyn, S.P. and P.E. Caines (1988). Asymptotic Behavior of Stochastic Systems Possessing Markovian Realizations. submitted to Siam J. Control and Optimization.

    Google Scholar 

  • Nummelin, E. (1984) General Irreducible Markov Chains and Non-Negative Operators Cambridge University Press, Cambridge.

    Google Scholar 

  • Nummelin, E. and P. Tuominen. (1982) “Geometric ergodicity of Harris recurrent Markov chains with applications to renewal theory”. Stochastic Processes and their Applications, 12 187–202.

    Google Scholar 

  • Orey, S. (1971) Limit Theorems for Markov Chain Transition Probabilities, Van Nostrand Reinhold Mathematical studies 34, Van Nostrand, London.

    Google Scholar 

  • Popov, V.M. (1973) Hyperstability of Control Systems, Springer-Verlag, Berlin, and Editura Academici Bucuresti.

    Google Scholar 

  • Revuz, D. (1975) Markov Chains, North Holland Publishing Co., Amsterdam, 1975.

    Google Scholar 

  • Rosenblatt, M. (1971) Markov Processes: Structure and Asymptotic Behaviour. Springer-Verlag, Berlin, Heidelberg, New York.

    Google Scholar 

  • Safonov, M.G. (1980) Stability and robustness of multivariable feedback systems MIT Press, Cambridge, Mass.

    Google Scholar 

  • Solo, V. (1978) Time series recursions and stochastic approximation, Ph.D. dissertation, The Australian National University, Sept. 1978.

    Google Scholar 

  • Sontag, E.D. (1983) “A Lyapunov-like characterization of asymptotic controllability” Siam J. Contr. Optimiz. 21 No. 3 pp. 462–471.

    Google Scholar 

  • Tuominen, P. and R.L. Tweedie (1979) “Markov chains with continuous components” Proc. London Math. Society, Series 3, Vol. 38.

    Google Scholar 

  • Tweedie, R.L. (1983) “Criteria for rates of convergence of Markov chains, with application to queueing and storage theory”, in Probability, Statistics and Analysis, London Math. Society Lecture Note Series 79 Ed. J.F.C. Kingman and G.E.H. Reuter, Cambridge University Press, Cambridge.

    Google Scholar 

  • Zames, G. (1966) “On the Input-Output Stability of Time Varying Non-Linear Feedback Systems”, Part I: IEEE Transactions on Automatic Control, AC-11, Number 2, pp. 228–238, April 1966. Part II: IEEE Transactions on Automatic Control, AC-11, Number 2, pp. 465–476, April 1966.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Norbert Christopeit Kurt Helmes Michael Kohlmann

Rights and permissions

Reprints and permissions

Copyright information

© 1989 Springer-Verlag

About this paper

Cite this paper

Meyn, S.P., Caines, P.E. (1989). Stochastic controllability and stochastic Lyapunov functions with applications to adaptive and nonlinear systems. In: Christopeit, N., Helmes, K., Kohlmann, M. (eds) Stochastic Differential Systems. Lecture Notes in Control and Information Sciences, vol 126. Springer, Berlin, Heidelberg. https://doi.org/10.1007/BFb0043789

Download citation

  • DOI: https://doi.org/10.1007/BFb0043789

  • Published:

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-51299-8

  • Online ISBN: 978-3-540-46188-3

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics