Skip to main content

On Successive Approximation of Optimal Control of Stochastic Dynamic Systems

  • Chapter
Modeling Uncertainty

Part of the book series: International Series in Operations Research & Management Science ((ISOR,volume 46))

Abstract

An approximation theory of optimal control for nonlinear stochastic dynamic systems has been established. Based on the generalized Hamilton-Jacobi-Bellman equation of the cost function of nonlinear stochastic systems, general iterative procedures for approximating the optimal control are developed by successively improving the performance of a feedback control law until a satisfactory suboptimal solution is achieved. A successive design scheme using upper and lower bounds of the exact cost function has been developed for the infinite-time stochastic regulator problem. The determination of the upper and lower bounds requires the solution of a partial differential inequality instead of equality. Therefore it provides a degree of flexibility in the design method over the exact design method. Stability of the infinite-time sub-optimal control problem was established under not very restrictive conditions, and stable sequences of controllers can be generated. Several examples are used to illustrate the applicati on of the proposed approximation theory to stochastic control. It has been shown that in the case of linear quadratic Gaussian problems, the approximation theory leads to the exact solution of optimal control.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 219.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  • Al’brekht, E. G. (1961). On the optimal stabilization of nonlinear systems, J. Appl. Math. Mech. (ZPMM), 25,5, 1254–1266.

    Article  MathSciNet  Google Scholar 

  • Aoki, M. (1967). Optimization of Stochastic Systems, Academic Press, N.Y.

    MATH  Google Scholar 

  • Bellman, R. (1956). Dynamic Programming, Princeton University Press, Princeton, N.J.

    MATH  Google Scholar 

  • Doob, J. L. (1953). Markov Processes. Wiley, N.Y.

    MATH  Google Scholar 

  • Dynkin, E. B. (1953). Stochastic Processes. Academic Press, N.Y.

    Google Scholar 

  • Itô, K. (1951). On Stochastic Differential Equations? Memn Amer. Math. Soc., 4.

    Google Scholar 

  • Jaynes, E. T. (1957). Information theory and statistical mechanics. Physical Review, 4, 106.

    MathSciNet  MATH  Google Scholar 

  • Kushner, H. (1971). Introduction to Stochastic Control, Holt, Reinhardt and Winston, NY.

    MATH  Google Scholar 

  • Kwakernaak, H. and R. Sivan. (1972). Linear Optimal Control Systems, Wiley, N.Y.

    MATH  Google Scholar 

  • Leake, R.J. and R.-W. Liu. (1967). Construction of suboptimal control sequences, SIAM J. Control, 5,1, 54–63.

    Article  MathSciNet  Google Scholar 

  • Lukes, D. L. (1969). Optimal regulation of nonlinear dynamical systems. SIAM J. Control, 7,1, 75–100.

    Article  MathSciNet  Google Scholar 

  • Nishikawa, Y., N. Sannomiya and H. Itakura. (1962). A method for suboptimal design of nonlinear feedback systems, Automatica, 7,6, 703–712.

    Article  MathSciNet  Google Scholar 

  • Ohsumi, A. (1984). Stochastic control with searching a randomly moving target, Proc. Of American control Conference, San Diego, CA, 500–504.

    Google Scholar 

  • Panossian, H. V. (1988). Algorithms and computational techniques in stochastic optimal control, C.T. Lenodes (ed.), Control and Dynamic Systems, 28,1

    Google Scholar 

  • Rekasius, Z.V. (1964). Suboptimal design of intentionally nonlinear controllers. IEEE Trans. Automatic Control, AC-9,4, 380–386.

    Article  MathSciNet  Google Scholar 

  • Sage, A. P. and C. C. White. (1977). Optimun Systems Control, Prentice-Hall, Englewood Cliffs, N.J.

    Google Scholar 

  • Saridis, G. N. and Fei-Yue Wang. (1994). Suboptimal control of nonlinear stochastic systems, Control — Theory and Advanced Technology, Vol. 10,No. 4, Part l, pp.847–871.

    MathSciNet  Google Scholar 

  • Saridis, G. N. (1988). Entropy formulation for optimal and adaptive control. IEEE Trans. Automatic Control, AC-33,8, 713–721.

    Article  MathSciNet  Google Scholar 

  • Saridis, G. N. and J. Balaram. (1986). Suboptimal control for nonlinear systems. Control-Thoery and Advanced Technology (C-TAT), 2.3, 547–562.

    Google Scholar 

  • Saridis, G. N. and C. S. G. Lee. (1976). An approximation theory of optimal control for trainable manipulators. IEEE Trans. Systems, Man, and Cybernetics, SMC-9,3, 152–159.

    Article  MathSciNet  Google Scholar 

  • Skorokhod, A. V. (1965). Studies in the Theory of Random Processes. Addison-Wesley, Reading, MA.

    MATH  Google Scholar 

  • Wang, Fei-Yue and G. N. Saridis. (1992). Suboptimal control for nonlinear stochastic systems, Proceedings of 31st Conference on Decision and control, Tucson, AZ, Dec.

    Google Scholar 

  • Wonham, W. M. (1970). Random differential equations in control theory. A. T. BharuchaReid (ed.), Probabilistic Methods in Applied Mathematics, Academic Press, NY.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2002 Springer Science + Business Media, Inc.

About this chapter

Cite this chapter

Wang, FY., Saridis, G.N. (2002). On Successive Approximation of Optimal Control of Stochastic Dynamic Systems. In: Dror, M., L’Ecuyer, P., Szidarovszky, F. (eds) Modeling Uncertainty. International Series in Operations Research & Management Science, vol 46. Springer, New York, NY. https://doi.org/10.1007/0-306-48102-2_16

Download citation

Publish with us

Policies and ethics