Abstract
The optimization of stochastic control systems has received increased attention in recent years [1, 2, 3, 4]. The stochastic nature of a control situation can arise from a variety of sources such as measurement noise, parameter drift, or other, possibly external disturbances. Such random properties represent a description of inaccurately known disturbances, signals, or system changes. If there exists much a priori information or assumptions concerning plant dynamics and/or statistical properties, then a stochastic controller selecting control policies on the basis of a state estimation procedure may be appropriate. With known statistical properties, the estimation is amenable to a Kaiman filter or a Bayesian approach. However, the lack of sufficient a priori assumptions concerning the stochastic plant dynamics requires a learning controller to interact with the stochastic plant to gain information for improving performance. Several algorithms for the learning controller for seeking optimal control policies have been considered elsewhere [5, 6, 7]. One such algorithm which has received considerable attention is stochastic approximation; it is a stochastic form of a more general class of “gradient” or hill-climbing search algorithms. Such algorithms, however, tend to have local optimization properties. Hence, although often providing rapid convergence, such an algorithm can “miss” an extremum while seeking a local minimum or maximum.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
H. F. Karreman, Stochastic Optimization and Control, John Wiley and Sons, New York, 1968.
R. S. Bucy and P. D. Joseph, Filtering for Stochastic Processes With Applications to Guidance, John Wiley and Sons, New York, 1968.
T. Kailath, et.al., Stochastic Problems in Control, Symposium held at 1968 JACC, University of Michigan, June 1968.
J. Meditch, Stochastic Optimal Linear Estimation and Control, McGraw-Hill, New York, 1969.
J. S. Riordon, “An Adaptive Automaton Controller for Discrete-Time Markov Processes,” in Preprints of the 1968 JACC, Univ. of Michigan, June 1968.
K. S. Fu and J. M. Mendel, Adaptive, Learning, and Pattern Recognition Systems, Academic Press, New York, 1970.
Proceedings of the U.S.-Japan Seminar on Learning Process in Control Systems, held August 18–20, 1970, in Nagoya, Japan
R. W. McLaren, “Application of a Continuous-Valued Control Algorithm to the On-Line Global Optimization of Stochastic Control Systems,” Proceedings of the National Electronics Conference, Chicago, Ill., 1969.
K. S. Fu and R. W. McLaren, “An Application of Stochastic Automata to the Synthesis of Learning Systems,” Purdue Univ. Tech. Rept. No. EE 65-17, School of Electrical Engineering, Purdue Univ., Lafayette, Ind., Sept. 1965.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 1971 Plenum Press, New York
About this chapter
Cite this chapter
McLaren, R.W. (1971). A Continuous-Valued Learning Controller for the global optimization of Stochastic Control Systems. In: Fu, K.S. (eds) Pattern Recognition and Machine Learning. Springer, Boston, MA. https://doi.org/10.1007/978-1-4615-7566-5_23
Download citation
DOI: https://doi.org/10.1007/978-1-4615-7566-5_23
Publisher Name: Springer, Boston, MA
Print ISBN: 978-1-4615-7568-9
Online ISBN: 978-1-4615-7566-5
eBook Packages: Springer Book Archive