Introduction
Optimal control is one of the main techniques of modern control design, as it has been for many years. The linear-quadratic theory of optimal control design is well established and has various forms including the receding horizon approach for a robust, easily implementable variation of the theory. It is also useful in H ∞ control in the well-known state-space game theoretic formulation. Obtaining the ‘best’ controller in any given circumstance is clearly important, but for general nonlinear systems, one is led to the solution of an extremely difficult (in general, non-smooth) partial differential equation. This makes the existing general nonlinear theory very difficult to apply.
In this chapter we shall show how to use the iteration technique developed above to solve nonlinear optimal control problems. In the next section we shall outline the classical linear quadratic regulator theory and derive the optimal feedback control in terms of the solution of a Riccati equation.We shall also indicate the modifications necessary to solve the linear tracking problem. The generalisation to nonlinear systems will be given in Section 6.3 and some examples will be presented in Section 6.4. Some comments on viscosity solutions of the Hamilton-Jacobi-Bellman (HJB) equation and the optimality of the solution will be given in Section 6.5.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Navarro-Hernandez, C., Banks, S.P., Aldeen, M.: Observer Design for Non-linear Systems using Linear Approximations. IMA J. Math. Cont and Inf. 20, 359–370 (2003)
Zheng, J., Banks, S.P., Alleyne, H.: Optimal Attitude Control for Three-Axis Stabilised Flexible Spacecraft. Acta Astronautica 56, 519–528 (2005)
Cimen, T., Banks, S.P.: Non-linear Optimal Tracking Control with Application to Super-Tankers for Autopilot Design. Automatica 40(11), 1845–1863 (2004)
Lions, P.L.: Generalized Solutions of Hamilton-Jacobi Equations. Pitman, Boston (1982)
Bardi, M., Capuzzo-Dolcetta, I.: Optimal Control and Viscosity Solutions of Hamilton-Jacobi-Bellman Equations. Birkhauser, Boston (1997)
Hugues-Salas, O., Banks, S.P.: Control of Chaos for Secure Communication. Int. J. Bifur. and Chaos (to appear, 2009)
Cimen, T., Banks, S.P.: Global Optimal Feedback Control for General Non-linear Systems with Non-quadratic Performance Criteria. Sys. Cont. Letts. 53, 327–346 (2004)
Rights and permissions
Copyright information
© 2010 Springer London
About this chapter
Cite this chapter
Tomás-Rodríguez, M., Banks, S.P. (2010). Optimal Control. In: Linear, Time-varying Approximations to Nonlinear Dynamical Systems. Lecture Notes in Control and Information Sciences, vol 411. Springer, London. https://doi.org/10.1007/978-1-84996-101-1_6
Download citation
DOI: https://doi.org/10.1007/978-1-84996-101-1_6
Publisher Name: Springer, London
Print ISBN: 978-1-84996-100-4
Online ISBN: 978-1-84996-101-1
eBook Packages: EngineeringEngineering (R0)