Abstract
We are concerned with a "simple" partially observed control problem first considered by Beneš and Karatzas in [1], where the relation between Zakai's equation and Mortensen's equation was revealed. In [1] it was pointed out that the dynamic programming techniques do not give us an existence theorem (- as it would be in the completely observed case -), but it was shown that the value function is a lower bound for the solution of Mortensen's equation, if such a solution exists.
In this contribution, we do not attack the existence problem, and our results do not contribute any idea towards such a result, as for as we see.
We consider Mortensen's equation and use the concept of viscosity solution for an attempt towards a better understanding of the important role this equation plays in stochastic control. A viscosity solution is defined to be a subsolution and a supersolution of the problem. We are thus trying to describe the sets of subsolutions and supersolutions and give some control-theoretical interpretation of these sets: Subsolutions may be seen as solutions of a suboptimal problem and supersolutions are related with superoptimal solutions of the control problem. Our results are related to those of Bensoussan in [2], and a lot of techniques are taken from [2] and [3]. The reference on viscosity solutions related to our problem is [6].
This article treats a slimmed problem which was described in more detail in [4]. Some of the technical details which are left out here, can be found there.
This work was supported by the SFB 72 of the DFG at the Universität Bonn, by the British Council during a visit to the Mathematical Institute at the University of Warwick, and by the AFF at the Universität Konstanz
Preview
Unable to display preview. Download preview PDF.
References
V.E. BENEŠ, I. KARATZAS, On the relation of ZAKAI's and MORTENSEN's equations, SIAM.J. Control and Opt. 21 (3) 1983, 472–489
A. BENSOUSSAN, Maximum principle and dynamic programming approaches of the optimal control of partially observed diffusions, Stochastics, 9 (3) 1983, 169–222
A. BENSOUSSAN, M. ROBIN, On the convergence of the discrete time dynamic programming equation for general semi groups, SIAM.J. Control and opt. 20 (5) 1982, 722–746
M. KOHLMANN, ɛ — optimal controls in partially observed control: a maximum principle and viscosity solutions for Mortensen's equation, preprint Universität Konstanz
P.L. LIONS, Generalized solutions of Hamilton-Jacobi equations, Pitman APP, Boston-London-Melbourne, 1982
P.L. LIONS, Optimal control of diffusion processes and Hamilton-Jacobi-Bellman equations, part 2: viscosity solutions and uniqueness, Comm. P.D.E 8 (11) 1983, 1229–1276
M.G. CRANDALL, P.L. LIONS, Hamilton-Jacobi equations in infinite dimensions, part I: uniqueness of viscosity solutions, preprint Université de Paris IX — Dauphine (1984)
I. EKELAND, Nonconvex minimization problems, Bull.AMS 1 (3) 1979, 443–474
M. KOHLMANN, Concepts for the derivation of optimal partially observed controls, preprint Universität Konstanz 1984
R.J. ELLIOTT, M. KOHLMANN, The variational principle and stochastic optimal control, Stochastics 3, 1980, 229–241
Author information
Authors and Affiliations
Editor information
Rights and permissions
Copyright information
© 1986 Springer-Verlag
About this paper
Cite this paper
Kohlmann, M. (1986). Viscosity solutions in partially observed control. In: Christopeit, N., Helmes, K., Kohlmann, M. (eds) Stochastic Differential Systems. Lecture Notes in Control and Information Sciences, vol 78. Springer, Berlin, Heidelberg. https://doi.org/10.1007/BFb0041166
Download citation
DOI: https://doi.org/10.1007/BFb0041166
Published:
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-16228-5
Online ISBN: 978-3-540-39767-0
eBook Packages: Springer Book Archive