Skip to main content

Monotone Optimal Policies for Left-Skip-Free Markov Decision Processes

  • Chapter
Applied Probability and Stochastic Processes

Abstract

In a previous paper (Stidham and Weber [9]), we considered a variety of models for optimal control of the service rate in a queueing system, in which the objective is to minimize the limiting average expected cost per unit time. By standard techniques, we showed how to convert such a problem into an equivalent problem in which the objective is to minimize the expected total (undiscounted) cost until the first entrance into state zero. Under weak assumptions on the one-stage (service plus holding) costs and transition probabilities, we showed that an optimal policy is monotonic, that is, a larger service rate is used in larger states. In contrast to previous models in the literature on control of queues, we assumed that the holding cost was nondecreasing, but not necessarily convex, in the state. A common assumption in all the models was that services take place one at a time, so that the state transitions are skip-free to the left: a one-step transition from state i to a state j < i − 1 is impossible. Many queueing models have this property, including all birth—death models, as well as a variety of M/GI/1-type models, including models with batch arrivals, phase-type service times, and LCFS-PR queue discipline.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 169.00
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 219.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 219.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Bertsekas, D. Dynamic Programming and Optimal Control, Vol. II. Athena Scientific, Belmont, MA, 1995.

    MATH  Google Scholar 

  2. Keilson, J. The use of Green’s functions in the study of bounded random walks with applications to queuing theory. J. Math. Phys. 41, 42–52, 1962.

    MathSciNet  MATH  Google Scholar 

  3. Keilson, J. Green’s Function Methods in Probability Theory. Griffin, London, 1965.

    MATH  Google Scholar 

  4. Kulkarni, V. G. Modeling and Analysis of Stochastic Systems. Chapman-Hall, London, 1995.

    MATH  Google Scholar 

  5. Lippman, S. A. Applying a new device in the optimization of exponential queuing systems. Oper. Res. 23, 687–710, 1975.

    Article  MathSciNet  MATH  Google Scholar 

  6. Puterman, M. Markov Decision Processes: Discrete Stochastic Dynamic Programming. Wiley, New York, 1994.

    MATH  Google Scholar 

  7. Schäl, M. Conditions for optimality in dynamic programming and for the limit of n-stage optimal policies to be optimal. Z. Wahrscheinlichkeitstheorie verw. Gerb. 32, 179–196, 1975.

    Article  MATH  Google Scholar 

  8. Serfozo, R. An equivalence between continuous and discrete-time markov decision processes. Oper. Res. 27, 616–620, 1979.

    Article  MathSciNet  MATH  Google Scholar 

  9. Stidham, S. Jr., and Weber, R. Monotonic and insensitive optimal policies for control of queues with undiscounted costs. Oper. Res. 87, 611–625, 1989.

    Article  MathSciNet  Google Scholar 

  10. Topkis, D. Minimizing a submodular function on a lattice. Oper. Res. 26, 305–321, 1978.

    Article  MathSciNet  MATH  Google Scholar 

  11. Wijngaard, J., and Stidham, S. Jr. Forward recursion for Markov decision processes with skip-free-to-the-right transitions, Part i: Theory and algorithms. Math. Oper. Res. 11, 295–308, 1986.

    Article  MathSciNet  MATH  Google Scholar 

Download references

Authors

Editor information

J. G. Shanthikumar Ushio Sumita

Rights and permissions

Reprints and permissions

Copyright information

© 1999 Springer Science+Business Media New York

About this chapter

Cite this chapter

Stidham, S., Weber, R.R. (1999). Monotone Optimal Policies for Left-Skip-Free Markov Decision Processes. In: Shanthikumar, J.G., Sumita, U. (eds) Applied Probability and Stochastic Processes. International Series in Operations Research & Management Science, vol 19. Springer, Boston, MA. https://doi.org/10.1007/978-1-4615-5191-1_13

Download citation

  • DOI: https://doi.org/10.1007/978-1-4615-5191-1_13

  • Publisher Name: Springer, Boston, MA

  • Print ISBN: 978-1-4613-7364-3

  • Online ISBN: 978-1-4615-5191-1

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics