Skip to main content

Part of the book series: International Series in Intelligent Technologies ((ISIT,volume 18))

Abstract

Model-based reinforcement learning methods are known to be highly efficient with respect to the number of trials required for learning optimal policies. In this article a novel fuzzy model-based reinforcement learning approach, fuzzy prioritized sweeping (F-PS), is presented. The approach is capable of learning strategies for Markov decision problems with continuous state and action spaces. The output of the algorithm are Takagi-Sugeno fuzzy systems approximating the Q-functions corresponding to the given control problems. From these Q-functions optimal control strategies can be easily derived. The effectiveness of the F-PS approach is shown by applying it to the task of selecting optimal framework signal plans in urban traffic networks. It is shown that the method outperforms existing model-based approaches.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 129.00
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 169.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 169.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  • Appl, M. (2000). Model-Based Reinforcement Learning in Continuous Environments. Ph.D. thesis. Technical University of Munich, Department of Computer Science. http://www.martinappl.de

  • Appl, M. and W. Brauer (2000). Indirect reinforcement learning with adaptive state space partitions. Proceedings of the Third European Symposium on Intelligent Techniques.

    Google Scholar 

  • Bertsekas, D. P. and J.N. Tsitsiklis (1996). Neuro-Dynamic Programming. Athena Scientific.

    Google Scholar 

  • Bingham, E. (1998). Neurofuzzy traffic signal control. Master’s thesis, Helsinki University of Technology.

    Google Scholar 

  • Davies, S. (1997). Multidimensional triangulation and interpolation for reinforcement learning. In M.C. Mozer, M.I. Jordan, and T. Petsche (Eds.), Advances in Neural Information Processing Systems, Volume 9, pp. 1005–1011. The MIT Press.

    Google Scholar 

  • Horiuchi, T., A. Fujino, O. Katai, and T. Sawaragi (1996). Fuzzy interpolation-based Q-learning with continuous states and actions. Proceedings of the Fifth IEEE International Conference on Fuzzy Systems, 594–600.

    Google Scholar 

  • Moore, A.W. and C.G. Atkeson (1993). Memory-based reinforcement learning: Converging with less data and less time. Robot Learning, 79–103.

    Google Scholar 

  • Sugeno, M. (1985). An introductory survey of fuzzy control. Information Sciences 36, 59–83.

    Article  MathSciNet  MATH  Google Scholar 

  • Sutton, R.S. and A.G. Barto (1998). Reinforcement Learning — An Introduction. MIT Press/Bradford Books, Cambridge, MA.

    Google Scholar 

  • Takagi, T. and M. Sugeno (1985). Fuzzy identification of systems and its application to modeling and control. In IEEE Transactions on Systems, Man and Cybernetics, Volume 15, pp. 116–132.

    Article  MATH  Google Scholar 

  • Thorpe, T. (1997). Vehicle Traffic Light Control Using SARSA. Ph.D. thesis. Department of Computer Science, Colorado State University.

    Google Scholar 

  • Watkins, CJ.C.H. (1989). Learning from Delayed Rewards. Ph.D. thesis, Cambridge University.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Hans-Jürgen Zimmermann Georgios Tselentis Maarten van Someren Georgios Dounias

Rights and permissions

Reprints and permissions

Copyright information

© 2002 Springer Science+Business Media New York

About this chapter

Cite this chapter

Appl, M., Brauer, W. (2002). Fuzzy Model-Based Reinforcement Learning. In: Zimmermann, HJ., Tselentis, G., van Someren, M., Dounias, G. (eds) Advances in Computational Intelligence and Learning. International Series in Intelligent Technologies, vol 18. Springer, Dordrecht. https://doi.org/10.1007/978-94-010-0324-7_15

Download citation

  • DOI: https://doi.org/10.1007/978-94-010-0324-7_15

  • Publisher Name: Springer, Dordrecht

  • Print ISBN: 978-94-010-3872-0

  • Online ISBN: 978-94-010-0324-7

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics