Skip to main content

An Algorithm for Path Planning Based on Improved Q-Learning

  • Conference paper
  • First Online:

Part of the book series: Advances in Intelligent Systems and Computing ((AISC,volume 1107))

Abstract

The path planning is a basic research question, where an Agent is going to look for an optimal path from the beginning to the terminal point. The earliest research problem was obstacle avoidance, that is, the Agent could not collide with obstacles or fall into traps in the process of path optimization. With rapid development of the AI (artificial intelligence) technology, the path planning has become an important application field to test intelligent learning algorithms, where reinforcement learning as an active learning method has the obvious advantage in path planning compared with traditional supervised learning and unsupervised learning. As is known to all, Q-Learning is one of the most successful reinforcement learning algorithms, and it surely can help solve the path planning. Aiming at the problems of slower convergence of the existing Q-Learning algorithm, this paper introduces dynamic search factor technology and puts forward a novel ɛ-Q-Learning algorithm. Experiments show that compared with the existing Q Learning algorithm, the ɛ-Q-Learning algorithm can not only present a better optimal solution, but also significantly reduce the number of iterations to quickly complete the optimal path generation.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   129.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   169.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

References

  1. Sutton, R.S., Barto, A.G.: Reinforcement Learning: An Introduction. MIT press, Cambridge (1998)

    MATH  Google Scholar 

  2. Gao, Y., Chen, S., Lu, X.: Review of reinforcement learning. J. Autom. 30(1), 86–100 (2004)

    Google Scholar 

  3. Watkins, C.J.C.H., Dayan, P.: Q-learning. Mach. Learn. 8(3–4), 279–292 (1992)

    MATH  Google Scholar 

  4. Qiao, J., Hou, Z., Yan, X.: Application of reinforcement learning based on neural network in obstacle avoidance. J. Tsinghua Univ. Nat. Sci. Ed. 48(S2), 1747–1750 (2008)

    Google Scholar 

  5. Fakoor, M., Kosari, A., Jafarzadeh, M.: Humanoid robot path planning with fuzzy Markov decision processes. J. Appl. Res. Technol. 14(5), 300–310 (2016)

    Article  Google Scholar 

  6. Kaelbling, L.P., Littman, M.L., Moore, A.W.: Reinforcement learning: a survey. J. Artif. Intell. Res. 4, 237–285 (1996)

    Article  Google Scholar 

  7. He, D., Sun, S.: An online self-learning fuzzy navigation method for mobile robots. J. Xi’an Univ. Technol. 27(4), 325–329 (2007)

    Google Scholar 

  8. Hao, C., Fang, Z., Li, P.: Three-dimensional track planning algorithm of UAV based on Q learning. J. Shanghai Jiaotong Univ. 46(12), 1931–1935 (2012)

    Google Scholar 

  9. Liu, Z., Li, H., Liu, Q.: Reinforcement learning algorithm research. Comput. Eng. Des. 29(22), 5805–5809 (2008)

    Google Scholar 

  10. Roozegar, M., et al.: XCS-based reinforcement learning algorithm for motion planning of a spherical mobile robot. Appl. Intell. 45, 736–746 (2016)

    Article  MathSciNet  Google Scholar 

  11. Chen, C.L.: Autonomous learning and navigation control of mobile robot based on reinforcement learning. University of Science and Technology of China (2006)

    Google Scholar 

  12. Castronovo, M., Maes, F., Fonteneau, R., et al.: Learning exploration/exploitation strategies for single trajectory reinforcement learning. In: European Workshop on Reinforcement Learning, pp. 1–10 (2013)

    Google Scholar 

Download references

Acknowledgment

This work is supported by the National Natural Science Foundation of China (61773415).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Shimin Gu .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2020 Springer Nature Singapore Pte Ltd.

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Gu, S. (2020). An Algorithm for Path Planning Based on Improved Q-Learning. In: Pan, JS., Lin, JW., Liang, Y., Chu, SC. (eds) Genetic and Evolutionary Computing. ICGEC 2019. Advances in Intelligent Systems and Computing, vol 1107. Springer, Singapore. https://doi.org/10.1007/978-981-15-3308-2_3

Download citation

Publish with us

Policies and ethics