Skip to main content

Abstraction of State-Action Space Utilizing Properties of the Body and Environment

Analysis of Policy Obtained by Three-Dimensional Snake-Like Robot Operating on Rubble

  • Chapter
  • First Online:
Book cover Novel Applications of Intelligent Systems

Part of the book series: Studies in Computational Intelligence ((SCI,volume 586))

  • 575 Accesses

Abstract

We focused on the autonomous control of a three-dimensional snake-like robot that moves on rubble. To realize an autonomous controller, we employed reinforcement learning. However, applying reinforcement learning in a conventional framework to a robot with many degrees of freedom and moving in a complex environment is difficult. There are three problems: state explosion, lack of reproducibility, and lack of generality. To solve these problems, we previously proposed abstracting the state-action space by utilizing the universal properties of the body and environment. The effectiveness of the proposed framework was demonstrated experimentally. Unfortunately, analysis of the obtained policy was lacking. In the present study, we analyzed the obtained policy (i.e., Q-values of Q-learning) to determine the mechanism for abstraction of the state-action space and confirmed that the three problems were solved.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 109.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. R.S. Sutton, Reinforcement Learning: An Introduction (The MIT Press, Massachusetts, 1988)

    Google Scholar 

  2. L.P. Kealbling, M.L. Littman, Reinforcement learning: a survey. J. Artif. Intell. Res. 4, 237–285 (1996)

    Google Scholar 

  3. C.J.H. Watkins, P. Dayan, Technical note Q-learning. Mach. Learn. 8, 279–292 (1992)

    MATH  Google Scholar 

  4. M. Asada, Y. Katoh, M. Ogino, K. Hosoda, A humanoid approaches to the goal: reinforcement learning based on rhythmic walking parameters, in Proceedings of the International Symposium of RoboCup2003 (2003, CD-ROM)

    Google Scholar 

  5. H. Kimura, T. Yamashita, S. Kobayashi, Reinforcement learning of walking behavior for a four-legged robot, in Proceedings of 40th IEEE Conference on Decision and Control (2001), pp. 411–416

    Google Scholar 

  6. K. Doya, H. Kimura, M. Kawato, Neural mechanism of learning and control. IEEE Control Syst. Mag. 21(4), 42–44 (2001)

    Article  Google Scholar 

  7. C. Anderson, Z. Hong, Reinforcement learning with modular neural networks for control, in Proceedings of NNACIP’94: IEEE International Workshop on Neural Networks Applied to Control and Image Processing (1994)

    Google Scholar 

  8. W. Sun, X. Wang, Y. Cheng, Reinforcement learning method for continuous state space based on dynamic neural network, in Proceedings of the 7th World Congress on Intelligent Control and Automation (2008), pp. 750–754

    Google Scholar 

  9. D. Gu, H. Hu, Reinforcement learning of fuzzy logic controller for quadruped walking robots. Presented at Proceedings of 15th IFAC World Congress, Barcelona, Spain, 21–26 July 2002

    Google Scholar 

  10. H.R. Berenji, Fuzzy learning for generalization of reinforcement learning, in Proceedings of the Fifth IEEE International Conference on Fuzzy Systems, vol. 3 (1996), pp. 2208–2214

    Google Scholar 

  11. W.M. Hinojosa, S. Nefti, U. Kaymak, Systems control with generalized probabilistic fuzzy-reinforcement learning. IEEE Trans. Fuzzy Syst. 19(1), 51–64 (2011)

    Article  Google Scholar 

  12. A. Likas, Reinforcement learning using the stochastic fuzzy min-max neural network. Neural Process. Lett. 13, 213–220 (2001)

    Article  MATH  Google Scholar 

  13. T. Inamura, M. Inada, H. Inoue, Integration model of learning mechanism and dialogue strategy based on stochastic experience representation using Bayesian network, in Proceedings of the 2000 IEEE International Workshop on Robot and Human Interactive Communication (2000), pp. 247–252

    Google Scholar 

  14. S. Ushio, M. Svinin, K. Ueda, S. Hosoe, An evolutionary approach to decentralized reinforcement learning for walking robots, in Proceedings of the 6th International Symposium on Artificial Life and Robotics (2001), pp. 176–179

    Google Scholar 

  15. K. Ito, F. Matsuno, Reinforcement learning for redundant robot: solution of state explosion problem in real world, in Proceedings of ROBIO’05 Workshop on Biomimetic Robotics and Biomimetic Control (2005), pp. 36–41

    Google Scholar 

  16. R. Pfeifer, Understand Intell, new edn. (The MIT Press, Massachusetts, 2001)

    Google Scholar 

  17. J.J. Gibson, The Ecological Approach to Visual Perception (Lawrence Erlbaum Associates, Hillsdale, NJ, 1987)

    Google Scholar 

  18. K. Ito, A. Takayama, T. Kobayashi, Hardware design of autonomous snake-like robot for reinforcement learning based on environment: discussion of versatility on different tasks, in Proceedings of the 2009 IEEE/RSJ International Conference on Intelligent Robots and Systems (2009), pp. 2622–2627

    Google Scholar 

  19. K. Ito, Y. Fukumori, A. Takayama, Autonomous control of real snake-like robot using reinforcement learning: abstraction of state-action space using properties of real world, in Proceedings of the International Conference on Intelligent Sensors, Sensor Networks and Information Processing (2007), pp. 389–394

    Google Scholar 

  20. K. Ito, S. Kuroe, T. Kobayashi, Abstraction of state-action space utilizing properties of the body and the environment: application to a three-dimensional snake-like robot that operates on rubble, in Proceedings of IEEE International Conference on Intelligent Systems (2012), pp. 114–120

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Kazuyuki Ito .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2016 Springer International Publishing Switzerland

About this chapter

Cite this chapter

Ito, K., Kuroe, S., Kobayashi, T. (2016). Abstraction of State-Action Space Utilizing Properties of the Body and Environment. In: Hadjiski, M., Kasabov, N., Filev, D., Jotsov, V. (eds) Novel Applications of Intelligent Systems. Studies in Computational Intelligence, vol 586. Springer, Cham. https://doi.org/10.1007/978-3-319-14194-7_3

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-14194-7_3

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-14193-0

  • Online ISBN: 978-3-319-14194-7

  • eBook Packages: EngineeringEngineering (R0)

Publish with us

Policies and ethics