Abstract
Slam (Simultaneous Localization and Mapping) is a passive system and in traditional slam algorithm robot’s path is not considered when improving localization uncertainty. However, improving localization accuracy while autonomously exploring unknown environments needs to get abundant feature points and make enough loop closures. To that end we propose a reinforcement learning based active slam framework that can add path planning to existing slam algorithms. In this framework a reinforcement learning agent plans the path while slam is processing. We have tested our framework in simulation environments built in Unreal engine with unrealcv plugin and we have got excellent results.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Mur-Artal, R., Montiel, J.M.M., TardĂ³s, J.D.: ORB-SLAM: a versatile and accurate monocular SLAM system. IEEE Trans. Robot. 31(5), 1147–1163 (2017)
Engel, J., Schöps, T., Cremers, D.: LSD-SLAM: large-scale direct monocular SLAM. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) ECCV 2014. LNCS, vol. 8690, pp. 834–849. Springer, Cham (2014). https://doi.org/10.1007/978-3-319-10605-2_54
Smart, W.D., Kaelbling, L.P.: Effective reinforcement learning for mobile robots. In: Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, pp. 3404–3410 (2008)
Gupta, S., Davidson, J., Levine, S., Sukthankar, R., Malik, J.: Cognitive Mapping and Planning for Visual Navigation (2017)
Mnih, V., Kavukcuoglu, K., Silver, D., et al.: Playing atari with deep reinforcement learning. Comput. Sci. (2013)
Lillicrap, T.P., Hunt, J.J., Pritzel, A., et al.: Continuous control with deep reinforcement learning. Comput. Sci. 8(6), A187 (2015)
Mnih, V., Badia, A.P., Mirza, M., et al.: Asynchronous methods for deep reinforcement learning (2016)
Wu, Y., Mansimov, E., Liao, S., et al.: Scalable trust-region method for deep reinforcement learning using Kronecker-factored approximation (2017)
Valencia, R., Miro, J.V., Dissanayake, G., Andrade-Cetto, J.: Active pose SLAM. In: 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 1885–1891 (2012)
Lauri, M., Ritala, R.: Planning for robotic exploration based on forward simulation. Robot. Auton. Syst. 83(C), 15–31 (2016)
Qiu, W., et al.: UnrealCV: virtual worlds for computer vision, pp. 1221–1224. ACM (2017)
Bai, S., Wang, J., Doherty, K., Englot, B.: Inference-enabled information-theoretic exploration of continuous action spaces. In: Bicchi, A., Burgard, W. (eds.) Robotics Research. Springer Proceedings in Advanced Robotics, vol. 3, pp. 419–433. Springer, Cham (2018). https://doi.org/10.1007/978-3-319-60916-4_24
Acknowledgement
This work is supported by National Key R&D Program of China under with Grant No. 2017YFB1302302.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2019 Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Li, B., Sun, F., Liu, H., Fang, B. (2019). Path Planning for Active V-Slam Based on Reinforcement Learning. In: Sun, F., Liu, H., Hu, D. (eds) Cognitive Systems and Signal Processing. ICCSIP 2018. Communications in Computer and Information Science, vol 1006. Springer, Singapore. https://doi.org/10.1007/978-981-13-7986-4_42
Download citation
DOI: https://doi.org/10.1007/978-981-13-7986-4_42
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-13-7985-7
Online ISBN: 978-981-13-7986-4
eBook Packages: Computer ScienceComputer Science (R0)