Abstract
In the astronaut-space robot interaction based on hand gestures, the detection and location of hands are the premise and basis of vision-based hand gesture recognition and hand tracking. In this paper, the SSD (Single Shot Multibox Detector) which is a kind of deep learning model is utilized to detect and locate astronaut’s hands for space human-robot interaction (SHRI) based on hand gestures. First of all, in order to meet the needs of hand detection and location, an improved SSD model is designed to detect hands when they are shown as small targets in images. Then, a platform for SHRI is built and a set of hand gestures for SHRI are designed. Finally, the proposed SSD model is validated experimentally on a homemade hand gesture database for proving the superiority of this improved SSD model to small target hands detection.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Fong, T., Nourbakhsh, I.: Interaction challenges in human-robot space exploration. Interactions 12(2), 42–45 (2005)
http://www.pingwest.com/leap-motion-meets-nasa/. Accessed 31 Jan 2018
Wolf, M.T., Assad, C., Vernacchia, M.T., et al.: Gesture-based robot control with variable autonomy from the JPL BioSleeve. In: 2013 IEEE International Conference on Robotics and Automation (ICRA), pp. 1160–1165. IEEE, Karlsruhe, Germany (2013)
Liu, J.G., Luo, Y.F., Ju, Z.J.: An interactive astronaut-robot system with gesture control. Comput. Intell. Neurosci. 2016, 11 (2016)
Grzejszczak, T., Kawulok, M., Galuszka, A.: Hand landmarks detection and localization in color images. Multimed. Tools Appl. 75(23), 16363–16387 (2016)
Grzejszczak, T., Łegowski, A., Niezabitowski, M.: Application of hand detection algorithm in robot control. In: 17th International Carpathian Control Conference (ICCC). IEEE, Tatranska Lomnica, Slovakia (2016)
Raheja, J.L., Chaudhary, A., Maheshwari, S.: Hand gesture pointing location detection. Int. J. Light Electron Opt. 125(3), 993–996 (2014)
Tompson, J., Stein, M., Lecun, Y., Perlin, K.: Real-time continuous pose recovery of human hands using convolutional networks. ACM Trans. Graph. 33(5), 169 (2014)
Ge, L., Liang, H., Yuan, J., Thalmann, D.: Robust 3D hand pose estimation in single depth images: from single-view CNN to multi-view CNNs. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3593–3601. IEEE, Seattle, WA (2016)
Yamashita, T., Watasue, T.: Hand posture recognition based on bottom-up structured deep convolutional neural network with curriculum learning. In: 2014 IEEE International Conference on Image Processing (ICIP), pp. 853–857. IEEE, Paris, France (2014)
Molchanov, P., Gupta, S., Kim, K.: Hand Gesture Recognition with 3D Convolutional Neural Networks. In: CVPR 2015. IEEE, Boston, America (2015)
Liu, W., Anguelov, D., Erhan, D., Szegedy, C.: SSD: single shot multibox detector. In: 14th European Conference on Computer Vision (ECCV), pp. 21–37. IEEE, Amsterdam, The Netherlands (2016)
Fu, C.Y., Liu, W., Ranga, A., Tyagi, A., Berg, A.C.: DSSD: Deconvolutional Single Shot Detector. Computer Vision and Pattern Recognition. arXiv:1701.06659 (2017)
Cao, G.M., Xie, X.M., Yang, W.Z., et al.: Feature-Fused SSD: Fast Detection for Small Objects. Computer Vision and Pattern Recognition. arXiv:1709.05054 (2017)
Liu, J.G., Gao, Q., Liu, Z.W., Li, Y.M.: Attitude control for astronaut assisted robot in the space station. Int. J. Control Autom. Syst. 14(4), 1082–1095 (2016)
Gao, Q., Liu, J.G., Tian, T.T., Li, Y.M.: Free-flying dynamics and control of an astronaut assistant robot based on fuzzy sliding mode algorithm. Acta Astronaut. 138, 462–474 (2017)
Gattupalli, S., Ghaderi, A., Athitsos, V.: Evaluation of deep learning based pose estimation for sign language recognition. In: 9th ACM International Conference on PErvasive Technologies Related to Assistive Environments. IEEE, Greece (2016)
Gao, Q., Liu, J., Ju, Z., Li, Y., Zhang, T., Zhang, L.: Static hand gesture recognition with parallel CNNs for space human-robot interaction. In: Huang, Y., Wu, H., Liu, H., Yin, Z. (eds.) ICIRA 2017, Part I. LNCS (LNAI), vol. 10462, pp. 462–473. Springer, Cham (2017). https://doi.org/10.1007/978-3-319-65289-4_44
Acknowledgments
The authors would like to acknowledge the support from the Research Fund of China Manned Space Engineering (050102), the Key Research Program of the Chinese Academy of Sciences (Y4A3210301), the Natural Science Foundation of China under Grant No. 51775541, 51575412, 51575338 and 51575407, the EU Seventh Framework Programme (FP7)-ICT under Grant No. 611391, and the Research Project of State Key Lab of Digital Manufacturing Equipment & Technology of China under Grant No. DMETKF2017003.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2018 Springer Nature Switzerland AG
About this paper
Cite this paper
Gao, Q., Liu, J., Ju, Z., Zhang, L., Li, Y., Liu, Y. (2018). Hand Detection and Location Based on Improved SSD for Space Human-Robot Interaction. In: Chen, Z., Mendes, A., Yan, Y., Chen, S. (eds) Intelligent Robotics and Applications. ICIRA 2018. Lecture Notes in Computer Science(), vol 10984. Springer, Cham. https://doi.org/10.1007/978-3-319-97586-3_15
Download citation
DOI: https://doi.org/10.1007/978-3-319-97586-3_15
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-97585-6
Online ISBN: 978-3-319-97586-3
eBook Packages: Computer ScienceComputer Science (R0)