Abstract
Recent years have witnessed great progress in depth sensor technology, which brings huge opportunities for action recognition field. This chapter gives an overview of the recent development of the 3D action recognition approaches, and presents the motivations of the 3D action recognition features, models, and representations in this book.
Keywords
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsReferences
Shotton, J., Fitzgibbon, A., Cook, M., Sharp, T., Finocchio, M., Moore, R., Kipman, A., Blake, A.: “Real-time human pose recognition in parts from single depth images”. In: CVPR (2011)
“CMU Graphics Lab Motion Capture Database”, http://mocap.cs.cmu.edu/.
Raptis, M., Kirovski, D., Hoppe, H.: “Real-time classification of dance gestures from skeleton animation”. In: Proceedings of the 2011 ACM SIGGRAPH/Eurographics Symposium on Computer Animation—SCA ’11, p. 147. ACM Press, New York, USA (2011)
Yang X., Tian, Y.: “EigenJoints-Based Action Recognition Using Naïve-Bayes-Nearest-Neighbor”. In: CVPR 2012 HAU3D, Workshop (2012)
Ofli, F., Chaudhr, R., Kurillo, G., Vidal, R., Bajcsy, R.: “Sequence of the most informative joints (smij): A new representation for human skeletal action recognition”. J. Vis. Commun. Image Represent. 25(1), 24–38 (2013)
Xia, L., Chen, C.C., Aggarwal, J.K.: “View invariant human action recognition using histograms of 3d joints”. In: IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), 2012, IEEE, pp. 20–27 (2012)
Chaudhry, R., Ofli, F., Kurillo, G., Bajcsy, R., Vidal, R.: “Bio-inspired dynamic 3D discriminative skeletal features for human action recognition”. In: HAU3D13 (2013)
Dalal, N., Triggs, B.: “Histograms of oriented gradients for human detection”. In: CVPR, IEEE, pp. 886–893 (2005)
Lowe, D.G.: “Distinctive image features from scale-invariant keypoints”. Int. J. Comput. Vision 60(2), 91–110 (2004)
Laptev, I., Marszalek, M., Schmid, C., Rozenfeld, B.: “Learning realistic human actions from movies”. In: CVPR, pp. 1–8 (2008)
Bo, L., Lai, K., Ren, X., Fox, D.: “Object recognition with hierarchical kernel descriptors”. In: CVPR, IEEE, pp. 1729–1736, ( 2011)
Lai, K., Bo, L., Ren, X., Fox, D.: “A large-scale hierarchical multi-view RGB-D object dataset”. In: Proceedings of the IEEE International Conference on Robotics and Automation (ICRA) (2011)
Yang, X., Zhang, C., Tian, Y.: “Recognizing actions using depth motion maps-based histograms of oriented gradients”. In: ACM Multimedia (2012)
Zhao, Y., Liu, Z., Lu, Y., Cheng, H.: “Combing rgb and depth map features for human activity recognition”. In: Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC), 2012 Asia-Pacific, IEEE, pp. 1–4 (2012)
Vieira, A.W., Nascimento, E.R., Oliveira, G.L., Liu, Z., Campos, M.M.: “STOP: space-time occupancy patterns for 3D action recognition from depth map sequences”. In: 17th Iberoamerican Congress on Pattern Recognition, Buenos Aires (2012)
Tang, S., Wang, X., Lv, X., Han, T.X., Keller, J., He, Z., Skubic, M., Lao, S.: “Histogram of oriented normal vectors for object recognition with a depth sensor”. In: Computer Vision-ACCV 2012, pp. 525–538. Springer (2013)
Oreifej, O., Liu, Z.: “HON4D: histogram of oriented 4D normals for activity recognition from depth sequences”. In: CVPR (2013)
Xia, L., Aggarwal, J.K.: “Spatio-temporal depth cuboid similarity feature for activity recognition using depth camera”. In: CVPR, IEEE, pp. 2834–2841 (2013)
Dollár, P., Rabaud, V., Cottrell, G., Belongie, S.: “Behavior recognition via sparse spatio-temporal features”. In: 2nd Joint IEEE International Workshop on Visual Surveillance and Performance Evaluation of Tracking and Surveillance, IEEE, pp. 65–72 (2005)
Ni, B., Wang, G., Moulin, P.: “Rgbd-hudaact: a color-depth video database for human daily activity recognition”. In: Consumer Depth Cameras for Computer Vision, pp. 193–208. Springer (2013)
Ren, Z., Yuan, J., Zhang, Z.: “Robust hand gesture recognition based on finger-earth mover’s distance with a commodity depth camera”. In: Proceedings of the 19th ACM International Conference on Multimedia. ACM, pp. 1093–1096 (2011)
Cheng, H., Dai, Z., Liu, Z.: “Image-to-class dynamic time warping for 3d hand gesture recognition”. In: IEEE International Conference on Multimedia and Expo (ICME), 2013, IEEE, pp. 1–6 (2013)
Li, W., Zhang, Z., Liu, Z.: Expandable data-driven graphical modeling of human actions based on salient postures. IEEE Trans. Circuits Syst. Video Technol. 18(11), 1499–1510 (2008)
Wang, J., Ying, W.: “Learning maximum margin temporal warping for action recognition”. In: ICCV, pp. 87–90. ACM (2013)
Li, W., Zhang, Z., Liu, Z.: “Action recognition based on a bag of 3d points”. In: Human Communicative Behavior Analysis Workshop (In Conjunction with CVPR) (2010)
Wang, J., Liu, Z., Wu, Y., Yuan, J.: “Mining actionlet ensemble for action recognition with depth cameras”. In: CVPR (2012)
Kurakin, A., Zhang, Z., Liu, Z.: “A real-time system for dynamic hand gesture recognition with a depth sensor”. In: EUSIPCO (2012)
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
Copyright information
© 2014 The Author(s)
About this chapter
Cite this chapter
Wang, J., Liu, Z., Wu, Y. (2014). Introduction. In: Human Action Recognition with Depth Cameras. SpringerBriefs in Computer Science. Springer, Cham. https://doi.org/10.1007/978-3-319-04561-0_1
Download citation
DOI: https://doi.org/10.1007/978-3-319-04561-0_1
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-04560-3
Online ISBN: 978-3-319-04561-0
eBook Packages: Computer ScienceComputer Science (R0)