kPose: A New Representation For Action Recognition

  • Zhuoli Zhou
  • Mingli Song
  • Luming Zhang
  • Dacheng Tao
  • Jiajun Bu
  • Chun Chen
Part of the Lecture Notes in Computer Science book series (LNCS, volume 6494)


Human action recognition is an important problem in computer vision. Most existing techniques use all the video frames for action representation, which leads to high computational cost. Different from these techniques, we present a novel action recognition approach by describing the action with a few frames of representative poses, namely kPose. Firstly, a set of pose templates corresponding to different pose classes are learned based on a newly proposed Pose-Weighted Distribution Model (PWDM). Then, a local set of kPoses describing an action are extracted by clustering the poses belonging to the action. Thirdly, a further kPose selection is carried out to remove the redundant poses among the different local sets, which leads to a global set of kPoses with the least redundancy. Finally, a sequence of kPoses is obtained to describe the action by searching the nearest kPose in the global set. And the proposed action classification is carried out by comparing the obtained pose sequence with each local set of kPose. The experimental results validate the proposed method by remarkable recognition accuracy.


Action Recognition Class Number Human Action Recognition Action Recognition Approach Speedup Robust Feature 
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.


Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.


  1. 1.
    Blank, M., Gorelick, L., Shechtman, E., Irani, M., Basri, R.: Actions as space-time shapes. In: Tenth IEEE International Conference on Computer Vision (2005)Google Scholar
  2. 2.
    Davis, J., Bobick, A.: The representation and recognition of action using temporal templates. In: IEEE Conference on Computer Vision and Pattern Recognition (1997)Google Scholar
  3. 3.
    Efros, A., Berg, A., Mori, G., Malik, J.: Recognizing action at a distance. In: Ninth IEEE International Conference on Computer Vision (2007)Google Scholar
  4. 4.
    Yamato, J., Ohya, J., Ishii, K.: Recognizing human action in time-sequential images using hidden Markov model. In: Proc. Comp. Vis. and Pattern Rec., pp. 379–385 (1992)Google Scholar
  5. 5.
    Hatun, K., Duygulu, P.: Pose sentences: A new representation for action recognition using sequence of pose words. In: 19th International Conference on Pattern Recognition (2008)Google Scholar
  6. 6.
    Mauthner, T., Roth, P., Bischof, H.: Instant action recognition. In: Salberg, A.-B., Hardeberg, J.Y., Jenssen, R. (eds.) SCIA 2009. LNCS, vol. 5575, pp. 1–10. Springer, Heidelberg (2009)CrossRefGoogle Scholar
  7. 7.
    Niebles, J., Fei-Fei, L.: A hierarchical model of shape and appearance for human action classification. In: IEEE Conference on Computer Vision and Pattern Recognition (2007)Google Scholar
  8. 8.
    Roth, P., Mauthner, T., Khan, I., Bischof, H.: Efficient Human Action Recognition by Cascaded Linear Classification (2010)Google Scholar
  9. 9.
    Schindler, K., Van Gool, L.: Action Snippets: How many frames does human action recognition require? In: IEEE Conference on Computer Vision and Pattern Recognition (2008)Google Scholar
  10. 10.
    Thurau, C., Hlavác, V.: Pose primitive based human action recognition in videos or still images. In: IEEE Conference on Computer Vision and Pattern Recognition (2008)Google Scholar
  11. 11.
    Dalai, N., Triggs, B.: Histograms of oriented gradients for human detection. In: IEEE Conference on Computer Vision and Pattern Recognition (2005)Google Scholar
  12. 12.
    Bay, H., Tuytelaars, T., Van Gool, L.: SURF: Speeded up robust features. In: Leonardis, A., Bischof, H., Pinz, A. (eds.) ECCV 2006. LNCS, vol. 3951, pp. 404–417. Springer, Heidelberg (2006)CrossRefGoogle Scholar
  13. 13.
    Laptev, I., Lindeberg, T.: SpaceCtime interest points. In: Tenth IEEE International Conference on Computer Vision (2003)Google Scholar
  14. 14.
    Harris, C., Stephens, M.: A combined corner and edge detector. In: Alvey Vision Conference, p. 50 (1988)Google Scholar
  15. 15.
    Dollar, P., Rabaud, V., Cottrell, G., Belongie, S.: Behavior recognition via sparse spatio-temporal features. In: 2nd Joint IEEE International Workshop on Visual Surveillance and Performance Evaluation of Tracking and Surveillance (2005)Google Scholar
  16. 16.
    Danafar, S., Gheissari, N.: Action recognition for surveillance applications using optic flow and SVM. In: Yagi, Y., Kang, S.B., Kweon, I.S., Zha, H. (eds.) ACCV 2007, Part II. LNCS, vol. 4844, pp. 457–466. Springer, Heidelberg (2007)CrossRefGoogle Scholar
  17. 17.
    Lu, W., Little, J.: Simultaneous tracking and action recognition using the pca-hog descriptor. In: Canadian Conference on Computer and Robot Vision (2006)Google Scholar
  18. 18.
    Patron-Perez, A., Reid, I.: A probabilistic framework for recognizing similar actions using spatio-temporal features. In: British Machine Vision ConferenceGoogle Scholar
  19. 19.
    Likas, A., Vlassis, N., et al.: The global k-means clustering algorithm. Pattern Recognition, 451–461 (2003)Google Scholar
  20. 20.
    Yu, L., Liu, H.: Feature selection for high-dimensional data: A fast correlation-based filter solution. Machine Learning-International, 856 (2003)Google Scholar
  21. 21.
    Fathi, A., Mori, G.: Action recognition by learning mid-level motion features. In: IEEE Conference on Computer Vision and Pattern Recognition (2008)Google Scholar

Copyright information

© Springer-Verlag Berlin Heidelberg 2011

Authors and Affiliations

  • Zhuoli Zhou
    • 1
  • Mingli Song
    • 1
  • Luming Zhang
    • 1
  • Dacheng Tao
    • 2
  • Jiajun Bu
    • 1
  • Chun Chen
    • 1
  1. 1.College of Computer ScienceZhejiang UniversityHangzhouChina
  2. 2.School of Computer EngineeringNanyang Technological UniversitySingapore

Personalised recommendations