Gesture Recognition from Two-Person Interactions Using Ensemble Decision Tree

  • Sriparna SahaEmail author
  • Biswarup Ganguly
  • Amit Konar
Conference paper
Part of the Advances in Intelligent Systems and Computing book series (AISC, volume 518)


The evolution of depth sensors has furnished a new horizon for human–computer interaction. An efficient two-person interaction detection system is proposed for an improved human−computer interaction using Kinect sensor. This device is able to identify twenty body joint coordinates in 3D space among which sixteen joints are selected and those have been adapted with certain weights to form four average points. The direction cosines of these four average points are evaluated followed by the angles made by x, y and z axes, respectively, i.e., twelve angles have been constructed for each frame. For recognition purpose, ensemble of tree classifiers with bagging mechanism is used. This novel work is widely acceptable for various gesture-based computer appliances and yields a recognition rate of 87.15%.


Human–computer interaction Kinect sensor Direction cosines Ensemble decision tree 



The research work is supported by the University Grants Commission, India, University with Potential for Excellence Program (Phase II) in Cognitive Science, Jadavpur University and University Grants Commission (UGC) for providing fellowship to the first author.


  1. 1.
    S. S. Rautaray and A. Agrawal, “Vision based hand gesture recognition for human computer interaction: a survey,” Artif. Intell. Rev., vol. 43, no. 1, pp. 1–54, 2015.Google Scholar
  2. 2.
    M. R. Andersen, T. Jensen, P. Lisouski, A. K. Mortensen, M. K. Hansen, T. Gregersen, and P. Ahrendt, “Kinect depth sensor evaluation for computer vision applications,” Tech. Rep. Electron. Comput. Eng., vol. 1, no. 6, 2015.Google Scholar
  3. 3.
    T. T. Dao, H. Tannous, P. Pouletaut, D. Gamet, D. Istrate, and M. C. H. B. Tho, “Interactive and Connected Rehabilitation Systems for E-Health,” IRBM, 2016.Google Scholar
  4. 4.
    S. Park and J. K. Aggarwal, “Event semantics in two-person interactions,” in Pattern Recognition, 2004. ICPR 2004. Proceedings of the 17th International Conference on, 2004, vol. 4, pp. 227–230.Google Scholar
  5. 5.
    A. Yao, J. Gall, G. Fanelli, and L. J. Van Gool, “Does Human Action Recognition Benefit from Pose Estimation?.,” in BMVC, 2011, vol. 3, p. 6.Google Scholar
  6. 6.
    K. Yun, J. Honorio, D. Chattopadhyay, T. L. Berg, and D. Samaras, “Two-person interaction detection using body-pose features and multiple instance learning,” in Computer Vision and Pattern Recognition Workshops (CVPRW), 2012 IEEE Computer Society Conference on, 2012, pp. 28–35.Google Scholar
  7. 7.
    S. Saha, A. Konar, and R. Janarthanan, “Two Person Interaction Detection Using Kinect Sensor,” in Facets of Uncertainties and Applications, Springer, 2015, pp. 167–176.Google Scholar
  8. 8.
    T. G. Dietterich, “An experimental comparison of three methods for constructing ensembles of decision trees: Bagging, boosting, and randomization,” Mach. Learn., vol. 40, no. 2, pp. 139–157, 2000.Google Scholar
  9. 9.
    T. G. Dietterich, “Ensemble methods in machine learning,” in Multiple classifier systems, Springer, 2000, pp. 1–15.Google Scholar
  10. 10.
    R. Drillis, R. Contini, and M. Maurice Bluestein, “Body segment parameters 1,” Artif. Limbs, p. 44, 1966.Google Scholar

Copyright information

© Springer Nature Singapore Pte Ltd. 2018

Authors and Affiliations

  1. 1.Electronics & Tele-Communication Engineering DepartmentJadavpur UniversityKolkataIndia

Personalised recommendations