Abstract
This paper proposes a vision-based Multi-user Human Computer Interaction (HCI) method for creating augmented reality user interfaces. In the HCI session, one of the users’ hands is selected as the active hand. The fingers of the active hand are employed as input devices to trigger functionalities of the application program. To share the token of interaction among the users, the HCI session is modeled as a Finite State Machine (FSM). The FSM is composed of the initial and steady states. In the initial state, the FSM identifies the active hand by tracking the hand with the maximum moving speed. Then the FSM enters the steady state to carry out the HCI session. At the end of each individual HCI cycle, the FSM polls requests from other hands for acquiring the role of active hand. If such requests are sensed, the FSM returns to the initial state to track a new active hand. Otherwise, the HCI session is continuously carried out by the current active hand. Test results show that the resultant user interface is efficient, flexible and practical for users with problems on using ordinary input devices. In a desk-top computer equipped with a 640 × 480 resolution web-camera, the HCI session can be successfully conducted when the operation distance ranges from 30 to 90 cm.
Similar content being viewed by others
References
Azuma RT (1997) A survey of augmented reality. Presence: Teleoperators and Virtual Environments 6(4):355–385
Erol A, Bebis G, Nicolescu M, Boyle RD, Twombly X (2007) Vision-based hand pose estimation: a review. Comput Vis Image Underst 108(1):52–73
Gutzeit E, Vahl M, Zhou Z, von Lukas U (2011) Skin cluster tracking and verification for hand gesture recognition. In Proceedings of ISPA 2011, pp. 241–246
Hasan MM, Mishra P (2012) Hand gesture modeling and recognition using geometric features: a review. Canadian Journal on Image Processing and Computer Vision 3(1):12–26
http://www.engadget.com/2014/10/08/kinect-for-windows-finger-tracking/, read 2015/09/28.
https://dev.windows.com/en-us/kinect/hardware, read 2015/09/28.
https://developer.leapmotion.com/documentation/java/devguide/Leap_Coordinate_Mapping.html, read 2015/08/10.
https://msdn.microsoft.com/zh-tw/library/hh438998.aspx, read 2015/08/10.
https://www.asus.com/tw/3D-Sensor/Xtion_PRO_LIVE/specifications/, read 2015/08/10.
Lee B, Chun J (2009) Manipulation of virtual objects in marker-less AR system by fingertip tracking and hand gesture recognition. In Proceedings of ICIS 2009, pp. 1110–1115
Lee D, Lee SG (2011) Vision-based finger action recognition by angle detection and contour analysis. ETRI J 33(3):415–422
Letessier J, Berard F (2004) Visual tracking of bare fingers for interactive surfaces. In Proceedings of ACM UIST’04, pp. 119–122
Manresa C, Varona J, Mas R, Perales FJ (2005) Real-time hand tracking and gesture recognition for human-computer interaction. Electronic Letters on Computer Vision and Image Analysis 5(3):96–104
Kay Elemetric Corporation (2003) Multi-Dimensional Voice Program (MDVP) operations manual, Model 5105
Raheja JL, Das K, Chaudhary A (2011) An efficient real time method of fingertip detection. In Proceedings of TIMA-2011, pp. 447–450
Rekimoto J (2002) SmartSkin: an infrastructure for freehand manipulation on interactive surfaces. In Proceedings of ACM CHI 2002, pp. 113–120
Ueng SK, Luo CM, Chang H (2012) Voice quality assessment and visualization. In Proceeding of CISIS 2012, pp. 618–623
Wilson AD (2010) Using a depth camera as a touch sensor. In Proceedings of ITS2010: Devices and Algorithms, pp. 69–72
Wu YW, Ai XY (2008) Face detection in color images using AdaBoost algorithm based on skin color information. In Proceedings of IEEE Workshop on Knowledge Discovery and Data Mining, pp. 339–342
Zhou H (2009) An OPS hand tracking algorithm based on optical flow. In Proceedings of World Congress on Software Engineering, pp. 190–194
Zhou H, Ruan Q, Chen H (2010) A new approach of hand tracking based on integrated optical flow analysis. In Proceedings of ICSP 2010, pp. 1194–1197
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Ueng, SK., Chen, GZ. Vision based multi-user human computer interaction. Multimed Tools Appl 75, 10059–10076 (2016). https://doi.org/10.1007/s11042-015-3061-z
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11042-015-3061-z