Abstract
Expressivity of hand movements is much greater than what current interaction techniques enable in touch-screen input. Especially for collaboration, hands are used to interact but also to express intentions, point to the physical space in which collaboration takes place, and communicate meaningful actions to collaborators. Various types of interaction are enabled by multi-touch surfaces (singe and both hands, single and multiple fingers, etc.), and standard approaches to tactile interactive systems usually fail in handling such complexity of expresion. The diversity of multi-touch input also makes designing multi-touch gestures a difficult task. We believe that one cause for this design challenge is our limited understanding of variability in multi-touch gesture articulation, which affects users’ opportunities to use gestures effectively in current multi-touch interfaces. A better understanding of multi-touch gesture variability can also lead to more robust design to support different users’ gesture preferences. In this chapter we present our results on multi-touch gesture variability. We are mainly concerned with understanding variability in multi-touch gestures articulation from a pure user-centric perspective. We present a comprehensive investigation on how users vary their gestures in multi-touch gestures even under unconstrained articulation conditions. We conducted two experiments from which we collected 6669 multi-touch gestures from 46 participants. We performed a qualitative analysis of user gesture variability to derive a taxonomy for users’ multi-touch gestures that complements other existing taxonomies. We also provide a comprehensive analysis on the strategies employed by users to create different gesture articulation variations for the same gesture type.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
McNeill D (1992) Hand and mind: What gestures reveal about thought. University of Chicago Press, Psychology/cognitive science
Kita S (2009) Cross-cultural variation of speech-accompanying gesture: a review. In: Language and cognitive processes, pp 145–167
Morris MR, Wobbrock JO, Wilson AD (2010) Understanding users’ preferences for surface gestures. Proceedings of GI 2010. ON, Canada, Toronto, pp 261–268
Wobbrock JO, Morris MR, Wilson AD (2009) User-defined gestures for surface computing. Proceeding of CHI 2009. ACM, New York, NY, USA, pp 1083–1092
Ruiz J, Li Y, Lank E (2011) User-defined motion gestures for mobile interaction. Proceeding of CHI 2011 ACM. NY, USA, New York, pp 197–206
Morris MR (2012) Web on the wall: insights from a multimodal interaction elicitation study. Proceedings of ITS 2012. ACM, New York, NY, USA, pp 95–104
Vatavu RD (2013) A comparative study of user-defined handheld vs. freehand gestures for home entertainment environments. J Ambient Intell Smart Environ:187–211
Vatavu R-D User-defined gestures for free-hand TV control. In: Proceedings of EuroiTV 2012. ACM, New York, NY, USA, pp 45–48
Bailly G, Müller J, Lecolinet E (2012) Design and evaluation of finger-count interaction: combining multitouch gestures and menus. Proceedings of IJHCS. Springer, Duluth, MN, USA, pp 673–689
Kin K, Agrawala M, DeRose T (2009) Determining the benefits of direct-touch, bimanual, and multifinger input on a multitouch workstation. Proceedings of GI 2009. ON, Canada, Toronto, pp 119–124
Wu M, Balakrishnan R Multi-finger and whole hand gestural interaction techniques for multi-user tabletop displays. In: Proceedings of UIST 2003. ACM, New York, NY, USA, pp 193–202
Anthony L, Vatavu R-D, Wobbrock JO (2013) Understanding the consistency of users’ pen and finger stroke gesture articulation. Proceedings of GI 2013. ON, Canada, Toronto, pp 87–94
Rekik Y, Grisoni L, Roussel N (2013) Towards many gestures to one command: a user study for tabletops. Proceedings of INTERACT. Springer, Berlin, Heidelber, pp 246–263
Hennecke F, Berwein F, Butz A (2011) Optical pressure sensing for tangible user interfaces. Proceedings of ITS 2011. ACM, New York, NY, USA, pp 45–48
Harrison C, Schwarz, J, Hudson SE TapSense: enhancing finger interaction on touch surfaces. In: Proceedings of UIST 2011. ACM, New York, NY, USA, pp 627–636
Lü H, Li Y (2012) Gesture coder: a tool for programming multi-touch gestures by demonstration. Proceedings of CHI 2012. ACM, New York, NY, USA, pp 2875–2884
Caramiaux B, Bevilacqua F, Tanaka A (2013) Beyond recognition: using gesture variation for continuous interaction. Proceedings of CHI 2013. ACM, New York, NY, USA, pp 2109–2118
Rekik Y, Vatavu R-D, Grisoni L (2014) Match-up & conquer: a two-step technique for recognizing unconstrained bimanual and multi-finger touch input. Proceedings of AVI. ACM, New York, NY, USA, pp 201–208
Hinrichs U, Carpendale S (2011) Gestures in the wild: studying multi-touch gesture sequences on interactive tabletop exhibits. Proceedings of CHI 2011. ACM, New York, NY, USA, pp 3023–3032
Freeman D, Benko H, Morris MR, Wigdor D (2009) ShadowGuides: visualizations for in-situ learning of multi-touch and whole-hand gestures. Proceedings of ITS 2009. ACM, New York, NY, USA, pp 165–172
Oh U, Findlater L (2013) The challenges and potential of end-user gesture customization. Proceedings of CHI 2013. ACM, New York, NY, USA, pp 1129–1138
Wu M, Shen C, Ryall K, Forlines C, Balakrishnan R (2006) Gesture registration, relaxation, and reuse for multi-point direct-touch surfaces. Proceedings of TABLETOP 2006. IEEE, Washington, DC, USA, pp 185–192
Ringel M, Ryall K, Shen C, Forlines C, Vernier F (2004) Release, relocate, reorient, resize: fluid techniques for document sharing on multi-user interactive tables. Proceedings of CHI EA 2004. ACM, New York, NY, USA, pp 1441–1444
Morris MR, Huang A, Paepcke A, Winograd T (2006) Cooperative gestures: multi-user gestural interactions for co-located groupware. Proceedings of CHI 2006. ACM, New York, NY, USA, pp 1201–1210
Bau O, Mackay WE (2008) OctoPocus: a dynamic guide for learning gesture-based command sets. Proceedings of UIST 2008. ACM, New York, NY, USA, pp 37–46
Kin K, Hartmann B, Agrawala M (2011) Two-handed marking menus for multitouch devices. Proceedings of TOCHI 2011. ACM, New York, NY, USA, pp 1073–0516
Kray C, Nesbitt D, Dawson J, Rohs M (2010) User-defined gestures for connecting mobile phones, public displays, and tabletops. Proceedings of mobile HCI 2010. ACM, New York, NY, USA, pp 239–248
Rekik Y, Vatavu R-D, Grisoni L (2014) Perceived difficulty understanding users’ of multi-touch gesture articulation. In: Proceedings of ICMI. ACM, New York, NY, USA
Huang TS, Pavlovic VI (1995) Hand gesture modeling, analysis, and synthesis. In: Workshop on automatic face and gesture recognition. IEEE, pp 73–79
Thieffry S (1981) Hand gestures. In: Tubiana R (ed) The hand. University of Chicago Press
Vatavu R-D, Anthony L, Wobbrock JO (2012) Gestures as point clouds: a $P recognizer for user interface prototypes. Proceedings of ICMI 2012. ACM, New York, NY, USA, pp 273–280
Tu H, Ren X, Zhai S (2012) A comparative evaluation of finger and pen stroke gestures. Proceedings of CHI 2012. ACM, New York, NY, USA, pp 1287–1296
Acknowledgements
R.-D. Vatavu acknowledges support from the project PN-II-RUTE-2014-4-1187 financed by UEFISCDI, Romania.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2016 Springer International Publishing Switzerland
About this chapter
Cite this chapter
Rekik, Y., Vatavu, RD., Grisoni, L. (2016). Spontaneous Gesture Production Patterns on Multi-touch Interactive Surfaces. In: Anslow, C., Campos, P., Jorge, J. (eds) Collaboration Meets Interactive Spaces. Springer, Cham. https://doi.org/10.1007/978-3-319-45853-3_3
Download citation
DOI: https://doi.org/10.1007/978-3-319-45853-3_3
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-45852-6
Online ISBN: 978-3-319-45853-3
eBook Packages: Computer ScienceComputer Science (R0)