Abstract
As technology matures, human-computer interfaces have changed to meet the needs of interacting with more complex systems in user-friendly ways. Gesture-based interfaces, a type of natural user interface (NUI), allow users to use their bodies to interact with computers or virtual/augmented reality (VR/AR) and offer a more natural and intuitive user experience; however, most gesture-based commands have been developed not by considering the user, but rather what is easiest for a computer to interpret. This methodology may lead to users having increased mental effort and low perceived system usability. In two studies, we took a user-centered approach to determine natural gestures for a set of common computer actions, such as selecting an object and moving an object. In Study One, participants performed gestures for a list of actions, and in Study Two a different group of participants rated how natural they found the gestures produced in Study One compared to a set of arbitrary gestures. Overall, participants produced very similar gestures for the actions in Study One, and these gestures were rated as most natural in Study Two. Taken together, this research provides a user-centered methodology for determining natural gestures and building a vocabulary of gestures for system designers to use when developing gesture-based NUIs.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Bailey, S.K.T., Johnson, C.I., Sims, V.K.: Using natural gesture interactions leads to higher usability and presence in a computer lesson. In: Bagnara, S., Tartaglia, R., Albolino, S., Alexander, T., Fujita, Y. (eds.) IEA 2018. AISC, vol. 826, pp. 663–671. Springer, Cham (2019). https://doi.org/10.1007/978-3-319-96065-4_70
Shiratuddin, M.F., Wong, K.W.: Non-contact multi-hand gestures interaction techniques for architectural design in a virtual environment. In: 5th Conference on information Technology & Multimedia, pp. 1–6. IEEE (2011). https://doi.org/10.1109/icimu.2011.6122761
van Dam, A.: Post-WIMP user interfaces. Commun. ACM 42, 63–67 (1997). https://doi.org/10.1145/253671.253708
Roupé, M., Bosch-Sijtsema, P., Johansson, M.: Interactive navigation interface for virtual reality using the human body. Comp. Env. Urban Sys. 43, 42–50 (2014). https://doi.org/10.1016/j.compenvurbsys.2013.10.003
Ullrich, D., Diefenbach, S.: From magical experience to effortlessness: an exploration of the components of intuitive interaction. In: 6th Nordic Conference on Human-Computer Interaction: Extending Boundaries, pp. 801–804. Association for Computing Machinery, NYC (2010). https://doi.org/10.1145/1868914.1869033
Norman, D.A.: The design of everyday things. Basic Books, NYC (2002)
Steuer, J.S.: Defining virtual reality: dimensions determining telepresence. J. Comm. 42, 73–93 (1992). https://doi.org/10.1111/j.1460-2466.1992.tb00812.x
Schwartz, R.N., Plass, J.L.: Click versus drag: user-performed tasks and the enactment effect in an interactive multimedia environment. Comp. Hum. Beh. 33, 242–255 (2014). https://doi.org/10.1016/j.chb.2014.01.012
McNamara, C., Proetsch, M., Lerma, N.: Investigating low-cost virtual reality technologies in the context of an immersive maintenance training application. In: Lackey, S., Shumaker, R. (eds.) VAMR 2016. LNCS, vol. 9740, pp. 621–632. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-39907-2_59
Nielsen, M., Störring, M., Moeslund, T.B., Granum, E.: A procedure for developing intuitive and ergonomic gesture interfaces for HCI. In: Camurri, A., Volpe, G. (eds.) GW 2003. LNCS (LNAI), vol. 2915, pp. 409–420. Springer, Heidelberg (2004). https://doi.org/10.1007/978-3-540-24598-8_38
Grandhi, S.A., Joue, G., Mittelberg, I.: Understanding naturalness and intuitiveness in gesture production: Insights for touchless gestural interfaces. In: Conference on Human Factors in Computing Systems, SIGCHI, pp. 821–824. (2010). https://doi.org/10.1145/1978942.1979061
Sousa-Poza, J.F., Rohrberg, R., Mercure, A.: Effects of type of information (abstract-concrete) and field dependence on asymmetry of hand movements during speech. Percept. Mot. Skills 48, 1323–1330 (1979). https://doi.org/10.2466/pms.1979.48.3c.1323
Bailey, S.K.T., Johnson, C.I., Schroeder, B.L., Marraffino, M.D.: Using virtual reality for training maintenance procedures. In: 17th Interservice/Industry Training, Simulation and Education Conference (2017)
Schroeder, B.L., Bailey, S.K.T., Johnson, C.I., Gonzalez-Holland, E.: Presence and usability do not directly predict procedural recall in virtual reality training. In: Stephanidis, C. (ed.) HCI 2017. CCIS, vol. 714, pp. 54–61. Springer, Cham (2017). https://doi.org/10.1007/978-3-319-58753-0_9
Mauney, D., Howarth, J., Wirtanen, A., Capra, M.: Cultural similarities and differences in user-defined gestures for touchscreen user interfaces. In: CHI Human Factors in Computing Systems, pp. 4015–4020. ACM (2010). https://doi.org/10.1145/1753846.1754095
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this paper
Cite this paper
Bailey, S.K.T., Johnson, C.I. (2020). A Human-Centered Approach to Designing Gestures for Natural User Interfaces. In: Kurosu, M. (eds) Human-Computer Interaction. Multimodal and Natural Interaction. HCII 2020. Lecture Notes in Computer Science(), vol 12182. Springer, Cham. https://doi.org/10.1007/978-3-030-49062-1_1
Download citation
DOI: https://doi.org/10.1007/978-3-030-49062-1_1
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-49061-4
Online ISBN: 978-3-030-49062-1
eBook Packages: Computer ScienceComputer Science (R0)