Advertisement

Abstract

Touchscreens and stylus input are becoming increasingly popular in modern computers. However, many programs still require keyboard input in order to take full advantage of their functionality. In order to increase the functionality of screen-based input, gestures are being adopted to input commands. However, current gesture interfaces are not sufficiently context-based, which limits their usefulness to a few applications, and they are not dynamic or interactive, which wastes time and keeps the user from feeling comfortable using them. In order to improve gesture functionality, we used off-the-shelf hardware and implemented a gesture-based system which changed gesture function based on currently running application, and based on recent gesture activity. The result of application-sensitivity was a system which is useable on many modern computers and provides useful functionality in a variety of applications. The results of sensitivity to previous gestures, however, was only minimal interactivity.

Keywords

gesture interface context-sensitive 

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. 1.
  2. 2.
    Cao, X., et al.: Modeling Human Performance of Pen Stroke Gestures. In: CHI 2007, San Jose, Calif., USA, April 28–May 3 (2007)Google Scholar
  3. 3.
    Castellucci, et al.: Graffiti vs. Unistrokes: An Empirical Comparison. In: CHI 2008, Florence, Italy (2008)Google Scholar
  4. 4.
    Grossman, et al.: Hover Widgets: Using the Tracking State to Extend the Capabilities of Pen-Operated Devices. In: CHI 2006, Montreal, Quebec, Canada, April 22–27 (2006)Google Scholar
  5. 5.
    Hinckley, et al.: Design and Analysis of Delimiters for Selection-Action Pen Gesture Phrases in Scribolie. In: CHI 2005, Portland, Oregon, USA, April 2–7 (2005)Google Scholar
  6. 6.
  7. 7.
    Liao, C., et al.: PapierCraft: A Gesture-Based Command System for Interactive Paper. In: TOCHI 2008, vol. 14(4) (January 2008)Google Scholar
  8. 8.
    Li, Y., et al.: Experimental Analysis of Mode Switching Techniques in Pen-based User Interfaces. In: SIGCHI Conf. on Human Factors in Computing Systems, April 02-07, pp. 461–470 (2005)Google Scholar
  9. 9.
  10. 10.
    Moyle, et al.: Analysing Mouse and Pen Flick Gestures. In: Proceedings of the SIGCHI-NZ Symposium On Computer-Human Interaction, New Zealand, July 11-12, pp. 19–24 (2002)Google Scholar
  11. 11.
    Opera web browser, http://www.opera.com (Cited March 27, 2010)
  12. 12.
    Roudaut, A., et al.: MicroRolls: Expanding Touch Screen Input Vocabulary by Distinguishing Rolls vs. Slides of the Thumb. In: CHI 2009, Boston, Mass, April 4–9 (2009)Google Scholar
  13. 13.
    Saund, E., et al.: Stylus Input and Editing Without Prior Selection of Mode. In: Symposium on User Interface Software and Technology, November 02–05 (2003)Google Scholar
  14. 14.
    Wobbrock, et al.: Maximizing the Guessability of Symbolic Input. In: CHI 2005, Portland, Oregon, USA, April 2-7 (2005)Google Scholar
  15. 15.
    Westerman, et al.: Writing using a Touch Sensor. US Patent Pub. No. US 2008/0042988 A1 (2007)Google Scholar
  16. 16.
    Zeleznik, R., et al.: Fluid Inking: Augmenting the Medium of Free-Form Inking with Gestures (2006)Google Scholar

Copyright information

© Springer-Verlag Berlin Heidelberg 2010

Authors and Affiliations

  • Daniel Demski
    • 1
  • Roger Lee
    • 1
  1. 1.Software Engineering and Information Technology Institute, Computer Science DepartmentCentral Michigan University 

Personalised recommendations