A Conceptual Design of Spatial Calibration for Optical See-Through Head Mounted Display Using Electroencephalographic Signal Processing on Eye Tracking

  • Azfar Tomi
  • Dayang Rohaya Awang RambliEmail author
Conference paper
Part of the Lecture Notes in Computer Science book series (LNCS, volume 10645)


One of vital issue in Optical See-Through Head Mounted Display (OST HMD) used in Augmented Reality (AR) systems is frequent (re)calibrations. OST HMD calibration that involved user interaction is time consuming. It will distract users from their application, which will reduce AR experience. Additionally, (re)calibration procedure will be prone to user errors. Nowadays, there are several approaches toward interaction-free calibration on OST HMD. In this proposed work, we propose a novel approach that uses EEG signal processing on eye movement into OST HMD calibration. By simultaneously recording eye movements through EEG during a guided eye movement paradigm, a few properties of eye movement artifacts can be useful for eye localization algorithm which can be used in interaction-free calibration for OST HMD. The proposed work is expected to enhance OST HMD calibration focusing on spatial calibration formulation in term reducing 2D projection error.


Optical See-Through Head Mounted Display Spatial calibration Electroencephalographic Eye-tracking 


  1. 1.
    Sutherland, I.: The ultimate display. In: Congress of the International Federation of Information Processing (IFIP), vol. 65, pp. 506–508 (1965)Google Scholar
  2. 2.
    Caudell, T.: Augmented reality: an application of heads-up display technology to manual manufacturing processes. In: Hawaii International Conferences on Systems Sciences, pp. 659–669 (1992)Google Scholar
  3. 3.
    Feiner, S.: Knowledge-based augmented reality. Commun. ACM 36, 53–62 (1993)CrossRefGoogle Scholar
  4. 4.
    Rolland, J.: A comparison of optical and video see-through head-mounted displays. In: SPIE Telemanipulator and Telepresence Technologies, pp. 293–307 (1994)Google Scholar
  5. 5.
    Azuma, R.: Predictive tracking for augmented reality. Ph.D. thesis. University of North Carolina, Chapel Hill, NC (1995)Google Scholar
  6. 6.
    Reiners, D.: Augmented reality for construction tasks: doorlock assembly. In: IWAR (1998)Google Scholar
  7. 7.
    Itoh, Y.: Calibration of head-mounted finger tracking to optical see-through head mounted display. In: ISMAR (2013)Google Scholar
  8. 8.
    Zhou, J.: Calibration of optical see through head mounted displays for augmented reality. Ph.D. thesis. Michigan State University (2007)Google Scholar
  9. 9.
    Klinker, G.: Augmented reality: a balancing act between high quality and real-time constraints. In: ISMR1999, pp. 325–346 (1999)Google Scholar
  10. 10.
    Tuceryan, M.: Single point active alignment method (SPAAM) for optical see-through HMD calibration for AR. In: ISAR, pp. 149–158 (2000)Google Scholar
  11. 11.
    Genc, Y.F.: Optical see-through HMD calibration: a stereo method validated with a video see-through system. In: ISAR, pp. 165–174 (2000)Google Scholar
  12. 12.
    Axholt, M.: Optical see-through head mounted display direct linear transformation calibration robustness in the presence of user alignment noise. In: Human Factors and Ergonomics Society 54th Annual Meeting, pp. 2427–2431 (2010)Google Scholar
  13. 13.
    Axholt, M.: Parameter estimation variance of the single point active alignment method in optical see-through head mounted display calibration. In: VR, pp. 27–34 (2011)Google Scholar
  14. 14.
    Genc, Y.M.: Practical solutions for calibration of optical see-through devices. In: ISMAR, pp. 169–175 (2002)Google Scholar
  15. 15.
    Owen, C.: Display-relative calibration for optical see-through head-mounted displays. In: ISMAR, pp. 70–78 (2004)Google Scholar
  16. 16.
    Itoh, Y.K.: Interaction-free calibration for optical see-through head-mounted displays based on 3D eye localization. In: 3D User Interfaces (3DUI), pp. 75–82. IEEE (2014)Google Scholar
  17. 17.
    Tsukada, A., Shino, M.: Illumination-free gaze estimation method for first-person vision wearable device. In: ICCV Workshops, pp. 2084–2091 (2011)Google Scholar
  18. 18.
    Tsukada, A., Kanade, T.: Automatic acquisition of a 3D eye model for a wearable first-person vision device. In: Eye Tracking Research and Applications (ETRA), pp. 213–216 (2012)Google Scholar
  19. 19.
    Ishiguro, Y.: Aided eyes: eye activity sensing for daily life. In: Augmented Human International Conference, p. 25. ACM (2010)Google Scholar
  20. 20.
    Schneider, E.: EyeSeeCam: an eye movement-driven head camera for the examination of natural visual exploration. In: Annuals of the New York Academy of Sciences, pp. 461–467 (2009)Google Scholar
  21. 21.
    Nitschke, C.: Corneal imaging revisited: an overview of corneal reflection analysis and applications. IPSJ Trans. Comput. Vis. Appl. 5, 1–18 (2013)CrossRefGoogle Scholar
  22. 22.
    Nitschke, C., Nakazawa, A.: Image-based eye pose and reflection analysis for advanced interaction techniques and scene understanding. In: Computer Vision and Image Media (CVIM), pp. 1–16 (2011)Google Scholar
  23. 23.
    Gao, J.: Automatic removal of eye-movement and blink artifacts from EEG signals. Brain Topogr. 23, 105–114 (2010). SpringerCrossRefGoogle Scholar
  24. 24.
    Plöchl, M.: Combining EEG and eye tracking: identification, characterization, and correction of eye movement artifacts in electroencephalographic data. In: Frontiers in Human Neuroscience, p. 278. Frontiers Media S.A. (2012)Google Scholar
  25. 25.
    Noureddin, B.: Online removal of eye movement and blink eeg artifacts using a high-speed eye tracker. Biomed. Eng. 59, 2103–2110 (2012). IEEEGoogle Scholar
  26. 26.
    Nikolaev, A.R.: Eye movement-related brain activity during perceptual and cognitive processing. In: Frontiers in Systems Neuroscience, p. 62. Frontiers Media S.A. (2014)Google Scholar
  27. 27.
    Kim, M.: Quantitative evaluation of a low-cost noninvasive hybrid interface based on EEG and eye movement. In: Neural Systems and Rehabilitation Engineering, pp. 159–168. IEEE (2015)Google Scholar
  28. 28.
    Swirski, L., Bulling, A., Dodgson, N.: Robust real-rime pupil tracking in highly off-axis images. In: Eye Tracking Research and Applications (ETRA), pp. 173–176 (2012)Google Scholar
  29. 29.
    Haji Samadi, M.R., Cooke, N.: EEG signal processing for eye tracking. In: 2014 Proceedings of the 22nd European Signal Processing Conference (EUSIPCO), pp. 2030–2034 (2014)Google Scholar
  30. 30.
    Navab, N.: An on-line evaluation system for optical see-through augmented reality. In: VR, pp. 245–246 (2004)Google Scholar
  31. 31.
    Grubert, J.: Comparative user study of two see-through calibration methods. In: VR, pp. 269–270. IEEE (2010)Google Scholar
  32. 32.
    Hansen, D.: In the eye of the beholder: a survey of models for eyes and gaze. Pattern Anal. Mach. Intell. 35, 478–500 (2010). IEEECrossRefGoogle Scholar
  33. 33.
    Kim, K.-N.: Vision-based eye-gaze tracking for human computer interface. In: Systems, Man, and Cybernetics, pp. 324–329. IEEE (1999)Google Scholar
  34. 34.
    Smith, K.: Tracking the multi person wandering visual focus of attention. In: 8th International Conference on Multimodal interfaces, pp. 265–272. ACM (2006)Google Scholar
  35. 35.
    Morimoto, C.H.: Eye gaze tracking techniques for interactive applications. Comput. Vis. Image Underst. 98, 4–24 (2005)CrossRefGoogle Scholar
  36. 36.
    Plopski, A., Itoh, Y., Nitschke, C., Kiyokawa, K., Klinker, G., Takemura, H.: Corneal-imaging calibration for optical see-through head-mounted displays. IEEE Trans. Vis. Comput. Graph. 21(4), 481–490 (2015)CrossRefGoogle Scholar

Copyright information

© Springer International Publishing AG 2017

Authors and Affiliations

  1. 1.Department of Computer Information SciencesUniversiti Teknologi PETRONASBandar Seri IskandarMalaysia

Personalised recommendations