Skip to main content

Vision-Based Tracking for Mobile Augmented Reality

  • Chapter
Book cover Multimedia Services in Intelligent Environments

Part of the book series: Studies in Computational Intelligence ((SCI,volume 120))

Summary

Augmented Reality Systems (ARS) attempt to enhance humans’ perception of their indoors and outdoors working and living environments and understanding of tasks that they need to carry out. The enhancement is effected by complementing the human senses with virtual input. For example, when the human visual sense is enhanced, an ARS allows virtual objects to be superimposed on a real world by projecting the virtual objects onto real objects. This provides the human user of the ARS with additional information that he/she could not perceive with his/her senses. In order to receive the virtual input and sense the world around them augmented with real time computer-generated features, users of an ARS need to wear special equipment, such as head-mounted devices or wearable computing gears. Tracking technologies are very important in an ARS and, in fact, constitute one challenging research and development topic. Tracking technologies involve both hardware and software issues, but in this chapter we focus on tracking computation. Tracking computation refers to the problem of estimating the position and orientation of the ARS user’s viewpoint, assuming the user to carry a wearable camera. Tracking computation is crucial in order to display the composed images properly and maintain correct registration of real and virtual worlds. This tracking problem has recently become a highly active area of research in ARS. Indeed, in recent years, several approaches to vision-based tracking using a wearable camera have been proposed, that can be classified into two main categories, namely “marker-based tracking” and “marker-less tracking.” In this chapter, we provide a concise introduction to vision-based tracking for mobile ARS and present an overview of the most popular approaches recently developed in this research area. We also present several practical examples illustrating how to conceive and to evaluate such systems.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 129.00
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 169.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 169.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. D.G. Lowe. Fitting Parameterized Three-Dimensional Models to Images. IEEE Transactions on Pattern Analysis and Machine Intelligence, 13, pp. 441-450, 1991.

    Article  Google Scholar 

  2. R.M. Haralick. Pose Estimation from Corresponding Point Data. IEEE Trans-actions on Systems, Man, and Cybernetics, 19(6), pp. 1426-1446, 1989.

    Article  Google Scholar 

  3. D.F. DeMenthon and L.S. Davis. Model-based Object Pose in 25 Lines of Code. International Journal of Computer Vision, 15(1-2), pp. 123-141, 1995.

    Article  Google Scholar 

  4. C.P. Lu, G. Hager, and E. Mjolsness. Fast and Globally Convergent Pose Esti-mation from Video Images. IEEE Transaction on Pattern Analysis and Machine Intelligence, 22(6), pp. 610-622, June 2000.

    Article  Google Scholar 

  5. Y. Hung, P. Yeh, and D. Harwood. Passive ranging to known planar point sets. In Proceedings of IEEE International Conference on Robotics and Automation, 1, pp. 80-85, St. Louis, Missouri, 1985.

    Google Scholar 

  6. L. Quan and Z. Lan. Linear n-Point Camera Pose Determination. IEEE Transac-tions. Pattern Analysis and Machine Intelligence, 21(7), pp. 774-780, July 1999.

    Article  Google Scholar 

  7. Y. Cho and U. Neumann. Multi-Ring Color Fiducial Systems for Scalable Fidu-cial Tracking Augmented Reality. In Proceedings of the Virtual Reality Annual International Symposium (VRAIS’98). pp. 212, Washington, DC, USA, 1998.

    Google Scholar 

  8. L. Naimark and E. Foxlin. Circular data matrix fiducial system and robust image processing for a wearable vision-inertial self-tracker. In Proceedings of ACM/IEEE International Symposium on Mixed and Augmented Reality (ISMAR 2002). pp. 27-36, Darmstadt, Germany, 2002.

    Chapter  Google Scholar 

  9. J. Rekimoto. Matrix: A Realtime Object Identification and Registration Method for Augmented Reality. In Proceedings of the Third Asian Pacific Computer and Human Interaction (APCHI’98). pp. 63-68, Washington DC, USA, 1998.

    Google Scholar 

  10. J. Rekimoto and Y. Ayatsuka. Cybercode: Designing Augmented Reality Envi-ronments with Visual Tags. In Proceedings of DARE 2000 on Designing Aug-mented Reality Environments, pp. 1-10, Elsinore, Denmark, 2000.

    Chapter  Google Scholar 

  11. H. Kato and M. Billinghurst. Marker Tracking and HMD Calibration for a Video-based Augmented Reality Conferencing System. In Proceedings of the 2nd ACM/IEEE International Workshop on Augmented Reality (IWAR’99), pp. 85-92, Washington DC, USA, 1999.

    Google Scholar 

  12. X. Zhang, S. Fronz, and N. Navab. Visual Marker Detection and Decoding in AR Systems: A Comparative Study. In Proceedings of the ACM/IEEE Inter-national Symposium on Mixed and Augmented Reality (ISMAR 2002), pp. 97, Washington, DC, USA, 2002.

    Google Scholar 

  13. C.B. Owen, X. Fan, and P. Middlin. What is the Best Fiducial? In Augmented Reality Toolkit, The First IEEE International Workshop. IEEE, 2002.

    Google Scholar 

  14. M. Fiala. Artag, A Fiducial Marker System Using Digital Techniques. In Pro-ceedings of the 2005 IEEE International Conference on Computer Vision and Pattern Recognition (CVPR’05). 2, pp. 590-596, Washington, DC, USA, 2005.

    Google Scholar 

  15. J. Canny. A computational approach to edge detection. IEEE Transactions on Pattern Analysis and Machine Intelligence, 8(6), pp. 679-698, 1986.

    Article  Google Scholar 

  16. Z. Zhang. Flexible Camera Calibration by Viewing a Plane from Unknown Ori-entations. In Proceedings of the International Conference on Computer Vision, 1, pp. 666, Corfu, Greece, 1999.

    Google Scholar 

  17. J.Y. Didier. Contributions à la déxtérité d’un système de réalité augmentée mo-bile appliquée à la maintenance industrielle. In PhD Thesis, Université d’ Évry, France, 2005.

    Google Scholar 

  18. G. Welch and G. Bishop. An introduction to the kalman filter. Technical Report No. TR 95-041, Department of Computer Science, University of North Carolina, USA, 2004.

    Google Scholar 

  19. H. Wuest, F. Vial, and D. Stricker. Adaptive Line Tracking with Multiple Hypotheses for Augmented Reality. In Proceedings of ACM/IEEE Interna-tional Symposium on Mixed and Augmented Reality (ISMAR 2005), pp. 62-69, Vienna, Austria, October 2005.

    Google Scholar 

  20. T. Drummond and R. Cipolla. Real-Time Visual Tracking of Complex Struc-tures, IEEE Transactions on Pattern Analysis and Machine Intelligence, 24(7), pp. 932-946, July 2002.

    Article  Google Scholar 

  21. Y. Yoon, A. Kosaka, J.B. Park, and A.C. Kak. A New Approach to the Use of Edge Extremities for Model-based Object Tracking. In Proceedings of the 2005 IEEE International Conference on Robotics and Automation (ICRA 2005). pp. 1883-1889, Barcelonna, Spain, April 2005.

    Google Scholar 

  22. A.I. Comport, E. Marchand, M. Pressigout, and F. Chaumette. Real-Time Markerless Tracking for Augmented Reality: The Virtual Visual Servoing Frame-work. IEEE Transactions on Visualization and Computer Graphics, 12(6), 615-628, July/August 2006.

    Google Scholar 

  23. V. Lepetit, L. Vacchetti, D. Thalmann, and P. Fua. Fully Automated and Stable Registration for Augmented Reality Applications. In Proceedings of ACM/IEEE International Symposium on Mixed and Augmented Reality (ISMAR 2003), p. 93, Tokyo, Japan, 2003.

    Google Scholar 

  24. L. Vacchetti, V. Lepetit, and P. Fua. Combining Edge and Texture Information for Real-Time Accurate 3D Camera Tracking. In Proceedings of ACM/IEEE International Symposium on Mixed and Augmented Reality (ISMAR 2004), pp. 48-57, Arlington, VA, November 2004.

    Google Scholar 

  25. M. Pressigout and E. Marchand. Real-Time 3D Model-based Tracking: Com-bining Edge and Texture Information. In Proceedings of the 2006 IEEE Inter-national Conference on Robotics and Automation (ICRA 06), pp. 2726-2731, Orlando, Florida, May 2006.

    Google Scholar 

  26. F. Ababsa and M. Mallem. Robust Line Tracking Using a Particle Filter for Camera Pose Estimation. In Proceedings of ACM Symposium on Virtual Re-ality Software and Technology (VRST 2006), pp. 207-211, Limassol, Cyprus, November 2006.

    Google Scholar 

  27. A. Beutelspacher and U. Rosenbaum. Projective Geometry: From Foundations to Applications, Cambridge University Press, Cambridge, 1998.

    Google Scholar 

  28. N.J. Gordon. A Hybrid Bootstrap Filter for Target Tracking in Clutter. IEEE Transactions on Aerospace and Electronic Systems, 33, pp. 353-358, 1997.

    Article  Google Scholar 

  29. A. Doucet, N. de Freitas, and N. Gordon. Sequential Monte Carlo Methods in Practice. Springer, Berlin Heidelberg New York, 2001.

    MATH  Google Scholar 

  30. M. Pupilli and A. Calway. Real-Time Camera Tracking Using a Particle Fil-ter. In Proceedings of the British Machine Vision Conference (BMVC 2005), pp. 519-528, Oxford, UK, September 2005.

    Google Scholar 

  31. A. State, G. Hirota, D.T. Chen, W.F. Garrett, and M.A. Livingston, Superior Augmented Reality Registration by Integrating Landmark Tracking and Mag-netic Tracking. In Proceedings of the 23rd Annual Conference on Computer Graphics and Interactive Techniques (SIGGRAPH’96), pp. 429-438, New York, NY, USA, 1996.

    Google Scholar 

  32. T. Auer and A. Pinz. Building a Hybrid Tracking System: Integration of Op- tical and Magnetic Tracking. In Proceedings of the 2nd ACM/IEEE International Workshop on Augmented Reality (IWAR’99), pp. 13-19, Washington, DC, USA, 1999.

    Google Scholar 

  33. S. You, U. Neumann, and R. Azuma. Hybrid Inertial and Vision Tracking for Augmented Reality Registration. In Proceedings of IEEE International Confer-ence on Virtual Reality (VR 99), pp. 260-267, 1999.

    Google Scholar 

  34. J. Chen and A. Pinz. Structure and Motion by Fusion of Inertial and Vision-based Tracking. In OCG. Proceedings of the 28th OAGM/AAPR Conference. Digital Imaging in Media and Education (W. Burger and J. Scharinger, eds.), pp. 55-62, 2004.

    Google Scholar 

  35. L. Chai, W. Hoff, and T. Vincent. Three-Dimensional Motion and Structure Estimation Using Inertial Sensors and Computer Vision for Augmented Reality. Presence: Teleoperators and Virtual Environments, pp. 474-492, 2002.

    Google Scholar 

  36. E. Foxlin and L. Naimark. VIS-Tracker: A Wearable Vision-Inertial Self-Tracker. In Proceedings of the IEEE Conference on Virtual Reality (VR 2003), pp. 193, Los Angeles, CA, USA. March 2003.

    Google Scholar 

  37. F. Ababsa and M. Mallem. Hybrid 3D Camera Pose Estimation Using Particle Filter Sensor Fusion. Advanced Robotics. International Journal of the Robotics Society of Japan (RSJ), pp. 21, 165-181, 2007.

    Google Scholar 

  38. S. Feiner, B. MacIntyre, T. Hollerer, and A. Webster. A Touring Machine: Pro-totyping 3D Mobile Augmented Reality Systems for Exploring the Urban Envi-ronment. In Proceedings of the 1st IEEE International Symposium on Wearable Computers (ISWC 97). pp. 74, Washington, DC, USA, 1997.

    Google Scholar 

  39. W. Piekarski and B. Thomas. Tinmith-metro: New Outdoor Techniques for Cre-ating City Models with an Augmented Reality Wearable Computer. In Proceed-ings of the 5th International Symposium on Wearable Computers, pp. 31-38, Zurich, 2001.

    Google Scholar 

  40. G.R. King, W. Piekarski, and B.H. Thomas. ARVino - Outdoor Augmented Reality Visualisation of Viticulture GIS Data. In Proceedings of the ACM/IEEE International Symposium on Mixed and Augmented Reality (ISMAR 2005), pp. 52-55, Washington, DC, USA, 2005.

    Google Scholar 

  41. G. Reitmayr and T. Drummond. Going out: Robust Model-based Tracking for Outdoor Augmented Reality. In Proceedings of the ACM/IEEE International Symposium on Mixed and Augmented Reality (ISMAR 2006), pp. 109-118, Santa Barbara, CA, USA, October 2006.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2008 Springer-Verlag Berlin Heidelberg

About this chapter

Cite this chapter

Ababsa, F., Maidi, M., Didier, JY., Mallem, M. (2008). Vision-Based Tracking for Mobile Augmented Reality. In: Tsihrintzis, G.A., Jain, L.C. (eds) Multimedia Services in Intelligent Environments. Studies in Computational Intelligence, vol 120. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-78502-6_12

Download citation

  • DOI: https://doi.org/10.1007/978-3-540-78502-6_12

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-78491-3

  • Online ISBN: 978-3-540-78502-6

  • eBook Packages: EngineeringEngineering (R0)

Publish with us

Policies and ethics