Abstract
People gain a physical sense of the environment surrounding them via visual information from the eyes; but from these observations alone we are not capable of determining the exact dimensions of the environment. Field robots use sensors such as Global Position System (GPS) or Inertial Measurement Units (IMU) to make accurate estimations of the position and attitude, but these instruments cannot provide accurate relative measurements with respect to a specific site without prior surveying. Computer vision techniques, i. e. using cameras as sensors; offer vision information that gives a physical sense of a robotic platform pose with respect to some targeted site, and are capable of making accurate estimates of relative positions and attitudes.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
S. Saripalli, J.F. Montgomery, and G.S. Sukhatme, “Vision-based autonomous landing of an unmanned aerial vehicle,” presented at IEEE International Conference on Robotics and Automation, 2002.
R.O.V. Shakernia, C.S. Sharp, Y. Ma, S. Sastry, “Multiple view motion estimation and control for landing an unmanned aerial vehicle,” presented at IEEE International Conference on Robotics and Automation, Washington, DC, 2002.
Z.F. Yang and W.H. Tsai, “Using parallel line information for vision-based landmark location estimation and an application to automatic helicopter landing,” Robotics and Computer-Integrated Manufacturing, vol. 14, pp. 297–306, 1998.
T. Amidi and K. Fujita, “Visual odometer for autonomous helicopter flight,” Robotics and Autonomous Systems, vol. 28, pp. 185–193, 1999.
T. Amidi and J.R. Miller, “Vision-Based Autonomous Helicopter Research at Carnegie Mellon Robotics Institute 1991–1997,” presented at American Helicopter Society International Conference, Heli, Japan, 1998.
A. Tsai, P. Gibbens, and R. Stone, “Terminal Phase Vision-Based Target Recognition and 3D Pose Estimation for a Tail-Sitter, Vertical Takeoff and Landing Unmanned Air Vehicle,” presented at Pacific-rim Symposium on Image and Video Technology, Hsin-Chu, Taiwan, 2006.
B. L. Stevens and F.L. Lewis, Aircraft Control and Simulation, 2 ed. Hoboken, New Jersey: John Wiley & Sons, 2003.
M. Hu, “Visual Pattern Recognition by Moment Invariants,” IRE Transactions on Information Theory, 1962.
R. Sivaramakrishna and N. S. Shashidharf, “Hu’s moment invariants: how invariant are they under skew and perspective transformations?” presented at IEEE WESCANEX 97: Communications, Power and Computing. 1997.
R.M. Haralick and L.G. Shapiro, Computer and Robot Vision, vol. II: Addison-Wesley, 1993.
R.H. Stone, “Configuration Design of a Canard Configuration Tail Sitter Unmanned Air Vehicle Using Multidisciplinary Optimization.” N.S.W.: PhD Thesis, University of Sydney, 1999.
“SPAN Technology System Characteristics and Performance,” NovAtel Inc. 2005.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2008 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Tsai, A., Gibbens, P., Hugh Stone, R. (2008). Visual Position Estimation for Automatic Landing of a Tail-Sitter Vertical Takeoff and Landing Unmanned Air Vehicle. In: Billingsley, J., Bradbeer, R. (eds) Mechatronics and Machine Vision in Practice. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-74027-8_14
Download citation
DOI: https://doi.org/10.1007/978-3-540-74027-8_14
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-74026-1
Online ISBN: 978-3-540-74027-8
eBook Packages: EngineeringEngineering (R0)