Abstract
The previous two chapters have shown that, in order to accomplish figure/ground segmentation based on the relative motion between objects in a scene, four distinct images are required to provide sufficient constraints on the four ‘unknowns’ of foreground, background, and their respective velocities or displacements. This is true irrespective of whether the relative object motion arises from independent movement of the objects or from camera motion. Essentially, a quadocular configuration is required, i.e. we need four images acquired at four equally-spaced positions x0, x1, x2, x3. If we are dealing with a translating camera rather than four distinct fixed cameras, then x i = vt i where v is the camera velocity and t i represents the time at which image i is acquired. Often, a binocular configuration is desirable. The question which arises then is: Can we effect image segmentation if we have two samples from each of two cameras? That is, can we combine spatial and temporal sampling in a unified spatio-temporal framework? Intuitively, it would seem that we can and the purpose of this chapter is to show formally how this can be achieved.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 2001 Springer Science+Business Media New York
About this chapter
Cite this chapter
Vernon, D. (2001). Articulated Binocular Vision. In: Fourier Vision. The Springer International Series in Engineering and Computer Science, vol 623. Springer, Boston, MA. https://doi.org/10.1007/978-1-4615-1413-8_5
Download citation
DOI: https://doi.org/10.1007/978-1-4615-1413-8_5
Publisher Name: Springer, Boston, MA
Print ISBN: 978-1-4613-5541-0
Online ISBN: 978-1-4615-1413-8
eBook Packages: Springer Book Archive