Abstract
In this paper, we propose a collaborative technique for face orientation estimation in smart camera networks. The proposed spatiotemporal feature fusion analysis is based on active collaboration between the cameras in data fusion and decision making using features extracted by each camera. First, a head strip mapping method is proposed based on a Markov model and a Viterbi-like algorithm to estimate the relative angular differences to the face between the cameras. Then, given synchronized face sequences from several camera nodes, the proposed technique determines the orientation and the angular motion of the face using two features, namely the hair-face ratio and the head optical flow. These features yield an estimate of the face orientation and the angular velocity through simple analysis such as Discrete Fourier Transform (DFT) and Least Squares (LS), respectively. Spatiotemporal feature fusion is implemented via key frame detection in each camera, a forward-backward probabilistic model, and a spatiotemporal validation scheme. The key frames are obtained when a camera node detects a frontal face view and are exchanged between the cameras so that local face orientation estimates can be adjusted to maintain a high confidence level. The forward-backward probabilistic model aims to mitigate error propagation in time. Finally, a spatiotemporal validation scheme is applied for spatial outlier removal and temporal smoothing. A face view is interpolated from the mapped head strips, from which snapshots at the desired view angles can be generated. The proposed technique does not require camera locations to be known in prior, and hence is applicable to vision networks deployed casually without localization.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Bai, X.-M., Yin, B.-C., Shi, Q., Sun, Y.-F.: Face recognition using extended fisherface with 3d morphable model. In: Proc. of the ICMLC, vol. 7, pp. 4481–4486 (2005)
Hu, Y., Jiang, D., Yan, S., Zhang, L., Zhang, H.: Automatic 3d reconstruction for face recognition. In: IEEE Conference on FGR, IEEE Computer Society Press, Los Alamitos (2004)
Kurata, D., Nankaku, Y., Tokuda, K., Kitamura, T., Ghahramani, Z.: Face recognition based on separable lattice hmms. In: Proc. of ICASSP (2006)
Liu, C., Wechsler, H.: Enhanced fisher linear discriminant models for face recognition. In: Proc. of ICPR, vol. 2, pp. 1368–1372 (1998)
Turk, M., Portland, A.: Eigenfaces for recognition. J. Cognition Nueralscience 3(1), 71–86 (1991)
Chang, C., Aghajan, H.: A LQR spatiotemporal fusion technique for face profile collection in smart camera surveillance. In: Proc. of ACIVS (2007)
Uchida, N., Shibahara, T., Aoki, T.: Face recognition using passive stereo vision. In: Proc. of ICIP (2005)
Bouguet, J.-Y.: Pyramidal Implementation of the Lucas Kanade Feature Tracker Description of the algorithm. In: Intel Corporation, Microprocessor Research Labs (2000)
Intel Corporation: Open Source Computer Vision Library 1.0 (2006)
Author information
Authors and Affiliations
Editor information
Rights and permissions
Copyright information
© 2007 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Chang, CC., Aghajan, H. (2007). Spatiotemporal Fusion Framework for Multi-camera Face Orientation Analysis. In: Blanc-Talon, J., Philips, W., Popescu, D., Scheunders, P. (eds) Advanced Concepts for Intelligent Vision Systems. ACIVS 2007. Lecture Notes in Computer Science, vol 4678. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-74607-2_14
Download citation
DOI: https://doi.org/10.1007/978-3-540-74607-2_14
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-74606-5
Online ISBN: 978-3-540-74607-2
eBook Packages: Computer ScienceComputer Science (R0)