Abstract
In facial emotion recognition, facial features of the person have to be detected first. For this, active appearance models (AAMs) are used in this work. The main drawback of AAM is that it cannot generalize to unseen faces. To overcome this drawback, the training images are pre-processed before using them for model construction in this work. For automatic initialization of the model, the output of Viola–Jones face detector is used. In literature, principal component analysis (PCA) is used to capture the main variations of the training data in AAM. As the variance in PCA is fixed, they produce large models which are difficult to optimize. In this work, fisher face method is used in which PCA is first used to reduce the dimensions of the feature space so that the resulting within class scatter matrix is non-singular and then apply the linear discriminant analysis (LDA) to reduce the dimensions still further. The experimental results on unseen and seen faces show that fitting accuracy is better and takes less time to fit when compared with existing AAMs.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Mehrabian, A.: Communication without words. Psychol. Today 2(4), 53–56 (1968)
Ekman, P., Huang, T., Sejnowski, T., Hager, J.: Final Report to NSF of the Planning Workshop on Facial Expression Understanding (1992)
Ekman, P., Friesen, W.V.: Manual for the Facial Action Coding System. Consulting Psychologists Press, Palo Alto (1977)
MPEG Video and SNHC: Text of ISO/IEC FDIS 14 496-3: Audio. In: Atlantic City MPEG Mtg (1998)
Moin, A. et al.: Weighted-PCA based multimodal medical image fusion in contourlet domain. In: Proceedings of the International Congress on Information and Communication Technology 2016, pp. 597–605. Springer, Singapore
Krishn, A. et al.: PCA based medical image fusion in ridgelet domain. In: Proceedings of the 3rd International Conference on Frontiers of Intelligent Computing: Theory and Applications (FICTA) 2014 2015, pp. 475–482. Springer, Cham
Bhateja, V. et al.: Medical image fusion in wavelet and ridgelet domains: a comparative evaluation. Int. J. Rough Sets Data Anal. (IJRSDA) 2(2), 78–91 (2015)
Belhumeur, P.N., Hespanha, J.P., Kriegman, D.J.: Eigenfaces vs. fisherfaces: recognition using class specific linear projection. IEEE Trans. Pattern Anal. Mach. Intell. 19(7), 711–720 (1997)
Lucey, P., Cohn, J.F., Saragih, J., Matthews, I., Ambadar, Z.: The extended Cohn-Kanade database: a complete facial expression database for both facial action units and emotion detection. In: IEEE CVPR4HB (2010)
ACTIVE APPEARANCE. MODELS: Theory, Extensions & Cases, 2nd edn. Mikkel Bille Stegmann. LYNGBY (2000)
Baker, S., Matthews, I.: Lucas-kanade 20 years on: a unifying framework. Int. J. Comput. Vis. 56(1), 221–255 (2004)
Matthews, Iain, Baker, Simon: Active appearance models revisited. Int. J. Comput. Vis. 60(2), 135–164 (2004)
Stegmann, M.B., Ersboll, B.K., Larsen, R.: FAME - a flexible appearance modeling environment. IEEE Trans. on Med. Imaging 22(10), 1319–1331 (2003)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2019 Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Videla, L.S., Rao, M.R.N., Anand, D., Vankayalapati, H.D., Razia, S. (2019). Deformable Facial Fitting Using Active Appearance Model for Emotion Recognition. In: Satapathy, S., Bhateja, V., Das, S. (eds) Smart Intelligent Computing and Applications . Smart Innovation, Systems and Technologies, vol 104. Springer, Singapore. https://doi.org/10.1007/978-981-13-1921-1_13
Download citation
DOI: https://doi.org/10.1007/978-981-13-1921-1_13
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-13-1920-4
Online ISBN: 978-981-13-1921-1
eBook Packages: EngineeringEngineering (R0)