Abstract
In this paper, we applied a reverse correlation approach to study the features that humans use to categorize facial expressions. The well-known portrait of Mona Lisa was used as the base image to investigate the features differentiating happy and sad expressions. The base image was blended with sinusoidal noise masks to create the stimulus. Observers were required to view each image and categorized it as happy or sad. Analysis of responses using superimposed classification images revealed both locations and identity of information required to represent each certain expression. To further investigate the results, a neural network based classifier was developed to identify the expression of the superimposed images from the machine learning perspective, which reveals that the pattern which humans perceive the expression is acknowledged by machines.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
Ahumada Jr., A.J., Lovell, J.: Stimulus features in signal detection. Journal of the Acoustical Society of America 49, 1751–1756 (1971)
Chauvin, A., Worsley, K., Schyns, P., Arguin, M., Gosselin, F.: Accurate statistical tests for smooth classification images. Journal of Vision 5, 655–667 (2005)
De Valois, R.L., De Valois, K.K.: Spatial vision. Oxford University Press, NY (1990)
Gosselin, F., Schyns, P.: Bubbles: A technique to reveal the use of information in recognition tasks. Vision Research 41, 2261–2271 (2001)
DeAngelis, G., Ohzawa, I., Freeman, D.: Neuronal mechanisms underlying stereopsis: how do simple cells in the visual cortex encode binocular disparity? Perception 24, 3–31 (1995)
Humphreys, K., Gosselin, F., Schyns, P.G., Johnson, M.H.: Using ‘Bubbles’ with babies: A new technique for investigating the informational basis of infant perception. Infant Behavior and Development 29, 471–475 (2006)
Leonid, L., Kontsevich, C.W.: Tyler “What makes Mona Lisa smile?”. Vision Research 44, 1493–1498 (2004)
Magini, M.C., Biederman, I.: Making the ineffable explicit: estimating the information employed for face classifications. Cognitive Science 28, 209–226 (2004)
Maurer, D., Le Grand, R., Mondloch, C.J.: The many faces of configural processing. Trends in Cognitive Sciences 6, 255–260 (2002)
Murray, R.F., Bennett, P.J., Sekuler, A.B.: Optimal methods for calculating classification images: Weighted sum. Journal of Vision 2, 79–104 (2002)
Ekman, P., Friesen, W.: Facial Action Coding System: A Technique for the Measurement of Facial Movement. Consulting Psychologists Press, Palo Alto (1978)
Zhang, X., Yin, L., Gerhardstein, P., Hipp, D.: Expression-driven salient features: bubble-based facial expression study by human and machine. In: IEEE International Conference on Multimedia and Expo (ICME 2010) (July 2010)
Yin, L., Chen, X., Sun, Y., Worm, T., Reale, M.: A high-resolution 3D dynamic facial expression database. In: IEEE International Conference on Automatic Face and Gesture Recognition (FG 2008) (2008)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2014 Springer International Publishing Switzerland
About this paper
Cite this paper
Zhang, X., Yin, L., Hipp, D., Gerhardstein, P. (2014). Evaluation of Perceptual Biases in Facial Expression Recognition by Humans and Machines. In: Bebis, G., et al. Advances in Visual Computing. ISVC 2014. Lecture Notes in Computer Science, vol 8888. Springer, Cham. https://doi.org/10.1007/978-3-319-14364-4_78
Download citation
DOI: https://doi.org/10.1007/978-3-319-14364-4_78
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-14363-7
Online ISBN: 978-3-319-14364-4
eBook Packages: Computer ScienceComputer Science (R0)