Skip to main content

Dimensional Emotion Prediction from Spontaneous Head Gestures for Interaction with Sensitive Artificial Listeners

  • Conference paper
Intelligent Virtual Agents (IVA 2010)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 6356))

Included in the following conference series:

Abstract

This paper focuses on dimensional prediction of emotions from spontaneous conversational head gestures. It maps the amount and direction of head motion, and occurrences of head nods and shakes into arousal, expectation, intensity, power and valence level of the observed subject as there has been virtually no research bearing on this topic. Preliminary experiments show that it is possible to automatically predict emotions in terms of these five dimensions (arousal, expectation, intensity, power and valence) from conversational head gestures. Dimensional and continuous emotion prediction from spontaneous head gestures has been integrated in the SEMAINE project [1] that aims to achieve sustained emotionally-colored interaction between a human user and Sensitive Artificial Listeners.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. The SEMAINE project, http://www.semaine-project.eu/

  2. Kendon, A.: Facial Expression of Emotion. In: Some functions of the face in a kissing round, pp. 117–152. Cambridge University Press, Cambridge (1990)

    Google Scholar 

  3. McClave, E.Z.: Linguistic functions of head movements in the context of speech. J. of Pragmatics 32, 855–878 (2000)

    Article  Google Scholar 

  4. Zeng, Z., et al.: A survey of affect recognition methods: Audio, visual, and spontaneous expressions. IEEE Tran. on PAMI 31, 39–58 (2009)

    Article  Google Scholar 

  5. Russell, J.A.: A circumplex model of affect. J. of Personality and Social Psychology 39, 1161–1178 (1980)

    Article  Google Scholar 

  6. Scherer, K.: Psychological models of emotion. In: The Neuropsychology of Emotion, pp. 137–162. Oxford University Press, Oxford (2000)

    Google Scholar 

  7. Fontaine, J.R., et al.: The world of emotion is not two-dimensional. Psychological Science 18, 1050–1057 (2007)

    Article  Google Scholar 

  8. Kawato, S., Ohya, J.: Real-time detection of nodding and head-shaking by directly detecting and tracking the between-eyes. In: IEEE FGR, pp. 40–45 (2000)

    Google Scholar 

  9. Kapoor, A., Picard, R.W.: A real-time head nod and shake detector. In: Workshop on Perceptive User Interfaces (2001)

    Google Scholar 

  10. Tan, W., Rong, G.: A real-time head nod and shake detector using hmms. Expert Systems with Applications 25(3), 461–466 (2003)

    Article  MathSciNet  Google Scholar 

  11. Morency, L.–P., et al.: Contextual recognition of head gestures. In: ICMI, pp. 18–24 (2005)

    Google Scholar 

  12. Glowinski, D., et al.: Technique for automatic emotion recognition by body gesture analysis. In: CVPR Workshops, pp. 1–6 (2008)

    Google Scholar 

  13. Kulic, D., Croft, E.A.: Affective state estimation for human-robot interaction. IEEE Tran. on Robotics 23(5), 991–1000 (2007)

    Article  Google Scholar 

  14. Chanel, G., et al.: Valence-arousal evaluation using physiological signals in an emotion recall paradigm. In: IEEE SMC, pp. 2662–2667 (2007)

    Google Scholar 

  15. Wollmer, M., et al.: Abandoning emotion classes - towards continuous emotion recognition with modelling of long-range dependencies. In: Interspeech, pp. 597–600 (2008)

    Google Scholar 

  16. Gunes, H., Pantic, M.: Automatic, dimensional and continuous emotion recognition. Int. Journal of Synthetic Emotions 1(1), 68–99 (2010)

    Article  Google Scholar 

  17. Douglas-Cowie, E., et al.: The Humaine database: addressing the needs of the affective computing community. In: Paiva, A.C.R., Prada, R., Picard, R.W. (eds.) ACII 2007. LNCS, vol. 4738, pp. 488–500. Springer, Heidelberg (2007)

    Chapter  Google Scholar 

  18. The SEMAINE database, http://semaine-db.eu/

  19. Schroder, M., et al.: A demonstration of audiovisual sensitive artificial listeners. In: ACII, pp. 263–264 (2009)

    Google Scholar 

  20. Viola, P., Jones, M.: Rapid object detection using a boosted cascade of simple features. In: IEEE CVPR, pp. 511–518 (2001)

    Google Scholar 

  21. Chang, C.C., Lin, C.J.: LIBSVM: a library for support vector machines (2001), http://www.csie.ntu.edu.tw/~cjlin/libsvm

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2010 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Gunes, H., Pantic, M. (2010). Dimensional Emotion Prediction from Spontaneous Head Gestures for Interaction with Sensitive Artificial Listeners. In: Allbeck, J., Badler, N., Bickmore, T., Pelachaud, C., Safonova, A. (eds) Intelligent Virtual Agents. IVA 2010. Lecture Notes in Computer Science(), vol 6356. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-15892-6_39

Download citation

  • DOI: https://doi.org/10.1007/978-3-642-15892-6_39

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-642-15891-9

  • Online ISBN: 978-3-642-15892-6

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics