Skip to main content

Expressive Body Animation Pipeline for Virtual Agent

  • Conference paper
Intelligent Virtual Agents (IVA 2012)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 7502))

Included in the following conference series:

Abstract

In this paper, we present our expressive body-gestures animation synthesis model for our Embodied Conversational Agent(ECA) technology. Our implementation builds upon a full body reach model using a hybrid kinematics solution. We describe the full pipeline of our model that starts from a symbolic description of behaviors, to the construction of a set of keyframes till the generation of the whole animation enhanced with expressive qualities. Our approach offers convincing visual quality results obtained with high real-time performance.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Baerlocher, P., Boulic, R.: An inverse kinematics architecture enforcing an arbitrary number of strict priority levels. Vis. Comput. 20(6), 402–417 (2004)

    Article  Google Scholar 

  2. Boulic, R., Thalmann, D.: Combined direct and inverse kinematic control for articulated figure motion editing. Computer Graphics Forum 11(4), 189–202 (1992)

    Article  Google Scholar 

  3. Chi, D., Costa, M., Zhao, L., Badler, N.: The EMOTE model for effort and shape. In: SIGGRAPH 2000, New York, USA, pp. 173–182 (2000)

    Google Scholar 

  4. Hartmann, B., Mancini, M., Pelachaud, C.: Implementing expressive gesture synthesis for embodied conversational agents, pp. 188–199 (2006)

    Google Scholar 

  5. Hecker, C., Raabe, B., Enslow, R.W., DeWeese, J., Maynard, J., van Prooijen, K.: Real-time motion retargeting to highly varied user-created morphologies. ACM Trans. Graph 27(3), 27:1–27:11 (2008)

    Google Scholar 

  6. Heloir, A., Kipp, M.: EMBR – A Realtime Animation Engine for Interactive Embodied Agents. In: Ruttkay, Z., Kipp, M., Nijholt, A., Vilhjálmsson, H.H. (eds.) IVA 2009. LNCS, vol. 5773, pp. 393–404. Springer, Heidelberg (2009)

    Chapter  Google Scholar 

  7. Kleinsmith, A., Bianchi-Berthouze, N.: Recognizing Affective Dimensions from Body Posture. In: Paiva, A.C.R., Prada, R., Picard, R.W. (eds.) ACII 2007. LNCS, vol. 4738, pp. 48–58. Springer, Heidelberg (2007)

    Chapter  Google Scholar 

  8. Kochanek, D.H.U., Bartels, R.H.: Interpolating splines with local tension, continuity, and bias control. In: SIGGRAPH (January 1984)

    Google Scholar 

  9. Kopp, S., Krenn, B., Marsella, S.C., Marshall, A.N., Pelachaud, C., Pirker, H., Thórisson, K.R., Vilhjálmsson, H.H.: Towards a Common Framework for Multimodal Generation: The Behavior Markup Language. In: Gratch, J., Young, M., Aylett, R.S., Ballin, D., Olivier, P. (eds.) IVA 2006. LNCS (LNAI), vol. 4133, pp. 205–217. Springer, Heidelberg (2006)

    Chapter  Google Scholar 

  10. McNeill: Hand and Mind: What Gestures Reveal About Thought. The University of Chicago press, Chicago (1992)

    Google Scholar 

  11. Neff, M., Fiume, E.: Modeling tension and relaxation for computer animation. In: SCA 2002, pp. 81–88. ACM, New York (2002)

    Chapter  Google Scholar 

  12. Neff, M., Fiume, E.: Artistically based computer generation of expressive motion. In: Proceedings of the AISB, pp. 29–39 (2004)

    Google Scholar 

  13. Neff, M., Fiume, E.: AER: aesthetic exploration and refinement for expressive character animation. In: SCA 2005, pp. 161–170. ACM, New York (2005)

    Chapter  Google Scholar 

  14. Niewiadomski, R., Bevacqua, E., Le, Q.A., Pelachaud, C.: Cross-media agent platform, pp. 11–19 (2011)

    Google Scholar 

  15. Edwards, A.D.N., Harling, P.A.: Hand tension as a gesture segmentation cue. In: In Proceedings of the Progress in Gestural Interaction, pp. 75–88. MIT mimeo (1997)

    Google Scholar 

  16. Tan, N., Clavel, C., Courgeon, M., Martin, J.-C.: Postural expressions of action tendencies. In: Proceedings of the 2nd International Workshop on Social Signal Processing. ACM, New York (2010)

    Google Scholar 

  17. Thiebaux, M., Marsella, S., Marshall, A.N., Kallmann, M.: Smartbody: behavior realization for embodied conversational agents. In: Proceedings of the 7th International Joint Conference on Autonomous Agents and Multiagent Systems, AAMAS 2008, vol. 1, pp. 151–158 (2008)

    Google Scholar 

  18. Tolani, D., Goswami, A., Badler, N.I.: Real-time inverse kinematics techniques for anthropomorphic limbs. Graph. Models Image Process (2000)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2012 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Huang, J., Pelachaud, C. (2012). Expressive Body Animation Pipeline for Virtual Agent. In: Nakano, Y., Neff, M., Paiva, A., Walker, M. (eds) Intelligent Virtual Agents. IVA 2012. Lecture Notes in Computer Science(), vol 7502. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-33197-8_36

Download citation

  • DOI: https://doi.org/10.1007/978-3-642-33197-8_36

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-642-33196-1

  • Online ISBN: 978-3-642-33197-8

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics