Expressive Body Animation Pipeline for Virtual Agent

  • Jing Huang
  • Catherine Pelachaud
Part of the Lecture Notes in Computer Science book series (LNCS, volume 7502)


In this paper, we present our expressive body-gestures animation synthesis model for our Embodied Conversational Agent(ECA) technology. Our implementation builds upon a full body reach model using a hybrid kinematics solution. We describe the full pipeline of our model that starts from a symbolic description of behaviors, to the construction of a set of keyframes till the generation of the whole animation enhanced with expressive qualities. Our approach offers convincing visual quality results obtained with high real-time performance.


Inverse Kinematic Hand Gesture Virtual Agent Forward Kinematic Conversational Agent 
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.


Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.


  1. 1.
    Baerlocher, P., Boulic, R.: An inverse kinematics architecture enforcing an arbitrary number of strict priority levels. Vis. Comput. 20(6), 402–417 (2004)CrossRefGoogle Scholar
  2. 2.
    Boulic, R., Thalmann, D.: Combined direct and inverse kinematic control for articulated figure motion editing. Computer Graphics Forum 11(4), 189–202 (1992)CrossRefGoogle Scholar
  3. 3.
    Chi, D., Costa, M., Zhao, L., Badler, N.: The EMOTE model for effort and shape. In: SIGGRAPH 2000, New York, USA, pp. 173–182 (2000)Google Scholar
  4. 4.
    Hartmann, B., Mancini, M., Pelachaud, C.: Implementing expressive gesture synthesis for embodied conversational agents, pp. 188–199 (2006)Google Scholar
  5. 5.
    Hecker, C., Raabe, B., Enslow, R.W., DeWeese, J., Maynard, J., van Prooijen, K.: Real-time motion retargeting to highly varied user-created morphologies. ACM Trans. Graph 27(3), 27:1–27:11 (2008)Google Scholar
  6. 6.
    Heloir, A., Kipp, M.: EMBR – A Realtime Animation Engine for Interactive Embodied Agents. In: Ruttkay, Z., Kipp, M., Nijholt, A., Vilhjálmsson, H.H. (eds.) IVA 2009. LNCS, vol. 5773, pp. 393–404. Springer, Heidelberg (2009)CrossRefGoogle Scholar
  7. 7.
    Kleinsmith, A., Bianchi-Berthouze, N.: Recognizing Affective Dimensions from Body Posture. In: Paiva, A.C.R., Prada, R., Picard, R.W. (eds.) ACII 2007. LNCS, vol. 4738, pp. 48–58. Springer, Heidelberg (2007)CrossRefGoogle Scholar
  8. 8.
    Kochanek, D.H.U., Bartels, R.H.: Interpolating splines with local tension, continuity, and bias control. In: SIGGRAPH (January 1984)Google Scholar
  9. 9.
    Kopp, S., Krenn, B., Marsella, S.C., Marshall, A.N., Pelachaud, C., Pirker, H., Thórisson, K.R., Vilhjálmsson, H.H.: Towards a Common Framework for Multimodal Generation: The Behavior Markup Language. In: Gratch, J., Young, M., Aylett, R.S., Ballin, D., Olivier, P. (eds.) IVA 2006. LNCS (LNAI), vol. 4133, pp. 205–217. Springer, Heidelberg (2006)CrossRefGoogle Scholar
  10. 10.
    McNeill: Hand and Mind: What Gestures Reveal About Thought. The University of Chicago press, Chicago (1992)Google Scholar
  11. 11.
    Neff, M., Fiume, E.: Modeling tension and relaxation for computer animation. In: SCA 2002, pp. 81–88. ACM, New York (2002)CrossRefGoogle Scholar
  12. 12.
    Neff, M., Fiume, E.: Artistically based computer generation of expressive motion. In: Proceedings of the AISB, pp. 29–39 (2004)Google Scholar
  13. 13.
    Neff, M., Fiume, E.: AER: aesthetic exploration and refinement for expressive character animation. In: SCA 2005, pp. 161–170. ACM, New York (2005)CrossRefGoogle Scholar
  14. 14.
    Niewiadomski, R., Bevacqua, E., Le, Q.A., Pelachaud, C.: Cross-media agent platform, pp. 11–19 (2011)Google Scholar
  15. 15.
    Edwards, A.D.N., Harling, P.A.: Hand tension as a gesture segmentation cue. In: In Proceedings of the Progress in Gestural Interaction, pp. 75–88. MIT mimeo (1997)Google Scholar
  16. 16.
    Tan, N., Clavel, C., Courgeon, M., Martin, J.-C.: Postural expressions of action tendencies. In: Proceedings of the 2nd International Workshop on Social Signal Processing. ACM, New York (2010)Google Scholar
  17. 17.
    Thiebaux, M., Marsella, S., Marshall, A.N., Kallmann, M.: Smartbody: behavior realization for embodied conversational agents. In: Proceedings of the 7th International Joint Conference on Autonomous Agents and Multiagent Systems, AAMAS 2008, vol. 1, pp. 151–158 (2008)Google Scholar
  18. 18.
    Tolani, D., Goswami, A., Badler, N.I.: Real-time inverse kinematics techniques for anthropomorphic limbs. Graph. Models Image Process (2000)Google Scholar

Copyright information

© Springer-Verlag Berlin Heidelberg 2012

Authors and Affiliations

  • Jing Huang
    • 1
  • Catherine Pelachaud
    • 1
  1. 1.Telecom ParisTech - CNRSParisFrance

Personalised recommendations