Skip to main content

Automatic Generation of Non-Verbal Facial Expressions from Speech

  • Conference paper
Advances in Modelling, Animation and Rendering

Abstract

Speech synchronized facial animation that controls only the movement of the mouth is typically perceived as wooden and unnatural. We propose a method to generate additional facial expressions such as movement of the head, the eyes, and the eyebrows fully automatically from the input speech signal. This is achieved by extracting prosodic parameters such as pitch flow and power spectrum from the speech signal and using them to control facial animation parameters in accordance to results from paralinguistic research.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 169.00
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 219.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 219.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. I. Albrecht, J. Haber, and H.-P. Seidel. Speech Synchronization for Physics-based Facial Animation. In Proc. WSCG 2002 pages 9–16, 2002.

    Google Scholar 

  2. D. Barr. Trouble in mind: Paralinguistic indices of effort and uncertainty in communication. InOralité et Gestualité. Actes du colloque ORAGE 2001 pages 597–600, 2001.

    Google Scholar 

  3. M. Brand. Voice Puppetry. In Proc. SIGGRAPH ‘89 pages 21–28, 1999.

    Google Scholar 

  4. J. Cassell and M. Stone. Living Hand to Mouth: Psychological Theories about Speech and Gesture in Interactive Dialogue Systems. Technical Report FS-99–03, AAAI Fall Symposium on Psychological Models of Communication in Collaborative Systems, 1999.

    Google Scholar 

  5. C. Cavé, I. Guatella, R. Bertrand, S. Santi, F. Harlay, and R. Espesser. About the relationship between eyebrow movements and f0 variations. In Proc. ICSLP ‘86 1996.

    Google Scholar 

  6. N. Chovil. Discourse-Oriented Facial Displays in Conversation. Research on Language and Social Interaction 25:163–194, 1991.

    Article  Google Scholar 

  7. M. M. Cohen and D. W. Massaro. Modeling Coarticulation in Synthetic Visual Speech. In N. M. Magnenat-Thalmann and D. Thalmann, editors, Models and Techniques in Computer Animation pages 139–156. 1993.

    Google Scholar 

  8. J. Cosnier. Les gestes de la question. In Kerbrat-Orecchioni, editor, La Question pages 163–171. Presses Universitaires de Lyon, 1991.

    Google Scholar 

  9. P. Ekman. About brows: emotional and conversational signals. In M. v. Cranach, K. Foppa, W. Lepenies, and D. Ploog, editors, Human Ethology: Claims and limits of a new discipline: contributions to the Colloquium. pages 169–248. 1979.

    Google Scholar 

  10. J. Haber, K. Kähler, I. Albrecht, H. Yamauchi, and H.-P. Seidel. Face to Face: From Real Humans to Realistic Facial Animation. InProc. Israel-Korea Binational Conference on Geometrical Modeling and Computer Graphics pages 73–82, Oct. 2001.

    Google Scholar 

  11. D. R. Hill, A. Pearce, and B. Wyvill. Animating Speech: An Automated Approach using Speech Synthesised by Rules.The Visual Computer3(5):277–289 Mar. 1988.

    Article  Google Scholar 

  12. D. House, J. Beskow, and B. Granström. Timing and Interaction of Visual Cues for Prominence in Audiovisual Speech Perception. In Proc. Eurospeech 2001 2001.

    Google Scholar 

  13. H. H. S. Ip and C. S. Chan. Script-Based Facial Gesture and Speech Animation Using a NURBS Based Face Model. Computers & Graphics 20(6):881–891, Nov. 1996.

    Article  Google Scholar 

  14. T. Johnstone and K. Scherer. The Effects of Emotions on Voice Quality. In Proc. XIVth International Congress of Phonetic Sciences, 2000. in press.

    Google Scholar 

  15. K. Kähler, J. Haber, and H.-P. Seidel. Geometry-based Muscle Modeling for Facial Animation. In Proc. Graphics Interface 2001 pages 37–46, June 2001.

    Google Scholar 

  16. P. Kalra, A. Mangili, N. Magnenat-Thalmann, and D. Thalmann. SMILE: A Multilayered Facial Animation System. In Proc. IFIP WG 5.10, Tokyo, Japan, pages 189–198, 1991.

    Google Scholar 

  17. J. P. Lewis and F. I. Parke. Automated Lip-Synch and Speech Synthesis for Character Animation. In Proc. Graphics Interface ‘87 pages 143–147, Apr. 1987.

    Google Scholar 

  18. M. Lundeberg and J. Beskow. Developing a 3D-agent for the AUGUST dialogue system. In Proc. Audio-Visual Speech Processing (AVSP) ‘89 1999.

    Google Scholar 

  19. A. Paeschke, M. Kienast, and W. Sendlmeier. FO-Contours in Emotional Speech. InProc. International Congress of Phonetic Sciences ‘89pages 929–931, 1999.

    Google Scholar 

  20. F. I. Parke and K. Waters, editors. Computer Facial Animation. A K Peters, Wellesley, MA, 1996.

    Google Scholar 

  21. A. Pearce, B. Wyvill, G. Wyvill, and D. R. Hill. Speech and Expression: A Computer Solution to Face Animation. In Proc. Graphics Interface ‘86 pages 136–140, May 1986.

    Google Scholar 

  22. C. Pelachaud, N. Badler, and M. Steedman. Generating Facial Expressions for Speech. Cognitive Science 20(1):1–46,1996.

    Article  Google Scholar 

  23. K. Sjölander. The Snack Sound Toolkit. http://www.speech.kth.se/snack/, 1997–2001.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2002 Springer-Verlag London

About this paper

Cite this paper

Albrecht, I., Haber, J., Seidel, HP. (2002). Automatic Generation of Non-Verbal Facial Expressions from Speech. In: Vince, J., Earnshaw, R. (eds) Advances in Modelling, Animation and Rendering. Springer, London. https://doi.org/10.1007/978-1-4471-0103-1_18

Download citation

  • DOI: https://doi.org/10.1007/978-1-4471-0103-1_18

  • Publisher Name: Springer, London

  • Print ISBN: 978-1-4471-1118-4

  • Online ISBN: 978-1-4471-0103-1

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics