Advertisement

A Study on Acquisition Method of Nonverbal Cues for Intelligent Agents: A Case Study on Facial Expression Analysis

Part of the Advances in Intelligent Systems and Computing book series (AISC, volume 194)

Abstract

The paper proposes a method to acquire nonverbal cues for an intelligent agent that can be used in communications with human partner. A framework to extract the information that affords a key to understand intension, mental state, affect, or signs evoked by a person through nonverbal communication between the person and an intelligent agent. The aim of the framework is to make intelligent agents to be able to extract the nonverbal cues through trial-and-error style learning. A limited functionality of the framework is implemented and applied to a facial expression analysis as a case study. The main advantage of the method is that it would acquire the cues as some time-series information inherent to particular facial expressions. Thus the agent Result in a preliminary experiment shows that this implemented method have potential to extract nonverbal cues concerning with fake and real smiles.

Keywords

nonverbal cue facial expression analysis intelligent agent time-series data 

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. 1.
    Argyle, M.: Bodily communication, 2nd edn., Methuen (1988)Google Scholar
  2. 2.
    Hall, J.A.: Nonverbal communication. In: Sander, D., Scherer, K.R. (eds.) The Oxford Companion to Emotion and The Affective Sciences, pp. 282–283. Oxford University Press, Oxford (2009)Google Scholar
  3. 3.
    Hakura, J., Kurematsu, M., Fujita, H.: Facial Expression Invariants for Estimating Mental States of Person. In: Proceedings of the 2009 Conference on New Trends in Software Methodologies, Tools and Techniques: Proceedings of the Eighth Somet 2009, pp. 518–530 (2009)Google Scholar
  4. 4.
    Fasel, B., Luettin, J.: Automatic facial expression analysis: a survey. Pattern Recognition 36, 259–276 (2003)MATHCrossRefGoogle Scholar
  5. 5.
    Pantic, M., Rothkrantz, L.J.M.: Toward an affect-sensitive multimodal human-computer interaction. Proceedings of the IEEE 91, 1370–1390 (2003)CrossRefGoogle Scholar
  6. 6.
    Tian, Y., Kanade, T., Cohn, J.F.: Recognizing action units for facial expression analysis. IEEE Transactions on Pattern Analysis and Machine Intelligence 23, 97–115 (2001)CrossRefGoogle Scholar
  7. 7.
    Oki Electronic Industry, Face Sensing Engine Version 4 Application Development Guide: FSE Feature Point Detector Edition (FP) (2008)Google Scholar
  8. 8.
    Cootes, T.F., Edwards, G.J., Taylor, C.J.: Active Appearance Models. IEEE Transactions on Pattern Analysis and Machine Intelligence 23(6), 681–685 (2001)CrossRefGoogle Scholar
  9. 9.
    Kohonen, T.: Self-organization and associative memory. Springer, Berlin (1989)CrossRefGoogle Scholar

Copyright information

© Springer-Verlag Berlin Heidelberg 2013

Authors and Affiliations

  1. 1.Graduate School of Software and Information ScienceIwate Prefectural UniversityTakizawaJapan

Personalised recommendations