Efficient Detection of Consecutive Facial Expression Apices Using Biologically Based Log-Normal Filters

  • Zakia Hammal
Part of the Lecture Notes in Computer Science book series (LNCS, volume 6938)

Abstract

The automatic extraction of the most relevant information in a video sequence made of continuous affective states is an important challenge for efficient human-machine interaction systems. In this paper a method is proposed to solve this problem based on two steps: first, the automatic segmentation of consecutive emotional segments based on the response of a set of Log-Normal filters; secondly, the automatic detection of the facial expression apices based on the estimation of the global face energy inside each emotional segment independently of the undergoing facial expression. The proposed method is fully automatic and independent from any reference image such as the neutral at the beginning of the sequence. The proposed method is the first contribution for the summary of the most important affective information present in a video sequence independently of the undergoing facial expressions. The robustness and efficiency of the proposed method to different data acquisition and facial differences has been evaluated on a large set of data (157 video sequences) taken from two benchmark databases (Hammal-Caplier and MMI databases) [1, 2] and from 20 recorded video sequences of multiple facial expressions (between three to seven facial expressions per sequence) in order to include more challenging image data in which expressions are not neatly packaged in neutral-expression-neutral.

Keywords

Facial expressions Apices Video affect summary Log-Normal filters 

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. 1.
    Hammal, Z., Couvreur, L., Caplier, A., Rombaut, M.: Facial expressions classification: A new approach based on transferable belief model. International Journal of Approximate Reasoning 46(3), 542–567 (2007)CrossRefGoogle Scholar
  2. 2.
    Pantic, M., Valstar, M.F., Rademaker, R., Maat, L.: Web-based database for facial expression analysis. In: Proc. IEEE Int. Conf. ICME 2005, Amsterdam, The Netherlands (July 2005)Google Scholar
  3. 3.
    Pantic, P., Rothkrantz, L.J.M.: Automatic Analysis of Facial Expressions: The State of the Art. IEEE Trans. PAMI. 22(12), 1424–1445 (2000)CrossRefGoogle Scholar
  4. 4.
    Zeng, Z., Pantic, M., Roisman, G.I., Huang, T.S.: A Survey of Affect Recognition Methods: Audio, Visual, and Spontaneous Expressions. IEEE Trans. PAMI 31(1), 39–58 (2009)CrossRefGoogle Scholar
  5. 5.
    Essa, I.A., Pentland, A.P.: Coding, Analysis, Interpretation, and Recognition of Facial Expressions. IEEE Trans. PAMI. 19(7), 757–763 (1997)CrossRefGoogle Scholar
  6. 6.
    Otsuka, T., Ohya, J.: Recognizing multiple persons’ facial expressions using HMM based on automatic extraction of significant frames from image sequences. In: Proc. IEEE Int. Conf. Image Processing, vol. 2, pp. 546–549 (1997)Google Scholar
  7. 7.
    Cohen, I., Cozman, F.G., Sebe, N., Cirelo, M.C., Huang, T.S.: Learning Bayesian network classifiers for facial expression recognition using both labeled and unlabeled data. In: Proc. IEEE CVPR (2003)Google Scholar
  8. 8.
    Hammal, Z., Massot, C.: Holistic and Feature-Based Information Towards Dynamic Multi-Expressions Recognition. In: Proc. Int. Conf. VISIGRAPP, Anger, France, (May 17-21, 2010)Google Scholar
  9. 9.
    Massot, C., Herault, J.: Model of Frequency Analysis in the Visual Cortex and the Shape from Texture Problem. International Journal of Computer Vision 76(2) (2008)Google Scholar
  10. 10.
    Hammal, Z., Eveno, N., Caplier, A., Coulon, P.-Y.: Parametric models for facial features segmentation. Signal processing 86, 399–413 (2006)CrossRefMATHGoogle Scholar
  11. 11.
    Ekman, P., Friesen, W.V.: The facial action coding system (FACS): A technique for the measurement of facial action. Consulting Psychologists Press, Palo Alto (1978)Google Scholar
  12. 12.
    Valstar, M.F., Pantic, M.: Induced Disgust, Happiness and Surprise: an Addition to the MMI Facial Expression Database. In: Proc. Int. Conf. LREC, Malta (May 2010)Google Scholar
  13. 13.
    Tian, Y.L., Kanade, T., Cohn, J.F.: Facial expression analysis. In: Li, S.Z., Jain, A.K. (eds.) Handboock of Face Recognition, pp. 247–276. Springer, NY (2005)CrossRefGoogle Scholar
  14. 14.
    Beaudot, W.: Le traitement neuronal de l’information dans la rétine des vertébrés: Un creuset d’idées pour la vision artificielle, Thèse de Doctorat INPG, TIRF, Grenoble France (1994)Google Scholar

Copyright information

© Springer-Verlag Berlin Heidelberg 2011

Authors and Affiliations

  • Zakia Hammal
    • 1
  1. 1.The Robotics institute of the Carnegie Mellon University CMUPittsburghUSA

Personalised recommendations