Automatic Partial Face Alignment in NIR Video Sequences

  • Jimei Yang
  • Shengcai Liao
  • Stan Z. Li
Part of the Lecture Notes in Computer Science book series (LNCS, volume 5558)

Abstract

Face recognition with partial face images is an important problem in face biometrics. The necessity can arise in not so constrained environments such as in surveillance video, or portal video as provided in Multiple Biometrics Grand Challenge (MBGC). Face alignment with partial face images is a key step toward this challenging problem.

In this paper, we present a method for partial face alignment based on scale invariant feature transform (SIFT). We first train a reference model using holistic faces, in which the anchor points and their corresponding descriptor subspaces are learned from initial SIFT keypoints and the relationships between the anchor points are also derived. In the alignment stage, correspondences between the learned holistic face model and an input partial face image are established by matching keypoints of the partial face to the anchor points of the learned face model. Furthermore, shape constraint is used to eliminate outlier correspondences and temporal constraint is explored to find more inliers. Alignment is finally accomplished by solving a similarity transform. Experiments on the MBGC near infrared video sequences show the effectiveness of the proposed method, especially when PCA subspace, shape and temporal constraint are utilized.

Keywords

Face Alignment Partial Faces SIFT MBGC 

References

  1. 1.
    Cootes, T., Cooper, D., Taylor, C., Gramham, J.: Active shape models - their training and application. Computer Vision and Image Understanding 61, 38–59 (1995)Google Scholar
  2. 2.
    Cootes, T., Adwards, G., Taylor, C.: Active appearance model. IEEE Transactions on Pattern Analysis and Machine Intelligence 23(6), 681–685 (2001)Google Scholar
  3. 3.
    Lowe, D.G.: Distinctive image features from scale-invariant keypoints. International Journal of Computer Vision 60(2) (2004)Google Scholar
  4. 4.
    Slot, K., Kim, H.: Keypoints derivation for object class detection with sift algorithm. In: Rutkowski, L., Tadeusiewicz, R., Zadeh, L.A., Żurada, J.M. (eds.) ICAISC 2006. LNCS, vol. 4029, pp. 850–859. Springer, Heidelberg (2006)Google Scholar
  5. 5.
    Carneiro, G., Jepson, A.D.: Pruning local feature correspondences using shape context. In: Proceedings of IEEE International Conference on Image Processing (2004)Google Scholar
  6. 6.
    Belongie, S., Malik, J., Puzicha, J.: Shape matching and object recognition using shape contexts. IEEE Transactions on Pattern Analysis and Machine Intelligence 24(4), 509–522 (2002)Google Scholar
  7. 7.
    Lowe, D.G.: Local feature view clustering for 3D object recognition. In: Proceedings of IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 682–688 (2001)Google Scholar
  8. 8.
    NIST: Multiple Biometric Grand Challenge, MBGC (2008), http://face.nist.gov/mbgc
  9. 9.
    Mikolajczyk, K., Tuytelaars, T., Schmid, C., Zisserman, A., Matas, J., Schaffalitzky, F., Kadir, T., Gool, L.V.: A comparison of affine region detectors. International Journal of Computer Vision 65(12), 43–72 (2005)Google Scholar
  10. 10.
    Mikolajczyk, K., Schmid, C.: A performance evaluation of local descriptors. IEEE Transactions on Pattern Analysis and Machine Intelligence 27(10), 1615–1630 (2005)Google Scholar

Copyright information

© Springer-Verlag Berlin Heidelberg 2009

Authors and Affiliations

  • Jimei Yang
    • 1
    • 2
  • Shengcai Liao
    • 2
  • Stan Z. Li
    • 2
  1. 1.University of Science and Technology of ChinaHefeiChina
  2. 2.Center for Biometrics and Security Research, Institute of AutomationChinese Academy of SciencesBeijingChina

Personalised recommendations