Towards Pose-Invariant 2D Face Classification for Surveillance
A key problem for “face in the crowd” recognition from existing surveillance cameras in public spaces (such as mass transit centres) is the issue of pose mismatches between probe and gallery faces. In addition to accuracy, scalability is also important, necessarily limiting the complexity of face classification algorithms. In this paper we evaluate recent approaches to the recognition of faces at relatively large pose angles from a gallery of frontal images and propose novel adaptations as well as modifications. Specifically, we compare and contrast the accuracy, robustness and speed of an Active Appearance Model (AAM) based method (where realistic frontal faces are synthesized from non-frontal probe faces) against bag-of-features methods (which are local feature approaches based on block Discrete Cosine Transforms and Gaussian Mixture Models). We show a novel approach where the AAM based technique is sped up by directly obtaining pose-robust features, allowing the omission of the computationally expensive and artefact producing image synthesis step. Additionally, we adapt a histogram-based bag-of-features technique to face classification and contrast its properties to a previously proposed direct bag-of-features method. We also show that the two bag-of-features approaches can be considerably sped up, without a loss in classification accuracy, via an approximation of the exponential function. Experiments on the FERET and PIE databases suggest that the bag-of-features techniques generally attain better performance, with significantly lower computational loads. The histogram-based bag-of-features technique is capable of achieving an average recognition accuracy of 89% for pose angles of around 25 degrees.
KeywordsFace Recognition Discrete Cosine Transform Face Image Gaussian Mixture Model Active Appearance Model
Unable to display preview. Download preview PDF.
- 1.McCahill, M., Norris, C.: Urbaneye: CCTV in London. Centre for Criminology and Criminal Justice, University of Hull, UK (2002)Google Scholar
- 2.Phillips, P., Grother, P., Micheals, R., Blackburn, D., Tabassi, E., Bone, M.: Face recognition vendor test 2002. In: Proc. Analysis and Modeling of Faces and Gestures, p. 44 (2003)Google Scholar
- 3.Blanz, V., Grother, P., Phillips, P., Vetter, T.: Face recognition based on frontal views generated from non-frontal images. In: Proc. IEEE Int. Conf. Computer Vision and Pattern Recognition, vol. 2, pp. 454–461. IEEE Computer Society Press, Los Alamitos (2005)Google Scholar
- 4.Shan, T., Lovell, B., Chen, S.: Face recognition robust to head pose from one sample image. In: Proc. 18th Int. Conf. Pattern Recognition (ICPR), vol. 1, pp. 515–518 (2006)Google Scholar
- 7.Lucey, S., Chen, T.: Learning patch dependencies for improved pose mismatched face verification. In: IEEE Conf. Computer Vision and Pattern Recognition, vol. 1, pp. 909–915. IEEE Computer Society Press, Los Alamitos (2006)Google Scholar
- 10.Csurka, G., Dance, C., Fan, L., Willamowski, J., Bray, C.: Visual cetegorization with bags of keypoints. In: Workshop on Statistical Learning in Computer Vision (co-located with ECCV 2004) (2004)Google Scholar
- 11.Sivic, J., Zisserman, A.: Video google: A text retrieval approach to object matching in videos. In: Proc. 9th International Conference on Computer Vision (ICCV), vol. 2, pp. 1470–1477 (2003)Google Scholar
- 15.Cootes, T., Taylor, C.: Active shape models - ‘smart snakes’. In: Proc. British Machine Vision Conference, pp. 267–275 (1992)Google Scholar
- 18.Cootes, T., Walker, K., Taylor, C.: View-based active appearance models. In: Proc. 4th IEEE International Conference on Automatic Face and Gesture Recognition, pp. 227–232. IEEE Computer Society Press, Los Alamitos (2000)Google Scholar
- 21.Gonzales, R., Woods, R.: Digital Image Processing. Addison-Wesley, Reading (1992)Google Scholar
- 23.Wallraven, C., Caputo, B., Graf, A.: Recognition with local features: the kernel recipe. In: Proc. 9th International Conference on Computer Vision (ICCV), vol. 1, pp. 257–264 (2003)Google Scholar