Tensor Factorization by Simultaneous Estimation of Mixing Factors for Robust Face Recognition and Synthesis
Facial images change appearance due to multiple factors such as poses, lighting variations, facial expressions, etc. Tensor approach, an extension of conventional matrix, is appropriate to analyze facial factors since we can construct multilinear models consisting of multiple factors using tensor framework. However, given a test image, tensor factorization, i.e., decomposition of mixing factors, is a difficult problem especially when the factor parameters are unknown or are not in the training set. In this paper, we propose a novel tensor factorization method to decompose the mixing factors of a test image. We set up a tensor factorization problem as a least squares problem with a quadratic equality constraint, and solve it using numerical optimization techniques. The novelty in our approach compared to previous work is that our tensor factorization method does not require any knowledge or assumption of test images. We have conducted several experiments to show the versatility of the method for both face recognition and face synthesis.
KeywordsFace Recognition Test Image Face Image Projection Method Training Image
Unable to display preview. Download preview PDF.
- 2.Vasilescu, M.A.O., Terzopoulos, D.: Multilinear independent components analysis. In: Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, vol. 1(20-25), pp. 547–553 (June 2005)Google Scholar
- 3.Vasilescu, M.A.O., Terzopoulos, D.: Multilinear subspace analysis of image ensembles. In: Proceedings of the 2003 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, vol. 2, pp. II-93–II-99 (June 2003)Google Scholar
- 4.Wang, H., Ahuja, N.: Facial expression decomposition. In: Proceedings of the Ninth IEEE International Conference on Computer Vision, vol. 2, pp. 958–965 (October 2003)Google Scholar
- 6.Lin, D., Xu, Y., Tang, X., Yan, S.: Tensor-based factor decomposition for relighting. In: IEEE International Conference on Image Processing, vol. 2(11–14), pp. 386–389 (2005)Google Scholar