Abstract
Least squares regression is one of the fundamental approaches for statistical analysis. And yet, its simplicity has often led to researchers overlooking it for complex recognition problems. In this chapter, we present a nonlinear regression framework on matrix manifolds. The proposed method is developed based upon two key attributes: underlying geometry and least squares fitting. The former attribute is vital since geometry characterizes the space for classification while the latter exhibits a simple estimation model. Considering geometric properties, we formulate the least squares regression as a composite function. The proposed framework can be naturally generalized to many matrix manifolds. We show that this novel formulation is applicable to matrix Lie groups, SPD(n), Grassmann manifolds, and the product of Grassmann manifolds for a number of computer vision applications ranging from visual tracking, object categorization, and activity recognition to human interaction recognition. Our experiments reveal that the proposed method yields competitive performance, including state-of-the-art results on challenging activity recognition benchmarks.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
Figure 14.2 is best viewed in color.
References
Absil PA, Mahony R, Sepulchre R (2008) Optimization algorithms on matrix manifolds. Princeton University Press, Princeton
Belinfante J, Kolman B (1972) A survey of lie groups and lie algebras with applications and computational methods. SIAM, Philadelphia
Conway J, Hardin R, Sloane N (1996) Packing lines, planes, etc.: packings in grassmannian spaces. Exp Math 5(2):139–159
Davis BC, Fletcher PT, Bullitt E, Joshi S (2007) Population shape regression from random design data. In: International conference on computer vision
Dollar P, Rabaud V, Cottrell G, Belongie S (2005) Behavior recognition via sparse spatio-temporal features. In: IEEE VS-PETS workshop, pp 65–72
Edelman A, Arias R, Smith S (1999) The geometry of algorithms with orthogonal constraints. SIAM J Matrix Anal Appl 2:303–353
Fiori S (2011) Visualization of riemannian-manifold-valued elements by multidimensional scaling. Neurocomputing 174:983–992
Fletcher PT (2011) Geodesic regression on riemannian manifolds. In: Third international workshop on mathematical foundations of computational anatomy
Gordon N, Salmond D, Smith A (1993) A novel approach to non-linear and non-gaussian Bayesian state estimation. IEE Proc F 140:107–113
Guo K, Ishwar P, Konrad J (2010) Action recognition using sparse representation on covariance manifold of optical flow. In: IEEE international conference on advanced video and signal-based surveillance, Boston
Hamm J, Lee D (2008) Grassmann discriminant analysis: a unifying view on subspace-based learning. In: International conference on machine learning, Helsinki, pp 376–383
Hastie T, Tibshirani R, Friedman J (2001) The elements of statistical learning: data mining, inference, and prediction. Springer, New York
He J, Balzano L, Szlam A (2012) Incremental gradient on the grassmannian for online foreground and background separation in subsampled video. In: IEEE conference on computer vision and pattern recognition
Ho H, Gopalan R (2014) Model-driven domain adaptation on product manifolds for unconstrained face recognition. Int J Comput Vis 109:110–125
Jayasumana S, Hartley R, Salzmann M, Li H, Harandi M (2015) Kernels on riemannian manifolds. In: Turaga P, Srivastava A (eds.) Riemannian computing in computer vision. Springer, New York (2015)
Karsten G, Hermann K (1973) How to conjugate c1-close group actions. Math Z 132:11–20
Kim TK, Kittler J, Cipolla R (2007) Discriminative learning and recognition of image set classes using canonical correlations. IEEE Trans Pattern Anal Mach Intell 29(6):1–14
Lathauwer LD, Moor BD, Vandewalle J (2000) A multilinear singular value decomposition. SIAM J Matrix Anal Appl 21:1253–1278
Lee J (2003) Introduction to smooth manifolds. Springer, New York
Leibe B, Schiele B (2003) Analyzing appearance and contour based methods for object categorization. In: IEEE conference on computer vision and pattern recognition
Lin D, Grimson E, Fisher J (2009) Learning visual flows: a lie algebraic approach. In: IEEE conference on computer vision and pattern recognition
Liu Y, Liu Y, Chan KCC (2011) Ordinal regression via manifold learning. In: Twenty-fifth AAAI conference on artificial intelligence, pp 398–403
Lui YM (2012) Advances in matrix manifolds for computer vision. Image Vis Comput 30:380–388
Lui YM (2012) Human gesture recognition on product manifolds. J Mach Learn Res 13:3297–3321
Lui YM, Beveridge JR, Kirby M (2010) Action classification on product manifolds. In: IEEE conference on computer vision and pattern recognition, San Francisco, pp 833–839
Ma Y, Kosecka J, Sastry S (1998) Motion estimation in computer vision: Optimization on stiefel manifolds. In: IEEE conference on decision and control, Tempa, pp 3751–3756
Meyer G, Bonnabel S, Sepulchre R (2011) Linear regression under fixed-rank constraints: a riemannian approach. In: International conference on machine learning
Pelletier B, Bataillon PE (2006) Non-parametric regression estimation on a closed riemannian manifold. J Nonparametric Stat 18:57–67
Pennec X, Fillard P, Ayache N (2006) A riemannian framework for tensor computing. Int J Comput Vis 66(1):728–735
Pham DS, Venkatesh S (2008) Robust learning of discriminative projection for multicategory classification on the stiefel manifold. In: IEEE conference on computer vision and pattern recognition
Porikli F, Pan P (2009) Regressed importance sampling on manifolds for efficient object tracking. In: IEEE international conference on advanced video and signal-based surveillance
Ross D, Lim J, Lin RS, Yang MH (2008) Incremental learning for robust visual tracking. Int J Comput Vis 77:125–141
Ryoo MS, Chen CC, Aggarwal JK, Roy-Chowdhu A (2010) An overview of contest on semantic description of human activities (sdha). Http://cvrc.ece.utexas.edu/SDHA2010/
Schüldt C, Laptev I, Caputo B (2004) Recognizing human actions: a local svm approach. In: International conference on pattern recognition, Cambridge
Schwartz WR, Kembhavi A, Harwood D, Davis LS (2009) Human detection using partial least squares analysis. In: International conference on computer vision
Srivastava A, Klassen E (2004) Bayesian, geometric subspace tracking. Adv Appl Probab 36(1):43–56
Su J, Srivastava A, de Souza F, Sarkar S (2014) Rate-invariant analysis of trajectories on riemannian manifolds with application in visual speech recognition. In: IEEE conference on computer vision and pattern recognition
Tuzel O, Porikli F, Meer P (2008) Learning on lie groups for invariant detection and tracking. In: IEEE conference on computer vision and pattern recognition
Vezzani R, Baltieri D, Cucchiara R (2010) Hmm based action recognition with projection histogram features. In: Recognizing patterns in signals, speech, images and videos, pp 286–293
Waltisberg D, Yao A, Gall J, Gool LV (2010) Variations of a hough-voting action recognition system. In: Recognizing patterns in signals, speech, images and videos, pp 306–312
Wei D, Lin D, Fisher J (2012) Learning deformations with parallel transport. In: European conference on computer vision (2012)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2016 Springer International Publishing Switzerland
About this chapter
Cite this chapter
Lui, Y.M. (2016). A General Least Squares Regression Framework on Matrix Manifolds for Computer Vision. In: Turaga, P., Srivastava, A. (eds) Riemannian Computing in Computer Vision. Springer, Cham. https://doi.org/10.1007/978-3-319-22957-7_14
Download citation
DOI: https://doi.org/10.1007/978-3-319-22957-7_14
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-22956-0
Online ISBN: 978-3-319-22957-7
eBook Packages: EngineeringEngineering (R0)