Matrix-variate data arise in many scientific fields such as face recognition, medical imaging, etc. Matrix data contain important structure information which can be ruined by vectorization. Methods incorporating the structure information into analysis have significant advantages over vectorization approaches. In this article, we consider the problem of two-class classification with high-dimensional matrix-variate data, and propose a novel portmanteau-local-feature discrimination (PLFD) method. This method first identifies local discrimination features of the matrix variate and then pools them together to construct a discrimination rule. We investigated the theoretical properties of the PLFD method and established its asymptotic optimality. We carried out extensive numerical studies including simulation and real data analysis to compare this method with other methods available in the literature, which demonstrate that the PLFD method has a great advantage over the other methods in terms of misclassification rate.
This is a preview of subscription content, access via your institution.
Buy single article
Instant access to the full article PDF.
Tax calculation will be finalised during checkout.
Anderson, T.W. (2003). An introduction to multivariate statistical analysis. Wiley, New York.
Guo, J. (2010). Simultaneous variable selection and class fusion for high-dimensional linear discriminant analysis. Biostatistics 11, 599–608.
Gupta, A. and Nagar, D. (1999). Matrix variate distributions. no. 104 in monographs and surveys in pure and applied mathematics. Chapman & Hall/CRC, Florida.
Gupta, A.K. and Varga, T. (1993). Elliptically contoured models in statistics. Springer, Berlin.
Hung, H. and Wang, C.C. (2013). Matrix variate logistic regression model with application to eeg data. Biostatistics 14, 189–202.
Koltchinskii, V., Lounici, K., Tsybakov, A.B. et al. (2011). Nuclear-norm penalization and optimal rates for noisy low-rank matrix completion. Ann. Stat. 39, 2302–2329.
Lai, Z., Xu, Y., Yang, J., Tang, J. and Zhang, D. (2013). Sparse tensor discriminant analysis. IEEE Trans. Image Process. 22, 3904–3915.
Li, B., Kim, M.K. and Altman, N. (2010). On dimension folding of matrix- or array-valued statistical objects. Ann. Stat. 38, 1094–1121.
Li, M and Yuan, B (2005). 2d-lda: A statistical linear discriminant analysis for image matrix, 26, p. 527–532.
Li, Q. and Schonfeld, D. (2014). Multilinear discriminant analysis for higher-order tensor data classification. IEEE Trans. Pattern Anal. Mach. Intell. 36, 2524–2537.
Luo, S. and Chen, Z. (2020). A procedure of linear discrimination analysis with detected sparsity structure for high-dimensional multi-class classification. J. Multivar. Anal. 179, 104641.
Masulli, F. and Rovetta, S. (2015). Clustering high-dimensional data. In Clustering High–Dimensional Data. Springer, Berlin, pp 1–13.
Molstad, A.J. and Rothman, A.J. (2019). A penalized likelihood method for classification with matrix-valued predictors. J. Comput. Graph. Stat. 28, 11–22.
Pan, Y., Mai, Q. and Zhang, X. (2019). Covariate-adjusted tensor classification in high dimensions. J. Am. Stat. Assoc. 114, 1305–1319.
Recht, B., Fazel, M. and Parrilo, P.A. (2010). Guaranteed minimum-rank solutions of linear matrix equations via nuclear norm minimization. SIAM Rev. 52, 471–501.
Shao, J., Wang, Y., Deng, X. and Wang, S. (2011). Sparse linear discriminant analysis by thresholding for high dimensional data. Ann. Stat. 1241–1265.
Witten, D.M. and Tibshirani, R. (2010). A framework for feature selection in clustering. J. Am. Stat. Assoc. 105, 713–726.
Xu, Z. (2020). Sparse linear discriminant analysis for high dimensional gaussian matrix-valued predictors. PhD thesis, Shanghai Jiao Tong University.
Zhang, X.L., Begleiter, H., Porjesz, B., Wang, W. and Litke, A. (1995). Event related potentials during object recognition tasks. Brain Res. Bull.38, 531–538.
Zheng, WS, Lai, JH and Li, SZ (2008). 1d-lda vs. 2d-lda: When is vector-based linear discriminant analysis better than matrix-based?. Patt. Recogn. 41, 2156–2172.
Zhong, W. and Suslick, K.S. (2015). Matrix discriminant analysis with application to colorimetric sensor array data. Technometrics A J. Sta. Phys. Chem. Eng. Sci. 57, 524.
Zhou, H. and Li, L. (2014). Regularized matrix regression. J. R. Stat. Soc. Ser. B-stat. Methodol. 76, 463–483.
Zhou, H., Li, L. and Zhu, H. (2013). Tensor regression with applications in neuroimaging data analysis. J. Am. Stat. Assoc. 108, 540–552.
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Electronic supplementary material
Below is the link to the electronic supplementary material.
About this article
Cite this article
Xu, Z., Luo, S. & Chen, Z. A Portmanteau Local Feature Discrimination Approach to the Classification with High-dimensional Matrix-variate Data. Sankhya A (2021). https://doi.org/10.1007/s13171-021-00255-2
- Matrix-variate data
- Feature selection
- Asymptotic optimality.
- Primary AMS62H; Secondary AMS62H30