Abstract
In many engineering applications, estimation of covariance and precision matrices is of great importance, helping researchers understand the dependency and conditional dependency between variables of interest. Among various matrix estimation methods, the modified Cholesky decomposition is a commonly used technique. It has the advantage of transforming the matrix estimation task into solving a sequence of regression models. Moreover, the sparsity on the regression coefficients implies certain sparse structure on the covariance and precision matrices. In this chapter, we first overview the Cholesky-based covariance and precision matrices estimation. It is known that the Cholesky-based matrix estimation depends on a prespecified ordering of variables, which is often not available in practice. To address this issue, we then introduce several techniques to enhance the Cholesky-based estimation of covariance and precision matrices. These approaches are able to ensure the positive definiteness of the matrix estimate and applicable in general situations without specifying the ordering of variables. The advantage of Cholesky-based estimation is illustrated by numerical studies and several real-case applications.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Bickel, P.J., Levina, E.: Covariance regularization by thresholding. Ann. Stat. 36(6), 2577–2604 (2008a)
Bickel, P.J., Levina, E.: Regularized estimation of large covariance matrices. Ann. Stat. 36(1), 199–227 (2008b)
Bien, J., Tibshirani, R.J.: Sparse estimation of a covariance matrix. Biometrika 98(4), 807–820 (2011)
Cai, T.T., Yuan, M.: Adaptive covariance matrix estimation through block thresholding. Ann. Stat. 40(40), 2014–2042 (2012)
Cai, T.T., Ren, Z., Zhou, H.H.: Estimating structured high-dimensional covariance and precision matrices: optimal rates and adaptive estimation. Electronic Journal of Statistics 10(1), 1–59 (2016)
Chang, C., Tsay, R.S.: Estimation of covariance matrix via the sparse Cholesky factor with lasso. J. Stat. Plann. Inference 140(12), 3858–3873 (2010)
Cochran, W.G.: Sampling Techniques. Wiley, New York (1977)
Deng, X., Lin, C.D., Liu, K.-W., Rowe, R.K.: Additive Gaussian process for computer models with qualitative and quantitative factors. Technometrics 59(3), 283–292 (2017)
Fan, J., Li, R.: Variable selection via nonconcave penalized likelihood and its oracle properties. J. Am. Stat. Assoc. 96 1348–1360 (2001)
Fan, J., Liao, Y., Liu, H.: An overview of the estimation of large covariance and precision matrices. Econ. J. 19(1), 1–32 (2016)
Francq, C., Zakoïan, J. M.: Estimating multivariate volatility models equation by equation. J. R. Stat. Soc. Ser. B (Stat Methodol.) 78(3), 613–635 (2016)
Friedman, J., Hastie, T., Tibshirani, R.: Regularization paths for generalized linear models via coordinate descent. J. Stat. Softw. 33, 1–22 (2010)
Huang, J.Z., Liu, N., Pourahmadi, M., Liu, L.: Covariance matrix selection and estimation via penalised normal likelihood. Biometrika 93, 85–98 (2006)
Kang, X., Xie, C., Wang, M.: A Cholesky-based estimation for large-dimensional covariance matrices. J. Appl. Stat. 47, 1017–1030 (2020)
Kang, X., Deng, X., Tsui, K., Pourahmad,i M.: On variable ordination of modified Cholesky decomposition for estimating time-varying covariance matrices. Int. Stat. Rev., 88(3), 616–641 (2020)
Lam, C., Fan, J.: Sparsistency and rates of convergence in large covariance matrix estimation. Ann. Stat. 37, 4254–4278 (2009)
Lan, Q., Sun, H., Robertson, J., Deng, X., Jin, R.: Non-invasive assessment of liver 1uality in transplantation based on thermal imaging analysis. Comput. Methods Prog. Biomed. 164, 31–47 (2018)
Leng, C., Li, B.: Forward adaptive banding for estimating large covariance matrices. Biometrika 98(4), 821–830 (2011)
Levina, E., Zhu, R.J.: Sparse estimation of large covariance matrices via a nested lasso penalty. Ann. Appl. Stat. 2(1), 245–263 (2008)
Liu, H., Wang, L., Zhao, T.: Sparse covariance matrix estimation with eigenvalue constraints. J. Comput. Graph. Stat. 23(2), 439–459 (2014)
Mohammadi, A., Wit, E.C.: Bayesian structure learning in sparse Gaussian graphical models. Bayesian Anal. 10(1), 109–138 (2015)
Nino-Ruiz, E.D., Sandu, A., Deng, X.: An ensemble Kalman filter implementation based on modified cholesky decomposition for inverse covariance matrix estimation. SIAM J. Sci. Comput. 40(2), A867–CA886 (2018)
Pedeli, X., Fokianos, K., Pourahmadi, M.: Two Cholesky-log-GARCH models for multivariate volatilities. Stat. Model. 15, 233–255 (2015)
Pourahmadi, M.: Joint mean-covariance models with applications to longitudinal data: unconstrained parameterisation. Biometrika 86, 677–690 (1999)
Rajaratnam, B., Salzman, J.: Best permutation analysis. J. Multivar. Anal. 121, 193–223 (2013)
Rigollet, P., Tsybakov, A.: Estimation of covariance matrices under sparsity constraints. Probl. Inf. Transm. 51(4), 32–46 (2012)
Rothman, A.J., Levina, E., Zhu, J.: Generalized thresholding of large covariance matrices. J. Am. Stat. Assoc. 104(485), 177–186 (2009)
Rothman, A.J., Levina, E., Zhu, J.: A new approach to Cholesky-based covariance regularization in high dimensions. Biometrika 97(3), 539–550 (2010)
Sapsanis, C., Georgoulas, G., Tzes, A., Lymberopoulos, D.: Improving EMG based classification of basic hand movements using EMD. In: Proceedings of the 2013 35th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), pp. 5754–5757 (2013)
Sun, H., Huang, S., Jin, R.: Functional graphical models for manufacturing process modeling. IEEE Trans. Autom. Sci. Eng. 14(4), 1612–1621 (2017)
Sun, H., Rao, P.K., Kong, Z., Deng, X., Jin, R.: Functional quantitative and qualitative models for quality modeling in a fused deposition modeling process. IEEE Trans. Autom. Sci. Eng. 15(1), 393–403 (2018)
Tan, L.S., Nott, D.J.: Gaussian variational approximation with sparse precision matrices. Stat. Comput. 28(2), 259–275 (2018)
Tibshirani, R.: Regression shrinkage and selection via the lasso. J. R. Stat. Soc. Ser. B (Stat Methodol.) 58, 267–288 (1996)
Wagaman, A., Levina, E.: Discovering sparse covariance structures with the Isomap. J. Comput. Graph. Stat. 18(3), 551–572 (2009)
Wu, H., Deng, X., Ramakrishnan, N.: Sparse estimation of multivariate poisson log-normal model and inverse covariance for counting data. Stat. Anal. Data Min. 11, 66–77 (2018)
Xue, L., Ma, S., Zou, H.: Positive-definite L1-penalized estimation of large covariance matrices. J. Am. Stat. Assoc. 107(500), 1480–1491 (2012)
Yu, P.L.H., Wang, X., Zhu, Y.: High dimensional covariance matrix estimation by penalizing the matrix-logarithm transformed likelihood. Comput. Stat. Data Anal. 114, 12–25 (2017)
Yuan, M., Lin, Y.: Model selection and estimation in the Gaussian graphical model. Biometrika 94, 19–35 (2007)
Zeng, L., Deng, X., Yang, J.: A constrained Gaussian process approach to modeling tissue-engineered scaffold degradation. IISE Trans. 50(5), 431–447 (2018)
Zheng, H., Tsui, K.-W., Kang, X., Deng, X.: Cholesky-based model averaging for covariance matrix estimation. Stat. Theory Relat. Fields 1(1), 48–58 (2017)
Acknowledgements
The authors would like to thank the editor and reviewers for the constructive and insightful comments, which have significantly enhanced the quality of this article.
Proof of Remark 1 and 2
Since the conclusions of Remarks 1 and 2 are much similar, we only provide the proof of Remark 2 here. Assume that there are n independent and identically distributed observations x1, …, xn, which are centered. Let \(\boldsymbol S = \frac {1}{n} \sum _{i=1}^{n} \boldsymbol x_{i} \boldsymbol x^{\prime }_{i}\) be the sample covariance matrix and assume that S is non-singular since n > p. We denote \(\hat {\boldsymbol \Sigma }_{0}\) as the estimated covariance matrix from (43.12) with tuning parameters equal to zeroes in (43.11). Then Remark 2 states that \(\hat {\boldsymbol \Sigma }_{0} = \boldsymbol S\) in spite of any permutation of x1, …, xn. Below is the proof.
Based on the sequential regression of (43.11), it is known that the first step is X1 = 𝜖1. It means that
Then the second step is to consider X2 = l21𝜖1 + 𝜖2, which provides
In general, the jth step is to consider the regression problem as
and we can obtain
Therefore, we can express the (s, t) entry of the covariance matrix estimate using the regression coefficients as
Note that
and the (s, t) entry of the sample covariance matrix is
The last equality holds because of
Thus, we can establish the result
□
Conditional Misclassification Error of LDA
Without loss of generality, we consider a two-class classification problem here. Suppose the binary classifier function for LDA is \(g(\boldsymbol x) = \log [ P(Y = 1|\boldsymbol X = \boldsymbol x) / P(Y = 2|\boldsymbol X = \boldsymbol x) ]\). Then
where π1 and π2 are the prior probabilities for class 1 and 2, respectively, i.e., π1 = P(Y = 1) and π2 = P(Y = 2). For a new observation x, we predict its class Y = 1 if g(x) > 0, and Y = 2 otherwise. Then the conditional misclassification error is
Since x|Y = 1 ∼ N(μ1, Σ), and x|Y = 2 ∼ N(μ2, Σ), obviously aTx|Y = 1 ∼ N(aTμ1, aT Σa), and aTx|Y = 2 ∼ N(aTμ2, aT Σa). Therefore,
where Φ(⋅) is the cumulative distribution function of the standard normal random variable. As a result, the conditional misclassification error is
Assume π1 = π2 = 1∕2. Then with the estimates of a and b through \(\hat {\boldsymbol \mu }_{1}, \hat {\boldsymbol \mu }_{2}, \hat {\boldsymbol \Sigma }\), the conditional misclassification error \(\gamma (\hat {\boldsymbol \Sigma }, \hat {\boldsymbol \mu }_{1}, \hat {\boldsymbol \mu }_{2})\) is
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Section Editor information
Rights and permissions
Copyright information
© 2023 Springer-Verlag London Ltd., part of Springer Nature
About this chapter
Cite this chapter
Kang, X., Zhang, Z., Deng, X. (2023). Covariance Estimation via the Modified Cholesky Decomposition. In: Pham, H. (eds) Springer Handbook of Engineering Statistics. Springer Handbooks. Springer, London. https://doi.org/10.1007/978-1-4471-7503-2_43
Download citation
DOI: https://doi.org/10.1007/978-1-4471-7503-2_43
Published:
Publisher Name: Springer, London
Print ISBN: 978-1-4471-7502-5
Online ISBN: 978-1-4471-7503-2
eBook Packages: Mathematics and StatisticsMathematics and Statistics (R0)