Abstract
In this article, we consider prediction of a univariate response from background data. The data may have a near-collinear structure and additionally group effects are assumed to exist. A two-step method is proposed. The first step summarizes the information in the predictors via a bilinear model. The bilinear model has a Krylov structured within individual design matrix, which is the link to classical partial least squares (PLS) analysis and a between-individual design matrix which handles group effects. The second step is the prediction step where a conditional expectation approach is used. The two-step method gives new insight into PLS. Explicit maximum likelihood estimators of the dispersion matrix and mean for the predictors are derived under the assumption that the covariance between the response and explanatory variables is known. It is shown that for within-sample prediction the mean squared error of the two-step method is always smaller than PLS.
Similar content being viewed by others
References
Björkström, A. (2010). Krylov sequences as a tool for analysing iterated regression algorithms. Scand. J. Stat., 37(1), 166–175.
Brown, P.J. (1993). Measurement, regression, and calibration. Oxford Statistical Science Series, vol. 12. The Clarendon Press Oxford University Press, New York.
Cook, R.D., Li, B. and Chiaromonte, F. (2010). Envelope models for parsimonious and efficient multivariate linear regression (with discussion). Statist. Sinica, 20, 927–1010 .
Fearn, T. (1983). A misuse of ridge regression in the calibration of a near infrared reflectance instrument. J. R. Stat. Soc. Ser. C, 32, 73–79.
Frank, I.E., and Friedman, J.H. (1993). A statistical view of some chemometrics regression tools. Technometrics, 35(2), 109–135.
Helland, I.S. (1988). On the structure of partial least squares regression. Comm. Statist. Simulation Comput., 17, 581–607.
Helland, I.S. (1990). Partial least squares regression and statistical models. Scand. J. Stat., 17, 97–114.
Helland, I.S. (1992). Maximum likelihood regression on relevant components. J. R. Stat. Soc. Ser. B, 54, 637–647.
Helland, I.S. (2010). Discussion of Cook, R.D., Li, B. and Chiaromonte, F. (2010). Envelope models for parsimonious and efficient multivariate linear regression. Statist. Sinica, 20, 978–981.
Hoerl, A.E., Kennard, R.W. and Hoerl, R.W. (1985). Practical use of ridge regression: A challenge met. J. R. Stat. Soc. Ser. C, 34, 114–120.
Kollo, T., and Von Rosen, D. (2005). Advanced Multivariate Statistics with Matrices. Mathematics and Its Applications (New York), vol. 579. Springer, Dordrecht.
Kondylis, A. and Whittaker, J. (2008). Spectral preconditioning of Krylov spaces: Combining PLS and PC regression. Comput. Statist. Data Anal., 52, 2588–2603.
Li, Y. and Von Rosen, D. (2012). Maximum likelihood estimators in a two step model for PLS. Comm. Statist. Theory Methods, 41, 2503–2511.
Ohlson, M., Ahmad, M.R. and Von Rosen, D. (2013). The multilinear normal distribution: Introduction and some basic properties. J. Multivariate Anal., 113, 37–47.
Potthoff, R.F. and Roy, S.N. (1964). A generalized multivariate analysis of variance model useful especially for growth curve problems. Biometrika, 51, 313–326.
Von Rosen, D. (1991). The growth curve model: A review. Comm. Statist. Theory Methods, 20, 2791–2822.
Von Rosen, D. (1994). PLS, linear models and invariant spaces. Scand. J. Stat., 21, 179–186.
Srivastava, M.S. and Khatri, C.G. (1979) An introduction to multivariate statistics. North-Holland, New York.
Srivastava, M.S. and Von Rosen, D. (1999). Growth curve models. In Multivariate Analysis, Design of Experiments, and Survey Sampling. Statist. Textbooks Monogr., vol. 159, pp. 547–578. Dekker, New York.
Stone, M. and Brooks, R.J. (1990). Continuum regression: Cross-validated sequentially constructed prediction embracing ordinary least squares, partial least squares and principal components regression. J. R. Stat. Soc. Ser. B, 52(2), 237–269. With discussion and a reply by the authors.
Sundberg, R. (1999). Multivariate calibration—direct and indirect regression methodology. Scand. J. Stat., 26, 161–207.
Tibshirani, R. (1996). Regression shrinkage and selection via the lasso. J. R. Statist. Soc. Ser. B, 58, 267–288.
Wold, S., Martens, H. and Wold, H. (1983). The multivariate calibration problem in chemistry solved by the PLS method. In Matrix Pencils. Lecture Notes in Mathematics, (B. Kågström and A. Ruhe, eds.), vol. 973, pp. 286–293. Springer Berlin / Heidelberg.
Woolson, R.F. and Leeper, J.D. (1980). Growth curve analysis of complete and incomplete longitudinal data. Comm. Statist. Theory Methods, 9, 1491–1513.
Author information
Authors and Affiliations
Corresponding author
Additional information
An erratum to this article is available at http://dx.doi.org/10.1007/s13171-015-0077-4.
A Appendix: Proof of Theorem 2
A Appendix: Proof of Theorem 2
Proof
Let us start with the proof of \(\boldsymbol{E} (\boldsymbol{y}-\hat{\boldsymbol{y}}_{a,TS})'(\hat{\boldsymbol{y}}_{a,TS}-\hat{\boldsymbol{y}}_{a,PLS})=0\). Based on the inverse binomial theorem \(\hat{{\boldsymbol{\Sigma}}}^{-1}\) can be split into two parts,
where
with
In order to compare the \(\hat{\boldsymbol{y}}_{a,PLS}\) with \(\hat{\boldsymbol{y}}_{a,TS}\), we replace \(\hat{\boldsymbol{G}}_a\) with \(n \boldsymbol{S}^{-1} \hat{\boldsymbol{A}}\), and using (A.1)
Since \(\boldsymbol{K}\hat{\boldsymbol{A}}(\hat{\boldsymbol{A}}'\boldsymbol{S}^{-1}\hat{\boldsymbol{A}})^{-}\hat{\boldsymbol{A}}'\boldsymbol{S}^{-1}\boldsymbol{X} \boldsymbol{P}_{\mathbf{1}}=\mathbf{0}\), (A.2) reduces to
For \(\boldsymbol{y}-\boldsymbol{y}_{a,TS}\),
Then we apply the result of (A.3) and (A.4),
which equals 0 if
For the second inequality, we need to prove that the sufficient condition \(\boldsymbol{E}(\boldsymbol{y}-\hat{\boldsymbol{y}}_L)'(\hat{\boldsymbol{y}}_L-\hat{\boldsymbol{y}}_{a,TS})=0 \) holds, where
where, (A.7) follows (A.6) since \( \hat{\boldsymbol{\mu}}_x'\boldsymbol{K}=\mathbf{0}. \) Furthermore,
since \(\boldsymbol{I}=\boldsymbol{S}^{-1} \boldsymbol{X}(\boldsymbol{I}-\boldsymbol{P}_{\mathbf{1}})\boldsymbol{X}'\). Thus \((\boldsymbol{y}-\hat{\boldsymbol{y}}_L)'(\hat{\boldsymbol{y}}_L-\hat{\boldsymbol{y}}_{a,TS})=0\).
Rights and permissions
About this article
Cite this article
Li, Y., Udén, P. & von Rosen, D. A two-step PLS inspired method for linear prediction with group effect. Sankhya A 75, 96–117 (2013). https://doi.org/10.1007/s13171-012-0022-8
Received:
Revised:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s13171-012-0022-8