Skip to main content

On the Bias in Eigenvalues of Sample Covariance Matrix

  • Conference paper
  • First Online:
Quantitative Psychology (IMPS 2017)

Part of the book series: Springer Proceedings in Mathematics & Statistics ((PROMS,volume 233))

Included in the following conference series:

Abstract

Principal component analysis (PCA) is a multivariate statistical technique frequently employed in research in behavioral and social sciences, and the results of PCA are often used to approximate those of exploratory factor analysis (EFA) because the former is easier to implement. In practice, the needed number of components or factors is often determined by the size of the first few eigenvalues of the sample covariance/correlation matrix. Lawley (1956) showed that if eigenvalues of population covariance matrix are distinct, then each sample eigenvalue contains a bias of order 1/N, which is typically ignored in practice. This article further shows that, under some regulatory conditions, the order of the bias term is p/N. Thus, when p is large, the bias term is no longer negligible even when N is large.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 99.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 129.00
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 169.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  • Anderson, T. W. (1963). Asymptotic theory for principal component analysis. Annals of Mathematical Statistics, 34, 122–148.

    Article  MathSciNet  Google Scholar 

  • Anderson, T. W. (2003). An introduction to multivariate statistical analysis (3rd ed.). New York: Wiley.

    MATH  Google Scholar 

  • Arruda, E. H., & Bentler, P. M. (2017). A regularized GLS for structural equation modeling. Structural Equation Modeling, 24, 657–665.

    Article  MathSciNet  Google Scholar 

  • Bendel, R. B. & Mickey, M. R. (1978). Population correlation matrices for sampling experiments. Communications in StatisticsSimulation and Computation, 7, 163–182.

    Article  Google Scholar 

  • Bentler, P. M., & Kano, Y. (1990). On the equivalence of factors and components. Multivariate Behavioral Research, 25, 67–74.

    Article  Google Scholar 

  • Davies, P. I., & Higham, N. J. (2000). Numerically stable generation of correlation matrices and their factors. BIT, 40, 640–651.

    Article  MathSciNet  Google Scholar 

  • Guttman, L. (1956). “Best possible” estimates of communalities. Psychometrika, 21, 273–286.

    Article  MathSciNet  Google Scholar 

  • Hotelling, H. (1933). Analysis of a complex of statistical variables into principal components. Journal of Educational Psychology, 24, 417–441, 498–520.

    Article  Google Scholar 

  • Hwang, H., & Takane, Y. (2004). Generalized structured component analysis. Psychometrika, 69, 81–99.

    Article  MathSciNet  Google Scholar 

  • IBM Corp. (2016). IBM SPSS statistics for windows, Version 24.0. Armonk, NY: IBM Corp.

    Google Scholar 

  • Johnstone, I. M., & Lu, A. Y. (2009). On consistency and sparsity for principal components analysis in high dimensions. Journal of the American Statistical Association, 104, 682–703.

    Article  MathSciNet  Google Scholar 

  • Krijnen, W. P. (2006). Convergence of estimates of unique variances in factor analysis, based on the inverse of sample covariance matrix. Psychometrika, 71, 193–199.

    Article  MathSciNet  Google Scholar 

  • Lawley, D. N. (1956). Tests of significance for the latent roots of covariance and correlation matrices. Biometrika, 43, 128–136.

    Article  MathSciNet  Google Scholar 

  • Muirhead, R. J. (1982). Aspects of multivariate statistical theory. New York: Wiley.

    Google Scholar 

  • SAS Institute. Appendix C: Generating random correlation matrices. Retrieved April 28, 2017, from https://support.sas.com/publishing/authors/extras/65378_Appendix_C_Generating_Random_Correlation_Matrices.pdf.

  • Schneeweiss, H., & Mathes, H. (1995). Factor analysis and principal components. Journal of Multivariate Analysis, 55, 105–124.

    Article  MathSciNet  Google Scholar 

  • Stewart, G. W. (1980). The efficient generation of random orthogonal matrices with an application to condition estimation. SIAM Journal on Numerical Analysis, 17, 403–409.

    Article  MathSciNet  Google Scholar 

  • Velicer, W. F., & Jackson, D. N. (1990). Component analysis versus common factor analysis: Some issues in selecting an appropriate procedure. Multivariate Behavioral Research, 25, 1–28.

    Article  Google Scholar 

  • Yuan, K.-H., & Chan, W. (2008). Structural equation modeling with near singular covariance matrices. Computational Statistics & Data Analysis, 52, 4842–4858.

    Article  MathSciNet  Google Scholar 

  • Yuan, K.-H., & Chan, W. (2016). Structural equation modeling with unknown population distribution: Ridge generalized least squares. Structural Equation Modeling, 23, 163–179.

    Article  MathSciNet  Google Scholar 

Download references

Acknowledgements

The authors are thankful to Dr. Dylan Molenaar for his very helpful comments. Ke-Hai Yuan’s work was supported by the National Science Foundation under Grant No. SES-1461355.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Kentaro Hayashi .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2018 Springer International Publishing AG, part of Springer Nature

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Hayashi, K., Yuan, KH., Liang, L. (2018). On the Bias in Eigenvalues of Sample Covariance Matrix. In: Wiberg, M., Culpepper, S., Janssen, R., González, J., Molenaar, D. (eds) Quantitative Psychology. IMPS 2017. Springer Proceedings in Mathematics & Statistics, vol 233. Springer, Cham. https://doi.org/10.1007/978-3-319-77249-3_19

Download citation

Publish with us

Policies and ethics