Skip to main content
Log in

Dimension reduction based on weighted variance estimate

  • Published:
Science in China Series A: Mathematics Aims and scope Submit manuscript

Abstract

In this paper, we propose a new estimate for dimension reduction, called the weighted variance estimate (WVE), which includes Sliced Average Variance Estimate (SAVE) as a special case. Bootstrap method is used to select the best estimate from the WVE and to estimate the structure dimension. And this selected best estimate usually performs better than the existing methods such as Sliced Inverse Regression (SIR), SAVE, etc. Many methods such as SIR, SAVE, etc. usually put the same weight on each observation to estimate central subspace (CS). By introducing a weight function, WVE puts different weights on different observations according to distance of observations from CS. The weight function makes WVE have very good performance in general and complicated situations, for example, the distribution of regressor deviating severely from elliptical distribution which is the base of many methods, such as SIR, etc. And compared with many existing methods, WVE is insensitive to the distribution of the regressor. The consistency of the WVE is established. Simulations to compare the performances of WVE with other existing methods confirm the advantage of WVE.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

  1. Cook R D. Regression Graphics. New York: Wiley, 1998

    Book  MATH  Google Scholar 

  2. Cook R D. On the interpretation of regression plots. J Amer Statist Assoc, 89: 177–189 (1994)

    Article  MATH  MathSciNet  Google Scholar 

  3. Cook R D. Graphics for regressions with a binary response. J Amer Statist Assoc, 91: 983–992 (1996)

    Article  MATH  MathSciNet  Google Scholar 

  4. Li K C. Sliced inverse regression for dimension reduction. J Amer Statist Assoc, 86: 316–342 (1991)

    Article  MATH  MathSciNet  Google Scholar 

  5. Cook R D, Weisberg S. Discussion of “sliced inverse regression for dimension reduction” by K-C Li. J Amer Statist Assoc, 86: 328–332 (1991)

    Article  Google Scholar 

  6. Li K C. On principal Hessian directions for the data visualization and dimension reduction: another application of Stein’s lemma. J Amer Statist Assoc, 87: 1025–1039 (1992)

    Article  MATH  MathSciNet  Google Scholar 

  7. Li B, Zha H, Chiaromente F. Contour Regression: a general approach to dimension reduction. Ann Statist, 33: 1580–1616 (2005)

    Article  MATH  MathSciNet  Google Scholar 

  8. Xia Y, Tong H, Li W K, et al. An adaptive estimation of dimension reduction space. J Roy Statist Soc Ser B, 64: 363–410 (2002)

    Article  MATH  MathSciNet  Google Scholar 

  9. Cook R D, Lee H. Dimension reduction in binary response regression. J Amer Statist Assoc, 94: 1187–1200 (1999)

    Article  MATH  MathSciNet  Google Scholar 

  10. Cook R D, Nachtsheim C J. Reweighting to achieve elliptically contoured covariates in regression. J Amer Statist Assoc, 89: 592–599 (1994)

    Article  MATH  Google Scholar 

  11. Ye Z, Weiss R E. Using the bootstrap to select one of a new class of dimension reduction methods. J Amer Statist Assoc, 98: 968–979 (2003)

    Article  MATH  MathSciNet  Google Scholar 

  12. Zhu L X, Ohtaki M, Li Y X. On hybrid method inverse regression-based algorithms. Comput Statist Data Anal, 51: 2621–2635 (2007)

    Article  MathSciNet  Google Scholar 

  13. Cook R D. SAVE: a method for dimension reduction and graphics in regression. Comm Statist Theory Methods, 29: 2109–2121 (2000)

    Article  MATH  Google Scholar 

  14. Zhao J L, Xu X Z, Ma J J. Extending SAVE and PHD. Comm Statist Theory Methods, 36: 1591–1606 (2007)

    Article  MATH  MathSciNet  Google Scholar 

  15. Li Y X, Zhu L X. Asymptotics for sliced average variance estimation. Ann Statist, 35: 41–69 (2007)

    Article  MathSciNet  Google Scholar 

  16. Zhu L P, Zhu L X. On kernel method for sliced average variance estimation. J Multivariate Anal, 98: 970–991 (2007)

    Article  MATH  MathSciNet  Google Scholar 

  17. Wang H S, Ni L Q, Tsai C L. Improving dimension reduction via contour-projection. Statist Sinica, 18: 299–311 (2008)

    MATH  Google Scholar 

  18. Li K C. Nonlinear confounding in high-dimensional regression. Ann Statist, 25: 577–612 (1997)

    Article  MATH  MathSciNet  Google Scholar 

  19. Zhu Y, Zeng P. Fourier methods for estimating the central subspace and the central mean subspace in regression. J Amer Statist Assoc, 101: 1638–1651 (2006)

    Article  MATH  MathSciNet  Google Scholar 

  20. Xia Y. A Constructive approach to the estimation of dimension reduction directions. Ann Statist, 35: 2654–2690 (2007)

    Article  MATH  MathSciNet  Google Scholar 

  21. Li B, Wang S. On directional regression for dimension reduction. J Amer Statist Assoc, 102: 997–1008 (2007)

    Article  MathSciNet  Google Scholar 

  22. Prendergast L A. Implications of influence function analysis for sliced inverse regression and sliced average variance estimation. Biometrika, 94: 585–601 (2007)

    Article  MATH  MathSciNet  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to JunLong Zhao.

Additional information

This work was supported by National Natural Science Foundation of China (Grant No. 10771015)

Rights and permissions

Reprints and permissions

About this article

Cite this article

Zhao, J., Xu, X. Dimension reduction based on weighted variance estimate. Sci. China Ser. A-Math. 52, 539–560 (2009). https://doi.org/10.1007/s11425-008-0130-z

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11425-008-0130-z

Keywords

MSC(2000)

Navigation