Skip to main content
Log in

On the performance of the ordinary least squares method under an error component model

  • Published:
Metrika Aims and scope Submit manuscript

Abstract

Starting from the one-dimensional results by Wang et al (1994) we consider the performance of the ordinary least squares estimator in comparison to the best linear unbiased estimator under an error component model with random effects in units and time. Upper bounds are derived for the first-order approximation to the difference between both estimators and for the spectral norm of the difference between their dispersion matrices.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

  • Baksalary JK, Kala R (1980) A new bound for the Euclidean norm of the difference between the least squares and the best linear unbiased estimators. Ann. Statist. 8: 679–681

    MATH  MathSciNet  Google Scholar 

  • Bloomfield P, Watson GS (1975) The inefficiency of least squares. Biometrika 62: 121–128

    Article  MATH  MathSciNet  Google Scholar 

  • Christensen R (1984) A note on ordinary least squares methods for two-stage sampling. J. Amer. Statist. Assoc. 79: 720–721

    Article  Google Scholar 

  • Fuller WA, Battese GF (1973) Transformations for estimation of linear models with nested-error structure. J. Amer. Statist. Assoc 68: 626–632

    Article  MATH  MathSciNet  Google Scholar 

  • Hannan RJ (1970) Multiple time series. Wiley, New York

    MATH  Google Scholar 

  • Knott M (1975) On the minimum efficiency of least squares. Biometrika 62: 129–132

    Article  MATH  MathSciNet  Google Scholar 

  • Krämer W (1980) Finite sample efficiency of ordinary least squares in the linear regression model with autocorrelated errors. J. Amer. Statist. Assoc. 75: 1005–1009

    Article  MATH  MathSciNet  Google Scholar 

  • Magness TA, McGuire JB (1962) Comparison of least squares and minimum variance estimates of regression parameters. Ann. Math. Statist. 33: 462–470

    MathSciNet  Google Scholar 

  • Magnus JR, Neudecker H (1988) Matrix differential calculus with applications in statistics and econometrics. Wiley, Chichester

    MATH  Google Scholar 

  • Rao CR (1973) Linear statistical inference and its applications. 2nd ed. Wiley, New York

    MATH  Google Scholar 

  • Schwabe R (1996) Optimum designs for multi-factor models. Lecture Notes in Statistics 113 Springer, New York

    MATH  Google Scholar 

  • Scott AJ, Holt D (1982) The effect of two-stage sampling on ordinary least squares methods. J. Amer. Statist. Assoc. 77: 848–854

    Article  MATH  Google Scholar 

  • Wang SG, Chow SC, Tse SK (1994) On ordinary least-squares methods for sample surveys. Statist. Probab. Letters 20: 173–182

    Article  MATH  MathSciNet  Google Scholar 

  • Zyskind G (1967) On canonical forms, non-negative covariance matrices and best and simple least squares linear estimators in linear models. Ann. Math. Statist 38: 1092–1109

    MathSciNet  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Parimal Mukhopadhyay.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Mukhopadhyay, P., Schwabe, R. On the performance of the ordinary least squares method under an error component model. Metrika 47, 215–226 (1998). https://doi.org/10.1007/BF02742874

Download citation

  • Received:

  • Issue Date:

  • DOI: https://doi.org/10.1007/BF02742874

Key Words

Navigation