Abstract
Random vectors, their expectation and dispersion matrix are reviewed. The concept of estimability in a linear model is introduced and the form of the best linear unbiased estimate of an estimable function is derived. The full rank case of the result, which is the Gauss–Markov Theorem, is stated. The Hadamard inequality for the determinant of a positive semidefinite matrix is proved and its application to weighing designs is discussed. An unbiased estimate for the error variance in terms of the residual sum of squares is obtained. The special case of one-way classification is described in greater detail. A general linear model with a possibly singular variance-covariance matrix is considered and the best linear unbiased estimate of an estimable function as well as an estimate of the error variance are obtained.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsAuthor information
Authors and Affiliations
Rights and permissions
Copyright information
© 2012 Springer-Verlag London Limited
About this chapter
Cite this chapter
Bapat, R.B. (2012). Linear Estimation. In: Linear Algebra and Linear Models. Universitext. Springer, London. https://doi.org/10.1007/978-1-4471-2739-0_7
Download citation
DOI: https://doi.org/10.1007/978-1-4471-2739-0_7
Publisher Name: Springer, London
Print ISBN: 978-1-4471-2738-3
Online ISBN: 978-1-4471-2739-0
eBook Packages: Mathematics and StatisticsMathematics and Statistics (R0)