Abstract
Until this point, the only difficulties with least squares estimation that we have considered have been associated with violations of Gauss-Markov conditions, These conditions only assure us that least squares estimates will be ‘best’ for a given set of independent variables; i.e., for a given X matrix. Unfortunately, the quality of estimates, as measured by their variances, can be seriously and adversely affected if the independent variables are closely related to each other. This situation, which (with a slight abuse of language) is called multicollinearity, is the subject of this chapter and is also the underlying factor that motivates the methods treated in Chapters 11 and 12.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 1990 Springer-Verlag New York Inc.
About this chapter
Cite this chapter
Sen, A., Srivastava, M. (1990). Multicollinearity. In: Regression Analysis. Springer Texts in Statistics. Springer, New York, NY. https://doi.org/10.1007/978-1-4612-4470-7_10
Download citation
DOI: https://doi.org/10.1007/978-1-4612-4470-7_10
Publisher Name: Springer, New York, NY
Print ISBN: 978-1-4612-8789-6
Online ISBN: 978-1-4612-4470-7
eBook Packages: Springer Book Archive