Abstract
Least squares estimation of parameters is considered. Sufficient conditions of strong consistency are obtained for a regression model in the presence of random errors.
Similar content being viewed by others
References
A. Albert, Regression, the Pseudoinverse, and Recursive Estimation [Russian translation], Moscow (1977).
N. F. Kirichenko and A. S. Slabospitskii, "Minimax filters in state estimation, parameter identification, and image recognition problems," Kibern. Vychisl. Tekh., No. 65, 23–33 (1985).
V. V. Petrov, "On the order of increase of the sum of dependent random variables," Teor. Veroyatn. Primen.,18, No. 2, 358–361 (1973).
A. S. Slabospitskii, "On energy interpretation of the convergence conditions of estimation operators," Kibern. Vychisl. Tekh., No. 73, 41–44 (1987).
T. W. Anderson and J. B. Taylor, "Strong consistency of least squares estimates in normal linear regressions," Ann. Stat.,4, No. 4, 488–490 (1976).
F. Eicker, "Asymptotic normality and consistency of the least squares estimators for families of linear regressions," Ann. Math. Stat.,34, No. 2, 447–456 (1963).
T. C. Hsia, System Identification. Least Squares Methods, Toronto (1977).
L. Ljung and T. Söderström, Theory and Practice of Recursive Identification, Cambridge (1983).
Additional information
Kiev University. Translated from Vychislitel'naya i Prikladnaya Matematika, No. 75, pp. 38–43, 1991.
Rights and permissions
About this article
Cite this article
Slabospitskii, A.S. Convergence conditions of the least squares method. J Math Sci 72, 3076–3079 (1994). https://doi.org/10.1007/BF01259474
Received:
Issue Date:
DOI: https://doi.org/10.1007/BF01259474