Abstract
Under weak dependence, a minimum distance estimate is obtained for a smooth function and its derivatives in a regression-type framework. The upper bound of the risk depends on the Kolmogorov entropy of the underlying space and the mixing coefficient. It is shown that the proposed estimates have the same rate of convergence, in the L 1-norm sense, as in the independent case.
Similar content being viewed by others
References
Beran, R. J. (1977). Minimum Hellinger distance estimates for parametric models, Ann. Statist., 5, 445–463.
Chanda, K. C. (1974). Strong mixing properties of linear stochastic processes, J. Appl. Probab., 14, 67–77.
Devroye, L. P. and Wagner, T. J. (1980). Distribution-free consistency results in nonparametric discrimination and regression function estimation, Ann. Statist., 8, 231–239.
Ibragimov, I. A. and Khas'minskii, R. Z. (1980). On nonparametric estimation of regression, Soviet Math. Dokl., 21, 130–131.
Kolmogorov, A. N. and Tikhomirov, V. M. (1961). ε-entropy and ε-capacity of sets of function spaces, Amer. Math. Soc. Transl., 17, 277–364 (translation).
LeCam, L. M. (1986). Asymptotic Methods in Statistical Decision Theory, Springer, New York.
Pham, D. T. (1986). The mixing property of linear and generalized random coefficient autoregressive models, Stochastic Process. Appl. 23, 291–300.
Pham, D. T. and Tran, L. T. (1985). Some strong mixing properties of time series models, Stochastic Process. Appl., 19, 297–303.
Robinson, P. M. (1986). On the consistency and finite-sample properties of nonparametric kernel time series regression, autoregression and density estimators, Ann. Inst. Statist. Math., 38, 539–549.
Roussas, G. G. (1990). Nonparametric regression estimation under mixing conditions, Stochastic Process. Appl., 36, 107–116.
Roussas, G. G. and Ioannides, D. (1987). Moment inequalities for mixing sequences of random variables, Stochastic Anal. Appl., 5, 61–120.
Roussas, G. G. and Ioannides, D. (1988). Probability bounds for sums in triangular arrays of random variables under mixing conditions, Statistical Theory and Data Analysis II (ed. K.Matusita), Elsevier Science (North Holland), New York.
Stone, C. J. (1980). Optimal rates of convergence for nonparametric estimators, Ann. Statist., 8, 1348–1360.
Stone, C. J. (1982). Optimal global rates of convergence for nonparametric estimators, Ann. Statist., 10, 1040–1053.
Tran, L. T. (1989). The L 1-convergence of kernel density estimates under dependence, Canad. J. Statist., 17, 197–208.
Tran, L. T. (1990). Kernel density estimation under dependence, Statist. Probab. Lett., 10, 193–201.
Tran, L. T. (1993). Nonparametric functional estimation for time series by local average estimators, Ann. Statist., 40, 1040–1057.
Wolfowitz, J. (1957). The minimum distance method, Ann. Math. Statist. 28, 75–88.
Yatracos, Y. G. (1985). Rates of convergence of minimum distance estimators and Kolmogorov's entropy, Ann. Statist. 13, 768–774.
Yatracos, Y. G. (1989a). A regression type problem, Ann. Statist., 17, 1597–1607.
Yatracos, Y. G. (1989b). On the estimation of the derivatives of a function with the derivatives of an estimate. J. Multivariate Anal., 28, 172–175.
Yatracos, Y. G. (1992). L 1-optimal estimates for a regression type function in, J. Multivariate Anal., 40, 213–220.
Yoshihara, Ken-ichi (1992). Weakly Dependent Stochastic Sequences and Their Applications, Vol. 1, Sanseido, Tokyo.
Author information
Authors and Affiliations
Additional information
This work was partially supported by a research grant from the Natural Sciences and Engineering Research Council of Canada.
About this article
Cite this article
Roussas, G.G., Yatracos, Y.G. Minimum distance regression-type estimates with rates under weak dependence. Ann Inst Stat Math 48, 267–281 (1996). https://doi.org/10.1007/BF00054790
Received:
Revised:
Issue Date:
DOI: https://doi.org/10.1007/BF00054790