Abstract
On the one hand, nonparametric regression approaches are flexible modeling tools in modern statistics. On the other hand, the lack of any parameters makes these approaches more challenging when assessing some statistical inference in these models. This is crucial especially in situations when one needs to perform some statistical tests or to construct some confidence sets. In such cases, it is common to use a bootstrap approximation instead. It is an effective alternative to more straightforward but rather slow plug-in techniques. In this contribution, we introduce a proper bootstrap algorithm for a robustified version of the nonparametric estimates, the so-called M-smoothers or M-estimates, respectively. We distinguish situations for homoscedastic and heteroscedastic independent error terms, and we prove the consistency of the bootstrap approximation under both scenarios. Technical proofs are provided and the finite sample properties are investigated via a simulation study.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Antoch, J., & Janssen P. (1989). Nonparametric regression m-quantiles. Statistics & Probability Letters, 8, 355–362.
Boente, G., Ruiz, M., & Zamar, R. (2010). On a robust local estimator for the scale function in heteroscedastic nonparametric regression. Statistics & Probability Letters, 80, 1185–1195.
Fan, J., & Gijbels, I. (1995). Local polynomial modelling and its applications (1st ed.). Boca Raton, FL: Chapman & Hall
Hall, P., Kay, J., & Titterington, D. (1990). Asymptotically optimal difference-based estimation of variance in nonparametric regression. Biometrica, 77, 521–528.
Härdle, W., & Gasser, T. (1984). Robust nonparametric function fitting. Journal of the Royal Statistical Society, Series B, 46, 42–51.
Härdle, W. K., & Marron, J. S. (1991). MBootstrap simultaneous error bars for nonparametric regression. The Annals of Statistics, 19(2), 778–796.
Hušková, M., & Maciak, M. (2017). Discontinuities in robust nonparametric regression with α-mixing dependence. Journal Nonparametric Statistics, 29(2), 447–475.
Hwang, R. C. (2002). A new version of the local constant M-smoother. Communications in Statistics: Theory and Methods, 31, 833–848
Leung, D. (2005). Cross-validation in nonparametric regression with outliers. Annals of Statistics, 33, 2291–2310.
Leung, D., Marriott, F., & Wu, E. (1993). Bandwidth selection in robust smoothing. Journal of Nonparametric Statistics, 2, 333–339.
Maciak, M. (2011). Flexibility, Robustness, and Discontinuity in Nonparametric Regression Approaches. Ph.D. thesis, Charles University, Prague.
Müller, H., & Stadtmüller, U. (1987). Estimation of heteroscedasticity in regression analysis. Annals of Statistics, 15, 610–625.
Nadaraya, E. (1964). On estimating regression. Theory Probability Applications, 9, 141–142.
Neumeyer, N. (2006). Bootstrap Procedures for Empirical Processes of Nonparametric Residuals. Habilitationsschrift, Ruhr-University Bochum, Germany.
Rue, H., Chu, C., Godtliebsen, F., & Marron, J. (1998). M-smoother with local linear fit. Journal of Nonparametric Statistics, 14, 155–168.
Serfling, R. (1980). Approximation theorems of mathematical statistics. New York: Wiley.
Watson, G. (1964). Smooth regression analysis. The Indian Journal of Statistics, Series A, 26, 359–372.
Acknowledgements
The author’s research was partly supported by the Grant P402/12/G097.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Appendix
Appendix
In this section we provide some technical details and the proof of the bootstrap consistency result stated in Theorem 2. Let \(\{(X_{i}, Y_{i}^{\star });~i = 1, \dots , n\}\) be the bootstrapped data where \(Y_{i}^{\star } = \widehat {m}(X_{i}) + \widehat {\sigma }(X_{i})\varepsilon _{i}^{\star }\), for \(\widehat {m}(X_{i})\) being the M-smoother estimate of m(X i), \(\widehat {\sigma }(X_{i})\) the estimate of σ(X i) in sense of (5), and the random error terms \(\{\varepsilon _{i}^{\star }\}_{i =1 }^{n}\) are defined in B3 step of the bootstrap algorithm in Sect. 3. Then, we can obtain the bootstrapped version of \(\widehat {m}(x)\), for some x ∈ (0, 1), given as a solution of the minimization problem
where \(\widehat {\boldsymbol {\beta }}_{x}^{\star } = (\widehat {\beta }_{0}^{\star }, \dots , \widehat {\beta }_{p}^{\star })^{\top }\), and \(\widehat {m}^{\star }(x) = \widehat {\beta }_{0}^{\star }\). Using the smoothness property of m(⋅) we can apply the Taylor expansion of the order p and given the model definition in (1) we can rewrite the minimization problem as an equivalent problem given by the following set of equations:
for ℓ = 0, …, p, where ψ = ρ′. Next, for any ℓ ∈{0, …, p} and \(\boldsymbol {b} \in \mathbb {R}^{p + 1}\) let us define an empirical process M n(b, ℓ) and its bootstrap counterpart \(M_{n}^{\star }(\boldsymbol {b}, \ell )\) as follows:
and
where for brevity we used the notation \(\xi _{i}^{\ell }(x) = \Big (\frac {X_{i} - x}{h_{n}}\Big )^{\ell }\). We need to investigate the behavior of \(M_{n}^{\star }(\boldsymbol {b}, \ell )\), conditionally on the sample {(X i, Y i); i = 1, …, n)}, and we will compare it with the behavior of M n(b, ℓ).
Let G ⋆(⋅) be the distribution function of the bootstrap residuals \(\{\varepsilon _{i}^{\star }\}_{i = 1}^{n}\) defined in B3. It follows from the definition that
where ϕ(⋅) and Φ(⋅) stand for the density and the distribution function of Z i’s, which are assumed to be normally distributed with zero mean and unit variance. It is easy to verify that G ⋆(⋅) is continuous, symmetric, and moreover, it satisfies Assumption A.2. Thus, for E ⋆ being the conditional expectation operator when conditioning by the initial data sample, we obtain the following:
where we used the symmetric property of the distribution function G ⋆. Next, we obtain
where the last term can be shown to be asymptotically negligible due to the properties of ψ(⋅) and the fact that \(\sup _{x \in \mathbb {R}} |G^{\star }(x) - G(x)| \to 0\) in probability (see Lemma 2.19 in [14]). For (9) we can use the Hölder property of λ G(⋅) (Assumption A.4) and we get that
and
where the first equation follows from the fact that \(\widehat {\sigma }(X_{i})\) is a consistent estimate of σ(X i) and the second follows from the fact that |X i − x|≤ h n. Both equations hold almost surely.
Putting everything together we obtain that
and, moreover, repeating the same steps also for the second moment \(E^{\star } \big [M_{n}^{\star }(\boldsymbol {b}, \ell )\big ]^2\) and applying (3) and (4) we also obtain that \(E^{\star } \big [M_{n}^{\star }(\boldsymbol {b}, \ell )\big ]^2 \to 0\) in probability.
To finish the proof we need the following lemma.
Lemma 1
Let the model in (1) hold and let Assumptions A.1–A.5 be all satisfied. Then the following holds:
where ℓ = 0, …, p, C > 0, ∥b∥ = |b 0| + ⋯ + |b p|, δ n = (nh n)−γ∕2, and γ ∈ (γ 0, 1], for some 0 < γ 0 ≤ 1.
Proof
Lemma 1 is a bootstrap version of Lemma 4 in [11] or a more general Lemma A.3 in [7]. The proof follows the same lines using the moment properties derived for \( M_n^\star (\delta _{n} \boldsymbol {b}, \ell )\). □
Lemma 4 in [11] allows us to express the classical M-smoother estimates \(\widehat {\beta }_{x}\) in terms of the asymptotic Bahadur representations as
while Lemma 1 allows us to express the bootstrapped counterparts \(\widehat {\beta }_{x}^\star \) in a similar manner as
where \(W_{n} = Diag\Bigg \{K\bigg (\frac {X_{1} - x}{h_{n}}\bigg ), \dots , K\bigg (\frac {X_{n} - x}{h_{n}}\bigg )\Bigg \}\), and \(X_{n} =\bigg (\big ( \frac {X_{i} - x}{h_{n}}\big )^j\bigg )_{i = 1, j = 0}^{n, p}\).
To finish the proof one just needs to realize that the sequences of random variables \(\{\xi _{n i}\}_{i = 1}^n\) and \(\{\xi _{n i}^\star \}_{i = 1}^n\) for \(\xi _{n i} = \frac {1}{\sqrt {n h_{n}}} \psi (\sigma (X_{i})\varepsilon _{i}) \big (\frac {X_{i} - x}{h_{n}}\big )^\ell K\big (\frac {X_{i} - x}{h_{n}}\big )\) and \(\xi _{n i}^\star = \frac {1}{\sqrt {n h_{n}}} \psi (\widehat {\sigma }(X_{i})\varepsilon _{i}^\star ) \big (\frac {X_{i} - x}{h_{n}}\big )^\ell K\big (\frac {X_{i} - x}{h_{n}}\big )\) both comply with the assumptions of the central limit theorem for triangular schemes and thus, random quantities \(\sum _{i = 1}^n \xi _{n i}\) and \(\sum _{i = 1}^n \xi _{n i}^\star \) both converge in distribution, conditionally on X i’s and the original data {(X i, Y i); i = 1, …, n}, respectively, to the normal distribution with zero mean and the same variance parameter. \(\square \)
Rights and permissions
Copyright information
© 2018 Springer Nature Switzerland AG
About this paper
Cite this paper
Maciak, M. (2018). Bootstrapping Nonparametric M-Smoothers with Independent Error Terms. In: Bertail, P., Blanke, D., Cornillon, PA., Matzner-Løber, E. (eds) Nonparametric Statistics. ISNPS 2016. Springer Proceedings in Mathematics & Statistics, vol 250. Springer, Cham. https://doi.org/10.1007/978-3-319-96941-1_16
Download citation
DOI: https://doi.org/10.1007/978-3-319-96941-1_16
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-96940-4
Online ISBN: 978-3-319-96941-1
eBook Packages: Mathematics and StatisticsMathematics and Statistics (R0)