Skip to main content
Log in

SiZer for smoothing splines

  • Published:
Computational Statistics Aims and scope Submit manuscript

Abstract

Smoothing splines are an attractive method for scatterplot smoothing. The SiZer approach to statistical inference is adapted to this smoothing method, named SiZerSS. This allows quick and sure inference as to “which features in the smooth are really there” as opposed to “which are due to sampling artifacts”, when using smoothing splines for data analysis. Applications of SiZerSS to mode, linearity, quadraticity and monotonicity tests are illustrated using a real data example. Some small scale simulations are presented to demonstrate that the SiZerSS and the SiZerLL (the original local linear version of SiZer) often give similar performance in exploring data structure but they can not replace each other completely.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Figure 1
Figure 2
Figure 3
Figure 4
Figure 5
Figure 6
Figure 7
Figure 8
Figure 9
Figure 10

Similar content being viewed by others

References

  1. Chaudhuri, P. and Marron, J. S. (1999), SiZer for exploration of structure in curves, Journal of the American Statistical Association, 94, 807–823.

    Article  MathSciNet  Google Scholar 

  2. Eubank, R. L. (2000), Nonparametric Regression and Spline Smoothing, Marcel Dekker, New York.

    MATH  Google Scholar 

  3. Fan, J. and Gijbels, I. (1996), Local Polynomial Modelling and Its Applications, Chapman and Hall, London.

    MATH  Google Scholar 

  4. Fan, J. and Marron, J. S. (1994), Fast implementations of nonparametric curve estimators, Journal of Computational and Graphical Statistics, 3, 35–56.

    Google Scholar 

  5. Green, P. J. and Silverman, B. W. (1994), Nonparametric Regression and Generalized Linear Models, Chapman and Hall, London.

    Book  Google Scholar 

  6. Härdle, W. (1990), Applied Nonparametric Regression, Cambridge University Press, Boston.

    Book  Google Scholar 

  7. Hastie, T.J. and Tibshirani, R. J. (1990), Generalized Additive Models, Chapman and Hall, London.

    MATH  Google Scholar 

  8. Loader, C. (1999), Local Regression and Likelihood, Springer Verlag, Berlin.

    MATH  Google Scholar 

  9. Marron, J. S. (1996), A personal view of smoothing and statistics, in Statistical Theory and Computational Aspects of Smoothing, eds. W. Härdie and M. Schimek, 1–9 (with discussion, and rejoinder 103–112).

    Google Scholar 

  10. Silverman, B. W. (1984), Spline smoothing: the equivalent kernel method. Ann. Statist., 12, 898–916.

    Article  MathSciNet  Google Scholar 

  11. Wahba, G. (1991), Spline Models for Observational Data, SIAM, Philadelphia.

    MATH  Google Scholar 

  12. Wand, M. P. and Jones, M. C. (1995), Kernel Smoothing, Chapman and Hall, London.

    Chapter  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Additional information

Marron’s research was supported by the Dept. of Stat. and Appl. Prob., National Univ. of Singapore, and by the National Science Foundation Grant DMS-9971649. Zhang’s research was supported by the National Univ. of Singapore Academic Research grant R-155-000-023-112. The Editor, the Associate Editor, and the referees are appreciated for their invaluable comments and suggestions that help improve the article significantly.

Appendix: Derivations of (5),(6), and (7)

Appendix: Derivations of (5),(6), and (7)

First of all, assume X1, X2, ⋯, Xn have been sorted so that X1 < X2 < < Xn. Write fi = f(Xi) and γi = f″(Xi) to be the values of f(x) and f″(x) at Xi for i = 1, 2, ⋯, n. Define f = (f1, ⋯, fn)T and γ = (γ2, ⋯, γn −1)T. Let hi = Xi+ 1Xi, i = 1, 2, ⋯, n − 1. Let Q : n × (n − 2), R: (n − 2) × (n − 2) and K: n × n be the matrices as defined in Green and Silverman (1994, pages 12–13). According to Theorem 2.1 of Green and Silverman (1994, page 13), f is a natural cubic spline with knots at Xi, i = 1, 2, ⋯, n if and only if

$$K=Q R^{-1} Q^{T}, \quad \gamma=R^{-1} Q^{T} \mathbf{f}, \quad \int f^{\prime \prime}(x)^{2} d x=\mathbf{f}^{T} K \mathbf{f}.$$
((11))

Simple calculation then leads to the following desired formula:

$$\hat{\mathbf{f}}=(W+\lambda K)^{-1} W \mathbf{Y} \equiv A_{\lambda} \mathbf{Y},$$
((12))

with the weight matrix W = diag(w1, w2, ⋯, wn), the hat matrix Aλ = (W + λK)−1 W, and the response vector Y = (Y1, Y2, ⋯, Yn)T.

Using (11) and (12), we are now ready to give the matrix formulas for computing \(\hat{f}_{\lambda}(x), \hat{f}_{\lambda}^{\prime}(x)\), and \(\widehat{sd}\left\{\hat{f}_{\lambda}^{\prime}(x)\right\}\) at a given grid of locations x = [x1, x2, ⋯, xN]T. By Green and Silverman (1994, pages 22–23), for any x, we can write \(\hat{f}(x)\) and \(\hat{f}{}^{\prime}(x)\) as linear combinations of \(\hat{\mathbf{f}}\) and \(\hat{\gamma}\). Let hi(x) = xXi, i = 1, 2, ⋯, n. When x < X1,

$$\hat{f}(x)=\hat{f}_{1}+h_{1}(x)\left\{\frac{\hat{f}_{2}-\hat{f}_{1}}{h_{1}}-\frac{h_{1}}{6} \hat{\gamma}_{2}\right\}, \quad \hat{f}^{\prime}(x)=\frac{\hat{f}_{2}-\hat{f}_{1}}{h_{1}}-\frac{h_{1}}{6} \hat{\gamma}_{2}.$$

When XixXi+ 1, let \(\delta_{i}(x)=[1+\frac{h_{i}(x)}{h_{i}}] \hat{\gamma}_{i+1}+[1-\frac{h_{i+1}(x)}{h_{i}}] \hat{\gamma}_{i}\) for some i = 1, 2, ⋯, n,

$$\begin{aligned} \hat{f}(x) &=\frac{h_{i}(x) \hat{f}_{i+1}-h_{i+1}(x) \hat{f}_{i}}{h_{i}}+\frac{h_{i}(x) h_{i+1}(x) \delta_{i}(x)}{6}, \\ \hat{f}^{\prime}(x) &=\frac{\hat{f}_{i+1}-\hat{f}_{i}}{h_{i}}+\frac{h_{i}(x) h_{i+1}(x)(\hat{\gamma}_{i+1}-\hat{\gamma}_{i})}{6 h_{i}}+\frac{\left[h_{i}(x)+h_{i+1}(x)\right] \delta_{i}(x)}{6} \end{aligned}$$

When x > Xn,

$$\hat{f}(x)=\hat{f}_{n}+\frac{h_{n}(x)}{6}\left\{\frac{\hat{f}_{n}-\hat{f}_{n-1}}{h_{n-1}}+h_{n-1} \hat{\gamma}_{n-1}\right\},$$
$$\hat{f}^{\prime}(x)=\frac{1}{6}\left\{\frac{\hat{f}_{n}-\hat{f}_{n-1}}{h_{n-1}}+h_{n-1} \hat{\gamma}_{n-1}\right\}.$$

It follows that \(\hat{f}(x)\) and \(\hat{f}^{\prime}(x)\) can be written respectively as \(c^{T} \hat{\mathbf{f}}-d^{T} \hat{\gamma}\) and \(\tilde{c}^{T} \hat{\mathbf{f}}-\tilde{d}^{T} \hat{\gamma}\) where \(c, \tilde{c}, d\)and \(\tilde{d}\) are coefficient vectors, depending on x and X1, X2, ⋯, Xn only. Let \(\hat{f}\left(x_{i}\right)=c_{i}^{T} \hat{\mathbf{f}}-d_{i}^{T} \hat{\gamma}, \hat{f}^{\prime}\left(x_{i}\right)=\tilde{c}_{i}^{T} \hat{\mathbf{f}}-\tilde{d}_{i}^{T} \hat{\gamma}, i=1,2, \cdots, n\). Define C = (c1, c2, ⋯, cn)T, D = (d1, d2, ⋯, dn)T, and define \(\tilde{C}\) and \(\tilde{D}\) similarly. Set \(\hat{\mathbf{f}}_{\mathbf{x}}=[\hat{f}(x_{1}), \cdots, \hat{f}(x_{N})]^{T}\) and \(\hat{\mathbf{f}}_{\mathbf{x}}^{\prime}=[\hat{f}^{\prime}(x_{1}), \cdots, \hat{f}^{\prime}(x_{N})]^{T}\). Then, using (11) and (12), we have

$$\begin{array}{l}{\hat{\mathbf{f}}_{\mathbf{x}}\quad =\quad C \hat{\mathbf{f}}-D \hat{\gamma}=[C-D R^{-1} Q^{T}] \hat{\mathbf{f}}=M \hat{\mathbf{f}}=M A_{\lambda} \mathbf{Y}}, \\ {\hat{\mathbf{f}}{}^{\prime}_{\mathbf{x}}\quad =\quad \tilde{C} \hat{\mathbf{f}}-\tilde{D} \hat{\gamma}=[\tilde{C}-\tilde{D} R^{-1} Q^{T}] \hat{\mathbf{f}}=\tilde{M} \hat{\mathbf{f}}=\tilde{M} A_{\lambda} \mathbf{Y}},\end{array}$$

where M = CDR− 1QT and \(\tilde{M}\) is similarly defined.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Marron, J.S., Zhang, J.T. SiZer for smoothing splines. Computational Statistics 20, 481–502 (2005). https://doi.org/10.1007/BF02741310

Download citation

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/BF02741310

Keywords

Navigation