Abstract
Because of its orthogonality, interpretability and best representation, functional principal component analysis approach has been extensively used to estimate the slope function in the functional linear model. However, as a very popular smooth technique in nonparametric/semiparametric regression, polynomial spline method has received little attention in the functional data case. In this paper, we propose the polynomial spline method to estimate a partial functional linear model. Some asymptotic results are established, including asymptotic normality for the parameter vector and the global rate of convergence for the slope function. Finally, we evaluate the performance of our estimation method by some simulation studies.
Similar content being viewed by others
References
Akaike H (1974) A new look at the statistical model identification. IEEE Trans Autom Control 19:716–723
Aneiros-Pérez G, Vieu P (2006) Semi-functional partial linear regression. Stat Probab Lett 76:1102–1110
Aneiros-Pérez G, Vieu P (2008) Nonparametric time series prediction: a semi-functional partial linear modeling. J Multivar Anal 99:834–857
Baíllo A, Grané A (2009) Local linear regression for functional predictor and scalar response. J Multivar Anal 100:102–111
Burba F, Ferraty F, Vieu P (2009) K-nearest neighbour method in functional nonparametric regression. J Nonparametric Stat 21:453–469
Cai TT, Hall P (2006) Prediction in functional linear regression. Annals Stat 34:2159–2179
Cai Z, Fan J, Yao Q (2000) Functional-coefficient regression model for nonlinear time series. J Am Stat Assoc 95:941–956
Cardot H, Ferraty F, Sarda P (1999) Functional linear model. Stat Prob Lett 45:11–22
Cardot H, Ferraty F, Sarda P (2003) Spline estimators for the functional linear model. Stat Sin 13:571–591
Cardot H, Sarda P (2008) Varying-coefficient functional linear regression models. Commun Stat Theory Methods 37:3186–3203
Chen H (1991) Polynomial splines and nonparametric regression. J Nonparametric Stat 1:143–156
Crambes C, Kneip A, Sarda P (2009) Smoothing splines estimators for functional linear regression. Annals Stat 37:35–72
de Boor C (2001) A practical guide to splines. Springer, New York
DeVore RA, Lorentz GG (1993) Constructive approximation. Springer, Berlin
Ferraty F, Vieu P (2006) Nonparametric functional data analysis. Springer, New York
Huang JZ (2003a) Asymptotics for polynomial spline regression under weak conditions. Stat Probab Lett 65:207–216
Huang JZ (2003b) Local asymptotics for polynomial spline regression. Annals Stat 31:1600–1635
Huang JZ, Wu CO, Zhou L (2004a) Polynomial spline estimation and inference for varying coefficient model with longitudinal data. Stat Sin 14:763–788
Huang JZ, Shen H (2004b) Functional coefficient regression models for non-linear time series: a polynomial spline approach. Scand J Stat 31:515–534
Hall P, Horowitz J (2007) Methodology and convergence rates for functional linear regression. Annals Stat 35:70–91
Li Y, Hsing T (2007) On rates of convergence in functional linear regression. J Multivar Anal 98:1782–1804
Lian H (2011) Functional partial linear model. J Nonparametric Stat 23:115–128
Newey WK (1997) Convergence rates and asymptotic normality for series estimators. J Econom 79:147–168
Ramsay J, Dalzell C (1991) Some tools for functional data analysis. J R Stat Soc Ser B 53:539–572
Ramsay J, Silverman B (1997) Functional data analysis. Springer, New York
Ramsay J, Silverman B (2005) Functional data analysis, 2nd edn. Springer, New York
Ramsay J, Hooker G, Graves S (2009) Functional data analysis with R and Matlab. Springer, New York
Rice JA, Silverman BW (1991) Estimating the mean and covariance structure nonparametrically when the data are curves. J R Stat Soc Ser B 53:233–243
Shin H (2009) Partial functional linear regression. J Stat Plan Inference 139:3405–3418
Stone CJ (1994) The use of polynomial splines and their tensor products in multivariate function estimation (with discussion). Annals Stat 22:118–171
Stone CJ, Hansen M, Kooperberg C, Truong YK (1997) Polynomial splines and their tensor products in extended linear modeling (with discussion). Annals Stat 25:1371–1470
Schwarz G (1978) Estimating the dimension of a model. Annals Stat 6:461–464
Zhang D, Lin X, Sowers M (2007) Two-stage functional mixed models for evaluating the effect of longitudinal covariate profiles on a scalar outcome. Biometrics 63:351–362
Zhou S, Shen X, Wolfe DA (1998) Local asymptotics for regression splines and confidence regions. Annals Stat 26:1760–1782
Zhou J, Chen M (2012) Spline estimators for semi-functional linear model. Stat Probab Lett 82:505–513
Acknowledgments
The work was supported by National Nature Science Foundation of China (Grant Nos. 10961026, 11171293, 11225103, 11301464), the PH.D. Special Scientific Research Foundation of Chinese University (20115301110004), the Key Fund of Yunnan Province (Grant No. 2010CC003) and the Scientific Research Foundation of Yunnan Provincial Department of Education (No. 2013Y360). We are grateful to the referees and the editors for their constructive remarks that greatly improved the manuscript.
Author information
Authors and Affiliations
Corresponding author
Appendix
Appendix
In the appendix, we give the proofs of the theorems and corollary in Sect. 3.
Set \(B_{s}={K_n}^{1/2}N_{s}^{b}, s=1,\ldots , K_n\), where \(N_{s}^{b}\) are the normalized B-splines. From the Theorem 4.2 of Chapter 5 of DeVore and Lorentz (1993), we have that for any spline function \(\sum _{s=1}^{K_n} b_s B_{s}\), there are positive constants \(M_1\) and \(M_2\) such that
where \(\Vert \cdot \Vert _2\) is Euclidean norm. Let \(\Vert r\Vert _{\infty }=\sup \nolimits _{x\in [0,1]} |r(x)|\).
In order to prove the theorems, we need the following two lemmas.
Lemma 1
If conditions (C1) and (C2) hold, then we have
-
(i)
$$\begin{aligned} \sup _{a\in S_{k,N_n}}\Big |\frac{\frac{1}{n}\sum _{i=1}^n \langle X_i,a \rangle ^2}{E \langle X,a \rangle ^2}-1\Big |=o_p(1). \end{aligned}$$
-
(ii)
there exists an interval \([M_3,M_4],0<M_3<M_4<\infty \) such that as \(n\rightarrow \infty \),
$$\begin{aligned} P\Big \{\text {all the eigenvalues of}~\frac{1}{n}B^TB~\text {fall in}~[M_3,M_4]\Big \}\rightarrow 1. \end{aligned}$$
Note that the Lemma 1 is a generalization of Lemma 1 and 2 in Huang and Shen (2004b) in functional data case. We give a brief proof in the following.
Proof
(i) Let \(\Gamma _n\) denote the empirical versions of operator \(\Gamma \), that is,
By the Cauchy–Schwarz inequality, condition (C2) and (28) in Cardot et al. (2003), we have
Then for an arbitrary constant \(\epsilon >0\), by Lemma 5.2 in Cardot et al. (1999), we have
together with (C2), which gives the result.
(ii) Let \(b=(b_1,\ldots ,b_{K_n})^T, a=\sum _{s=1}^{K_n} b_sB_s\). It follows from (i) that except an event whose probability tends to zero as \(n\rightarrow \infty \),
By the Cauchy–Schwarz inequality, (28) in Cardot et al. (2003) and (7),
Thus, except an event whose probability tends to zero, \(\frac{1}{n}b^TB^TBb\asymp \Vert b\Vert _2^2,\) holds uniformly for all b, which yields the result. \(\square \)
Lemma 2
Under conditions (C1)–(C5), as \(n\rightarrow \infty \), we have
Proof
Let \(\mu _j(X_i)=E(Z_{ij}|X_i)=\langle X_i,g_j\rangle , \eta _{ij}=Z_{ij}-\mu _j(X_i)\),
We also define \(V=(\tilde{V_1},\ldots ,\tilde{V_p})\), \(\eta =(\tilde{\eta _1},\ldots ,\tilde{\eta _p})\). Then, \(\mathbf{Z}=\eta +V\) and
For the (j, l)th element of \(I_1\)
By independence and the Cauchy–Schwarz inequality, we have
Further, by \(C_r\) inequality and (C2)–(C4), we have
Thus,
Note that \(A\ge 0\), then we have
By Lemma 1, we can know that except an event whose probability tends to zero,
Also note that \(E\langle X_i,B_s\rangle \eta _{ij}=E\langle X_i,B_s\rangle E(\eta _{ij}|X_i)=0\). Then, by (7) and conditions (C2)–(C4), we have there exists a positive constant C such that
Thus, for \(j,l=1,\ldots ,p\),
which together with (8) yields
For the (j, l)-th element of \(I_4\), \(j,l=1,\ldots ,p\),
by Cauchy–Schwartz inequality,
It follows from Theorem XII.1 of de Boor (2001) that there exist positive constant \(C_j\) and spline function \(g_j^*\in S_{k,N_n}, j=1,\ldots ,p\) such that
Set \(g_j^*=\sum _{s=1}^{K_n} b_{js}^*B_{s}, \quad b_j^*=(b_{j1}^*,\ldots ,b_{jK_n}^*)^T,\quad j=1,\ldots ,p\), then,
As A is an orthogonal projection matrix,
From the above results and (C1), we have
that is,
For the (j, l)-th element of \(I_2\) and \(I_3\), \(j,l=1,\ldots ,p\), we have
Using (9) and (10), we can infer that
The combination of (9)–(11) allows us to finish the proof of Lemma 2. \(\square \)
Proof of Theorem 1
Denote \(\Phi =\Big (\langle X_1,\alpha \rangle ,\ldots ,\langle X_n,\alpha \rangle \Big )^T\), \(\varepsilon =(\varepsilon _1,\ldots ,\varepsilon _1)^T\). Then, \(Y=\mathbf{Z}\beta +\Phi +\varepsilon \). We can write
Observe that
For \(\Delta _{11}\), as \(\mathbf{Z}=\eta +V\),
By (C4) and the Theorem XII.1 of de Boor (2001), we know that there is a spline function \(\alpha ^*=\sum _{s=1}^{K_n} b_s^*B_s\in S_{k,N_n}\) and positive constant C such that
Set \(\Phi ^*=(\langle X_1,\alpha ^*\rangle ,\ldots ,\langle X_n,\alpha ^*\rangle )^T\) and \(b^*=(b_1^*,\ldots ,b_{K_n}^*)^T\), we have \(\Phi ^*=Bb^*\). For \(j=1,\ldots ,p\), by conditions (C1), (C2), (C4) and Theorem XII.1 of de Boor (2001), we can infer
Thus, by (C1) we have
Observe that for \(j=1,\ldots ,p\),
As \(E\Big (\eta _{ij}\langle X_i,\alpha -\alpha ^*\rangle \Big )=E\Big [\langle X_i,\alpha -\alpha ^*\rangle E(\eta _{ij}|X_i)\Big ]=0\) and
we can infer
Further, under the Lemma 1, (C1) and (13), we can show
By (12), (14)–(16) and Lemma 2, we have
\(\Delta _{21}\) can be expressed as
Let \(\epsilon _i=\eta _i \varepsilon _i\). Since \(\varepsilon _i\) is independent of \((X_i,\mathbf{Z}_i)\) and \((X_i,\mathbf{Z}_i,Y_i)\) is i.i.d. sequence, the \(\epsilon _i\) are i.i.d. random variables with \(E\epsilon _i=0\) and \(Var(\epsilon _i)=\sigma ^2\Sigma \).
Observe that
Then, by the central limit theorem,
Also note that
Then, it follows from Lemma 1 that
Since \(E\langle X_i,B_s\rangle \varepsilon _i\langle X_j,B_s\rangle \varepsilon _j=0, i\ne j\), we have
that is, \(\varepsilon ^TA\varepsilon =O_p(K_n)\). In addition, we can know from the proof of Lemma 2 that
Thus,
which, together with (18), yields
For the jth element of \(R_2\), \(j=1,\ldots ,p\), we have
Since \(\varepsilon _i\) is independent \((X_i,\mathbf{Z}_i)\), we have
Then,
Also, observe that
Then, by (C1), we have
From the above results, we can infer
Now, by Lemma 2, (17), (19), (20) and Slutsky theorem, we can obtain the Theorem 1. \(\square \)
Proof of Theorem 2
Observe that
Let \(\tilde{Y}=\mathbf{Z}(\beta -\widehat{\beta })+\Phi \). Denote \(\tilde{b}=(B^TB)^{-1}B^T\tilde{Y}\) and \(\tilde{\alpha }(t)=\sum _{s=1}^{K_n} \tilde{b_s}B_s(t)\), where \(\tilde{b}=(\tilde{b_1},\ldots ,\tilde{b_{K_n}})^T\). Then, \(\widehat{b}-\tilde{b}=(B^TB)^{-1}B^T\varepsilon \). By Lemma 1, we have
except on an event whose probability tends to zero as \(n\rightarrow \infty \). Thus, by (7), we can infer
Also, it follows from the Theorem XII.1 of de Boor (2001) that there exists a spline function \(\alpha ^*(t)=\sum _{s=1}^{K_n} b_s^*B_s(t)\in S_{k,N_n}\) where \(b^*=(b_1^*,\ldots ,b_{K_n}^*)^T\) and constant \(C>0\) such that
By the Theorem XII.1 of de Boor (2001) and (7), we have
Observe that \(B\tilde{b}=B(B^TB)^{-1}B^T\tilde{Y}\) and \(B(B^TB)^{-1}B^T\) is an orthogonal projection matrix. Thus,
Applying (C2), (22) and the Cauchy–Schwarz inequality, we obtain that
that is,
In addition, note that
Then, it follows from Theorem 1 and (C4) that
which together with (23) yields
Further, we can infer that
Then, the combination of (21), (22), (25) and (26) allows us to complete the proof of Theorem 2. \(\square \)
Proof of Theorem 3
We can write
Observe that
Then, by (C1), (C2), theorem 2, we have
It follows from (24) that
For \(R_{n3}\), since \(E(\varepsilon _1^2-\sigma ^2)=0\) and \(\Lambda ^2=E(\varepsilon _1^2-\sigma ^2)^2<\infty \), it follows from the central limit theorem that
For \(R_{n4}\), we have
Then, applying (C1), (C2) and Theorem 2, we can obtain
Note that
Thus, using (C3) and Theorem 1, we have
Also, observe that
Then, by (C1)–(C3), Theorem 1 and 2, we can get
Finally, using (27)–(32), we can complete the proof of Theorem 3. \(\square \)
Proof of Corollary 1
It follows form Theorem 1 that
Also, by Lemma 2 and Theorem 3, we have that
Then, by the Slutsky theorem, we obtain the Corollary 1. \(\square \)
Rights and permissions
About this article
Cite this article
Zhou, J., Chen, Z. & Peng, Q. Polynomial spline estimation for partial functional linear regression models. Comput Stat 31, 1107–1129 (2016). https://doi.org/10.1007/s00180-015-0636-0
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00180-015-0636-0