Abstract
A semivarying coefficient model with measurement error in the nonparametric part was proposed by Feng and Xue (Ann Inst Stat Math 66:121–140, 2014), but its inferences have not been systematically studied. This paper applies empirical likelihood method to construct confidence regions/intervals for the regression parameter and coefficient function. When some auxiliary information about the parametric part is available, the empirical log-likelihood ratio statistic for the regression parameter is introduced based on the corrected local linear estimator of the coefficient function. Furthermore, corrected empirical log-likelihood ratio statistic for coefficient function is also investigated with the use of auxiliary information. The limiting distributions of the resulting statistics both for the regression parameter and coefficient function are shown to have standard Chi-squared distribution. Simulation experiments and a real data set are presented to evaluate the finite sample performance of our proposed method.
Similar content being viewed by others
References
Banerjee, A., Pitarakis, J.Y.: Functional cointegration: definition and nonparametric estimation. Stud. Nonlinear Dyn. Econom. 18, 507–520 (2014)
Carroll, R.J., Ruppert, D., Stefanski, L.A.: Measurement Error in Nonlinear Models. Chapman & Hall, London (1995)
Chen, J.H., Qin, J.: Empirical likelihood estimation for finite populations and the effective usage of auxiliary information. Biometrika 80, 10–116 (1993)
Cui, H.J., Li, R.C.: On parameter estimation for semi-linear error-in-variables models. J. Multivar. Anal. 64, 1–24 (1998)
Engle, R.F., Granger, W.J., Rice, J., Weiss, A.: Semiparametric estimates of the relation between weather and electricity sales. J. Am. Stat. Assoc. 80, 310–320 (1986)
Fan, G.L., Xu, H.X., Liang, H.Y.: Empirical likelihood inference for partially time-varying coefficient errors-in-variables models. Electron. J. Stat. 6, 1040–1058 (2012)
Fan, J., Huang, T.: Profile likelihood inferences on semiparametric varying-coefficient partially linear models. Bernoulli 11, 1031–1057 (2005)
Feng, S.Y., Xue, L.G.: Bias-corrected statistical inference for partially linear varying coefficient errors-in-variables models with restricted condition. Ann. Inst. Stat. Math. 66, 121–140 (2014)
Fuller, W.A.: Measurement Error Models. Wiley, New York (1987)
Härdle, W., Liang, H., Gao, J.T.: Partially Linear Models. Verlag, Heidelberg (2000)
Hastie, T.J., Tibshirani, R.: Varying-coefficient models. J. R. Stat. Soc. Ser. B 55, 757–796 (1993)
Huang, Z.S., Zhang, R.Q.: Profile empirical-likelihood inferences for the single-index-coefficient regression model. Stat. Comput. 23, 455–465 (2013)
Hwang, J.T.: The multiplicative errors-in-variables models with applications to the recent data released by the US Department of Energy. J. Am. Stat. Assoc. 81, 680–688 (1986)
Kai, B., Li, R., Zou, H.: New efficient estimation and variable selection methods for semiparametric varying-coefficient partially linear models. Ann. Stat. 39(1), 305–332 (2011)
Koenker, R.W.: Quantile Regression. Cambridge University Press, Cambridge (2005)
Liang, H., Hädle, W., Carroll, R.J.: Estimation in a semiparametric partially linear errors-in-variables model. Ann. Stat. 27, 1519–1535 (1999)
Niu, C.Z., Guo, X., Xu, W.L., Zhu, L.X.: Empirical likelihood inference in linear regression with nonignorable missing response. Comput. Stat. Data Anal. 79, 91–112 (2014)
Owen, A.B.: Empirical likelihood ratio confidence intervals for a single functional. Biometrika 75, 237–249 (1988)
Owen, A.B.: Empirical likelihood ratio confidence regions. Ann. Stat. 18, 90–120 (1990)
Qin, J., Lawless, J.F.: Empirical likelihood and general estimating equations. Ann. Stat. 22, 300–325 (1994)
Tosteson, T., Stefanski, L.A., Schafer, D.W.: A measurement error model for binary and ordinal regression. Stat. Med. 8, 1139–1147 (1989)
Wang, N., Carroll, R.J., Liang, K.Y.: Quasi-likelihood and variance functions in measurement error models with replicates. Biometrics 52, 423–432 (1996)
Wang, X.L., Li, G.R., Lin, L.: Empirical likelihood inference for semi-parametric varying-coefficient partially linear EV models. Metrika 73, 171–185 (2011)
Wei, C.H.: Statistical inference for restricted partially linear varying coefficient errors-in-variables models. J. Stat. Plan. Inference 142, 2464–2472 (2012)
Yan, L., Chen, X.: Empirical likelihood for partly linear models with errors in all variables. J. Multivar. Anal. 130, 275–288 (2014)
Yang, S.J., Park, B.U.: Efficient estimation for partially linear varying coefficient models when coefficient functions have different smoothing variables. J. Multivar. Anal. 126, 100–113 (2014)
Yang, Y., Li, G., Peng, H.: Empirical likelihood of varying coefficient errors-in-variables models with longitudinal data. J. Multivar. Anal. 127, 1–18 (2014)
You, J.H., Chen, G.M.: Empirical likelihood for semiparametric varying-coefficient partially linear regression models. Stat. Probab. Lett. 76, 412–422 (2006)
You, J.H., Zhou, Y., Chen, G.M.: Corrected local polynomial estimation in varying-coefficient models with measurement errors. Can. J. Stat. 34, 391–410 (2006)
Zhang, B.: Confidence intervals for a distribution function in the presence of auxiliary information. Comput. Stat. Data Anal. 21, 327–342 (1996)
Zhang, W., Lee, S.Y., Song, X.: Local polynomial fitting in semivarying coefficient models. J. Multivar. Anal. 82, 166–188 (2002)
Zhou, Y., Liang, H.: Statistical inference for semiparametric varying-coefficient partially linear models with error-prone linear covariates. Ann. Stat. 37, 427–458 (2009)
Zhu, L., Cui, H.: A semiparametric regression model with errors in variables. Scand. J. Stat. 30, 429–442 (2003)
Acknowledgments
The authors would like to thank the Editor and two referees for their truly helpful comments and suggestions which led to a much improved presentation. This research was supported by the National Natural Science Foundation of China (11401006, 11226218, 71171003, 71271003, 11471160, 11101114), the National Statistical Science Research Key Program of China (2013LZ45), the Programming Fund Project of the Humanities and Social Sciences Research of the Ministry of Education of China (12YJA790041), Jiangsu Provincial Basic Research Program (Natural Science Foundation) (BK20131345) and the Fundamental Research Funds for the Central Universities (30920130111015).
Author information
Authors and Affiliations
Corresponding author
Appendix A: Assumptions and proofs
Appendix A: Assumptions and proofs
For the convenience and simplicity, let \(Z=(X_1,\ldots ,X_n)^\mathrm{T}\), \(\varepsilon =(\varepsilon _1,\ldots ,\varepsilon _n)^\mathrm{T}\), \(\eta =(\eta _1,\ldots ,\eta _n)^\mathrm{T}\), \(\widetilde{Z}=(I-S)Z\), \(\widetilde{M}=(I-S)M\), \(\widetilde{\varepsilon }=(I-S)\varepsilon \), \(\widetilde{\eta }=(I-S)\eta \), \(a_n=\{\frac{\log (1/h_1)}{nh_1}\}^{1/2}+h_1^2\) and \(C\) denote positive constant whose value may vary at each occurrence. Before proving the main theorems, we begin this section with making the following assumptions.
-
(C1)
The kernel \(K(\cdot )\) is a symmetric probability function with bounded support.
-
(C2)
The variable \(U\) has a bounded support \(\fancyscript{U}\) and its density function \(p(u)>0\) is Lipschitz continuous and bounded away from zero on \(\fancyscript{U}\).
-
(C3)
The matrixes \(\Gamma (u)\) is non-singular. \(\Gamma ^{-1}(u)\), \(\Phi (u)\) and \(E(X_1X_1^\mathrm{T}|U_1=u)\) are all Lipschitz continuous. \(E\Vert Z_1\Vert ^{2s}<\infty \), \(E\Vert X_1\Vert ^{2s}<\infty \), \(E|\varepsilon _1|^{2s}<\infty \) and \(E\Vert \eta _1\Vert ^{2s}<\infty \) for some \(s>2\), where \(\Vert \cdot \Vert \) is the \(L_2\) norm.
-
(C4)
\(\{\alpha _j(\cdot ),j=1,\ldots ,q\}\) have continuous second derivative on \(\fancyscript{U}\).
-
(C3)
There exist a \(\delta <2-s^{-1}\) such that \(\lim _{n\rightarrow \infty }n^{2\delta -1}h_1=\infty .\)
-
(C4)
The bandwidth \(h_1\) satisfies that \(nh_1^2(\log n)^{-2}\rightarrow \infty \), and \(nh_1^8\rightarrow 0\).
-
(C5)
The bandwidth \(h_2\) satisfies that \(nh_2\rightarrow \infty \) and \(h_2\rightarrow 0.\)
Lemma 5.1
Assume that conditions (C1)–(C5) are satisfied. Then we have
Lemma 5.1 can be proved as Lemmas 2 and 3 in Feng and Xue (2014).
Lemma 5.2
Let \(D_1,\ldots ,D_n\) be independent and identical distributed random variables. If \(E|D_1|^s<\infty \) for \(s>1,\) then \(\max _{1\le i\le n}|D_i|=o(n^{1/s})\) a.s.
Lemma 5.2 can be proved as Lemma 3 in Owen (1990).
Lemma 5.3
Under the conditions of Theorem 2.1, if \(\beta \) is the true value of the parameter\(,\) then we have
where \(V_{AI}=\bigg (\begin{array}{ll} V_{A}&{}0\\ 0&{}V_1 \end{array}\bigg ),\) \(V_1 =E\big \{\big [E(X_1X_1^\mathrm{T}|U_1)-\Phi ^\mathrm{T}(U_1)\Gamma ^{-1}(U_1)\Phi (U_1)\big ](\varepsilon _1-\eta _1^\mathrm{T}\alpha (U_1))^2\big \} +\sigma ^2E\big \{\Phi ^\mathrm{T}(U_1)\Gamma ^{-1}(U_1)\Sigma _{\eta }\Gamma ^{-1}(U_1)\Phi (U_1)\big \} +E\big \{\Phi ^\mathrm{T}(U_1)\Gamma ^{-1}(U_1)(\eta _1\eta _1^\mathrm{T}-\Sigma _{\eta })\big \}^{\otimes 2}\) and \(V_A=E\{A(X_1)A^\mathrm{T}(X_1)\}\).
Proof
From the definition of \(\varphi _i(\beta )\) and Lemma 5.1, by simply calculation and the proof in Lemma 4 in Feng and Xue (2014), we have
It is easy to see that \(G_{i}\) is independent and identical distributed with mean zero and \(\text{ Var }(G_{i})=V_1\). Thus, by the Slutsky theorem and the central limit theorem, we obtain
Also, we find \(\frac{1}{\sqrt{n}}\sum _{i=1}^nA(X_i)\mathop {\longrightarrow }\limits ^{\mathcal {D}}N(0,V_{A})\) and \(\mathrm{Cov}\big (\frac{1}{\sqrt{n}}\sum _{i=1}^nA(X_i), \frac{1}{\sqrt{n}}\sum _{i=1}^n\varphi _i(\beta )\big ) \rightarrow 0\) by (5.1), which together with (5.2) and the central limit theorem yields Lemma 5.3(i).
As to Lemma 5.3(ii), observe that
The law of large numbers implies \(\frac{1}{n}\sum _{i=1}^nA(X_i)A^\mathrm{T}(X_i)\mathop {\longrightarrow }\limits ^{\mathcal {P}}V_A.\) On the other hand, by Lemma 5.1 and the proof in (5.1), it follows that
Similar to the proof of (5.1), we can derive
follows from \(E\Vert n^{-1}\sum _{i=1}^nA(X_i)G^\mathrm{T}_i\Vert ^2=n^{-1}E\Vert A(X_i)G_i^\mathrm{T}\Vert ^2\rightarrow 0\) and Markov inequality. Thus, Lemma 5.3(ii) holds.
As to Lemma 5.3(iii), note that \(A(X_i)\) is i.i.d. and \(E\{A(X_i)A^\mathrm{T}(X_i)\}\) is positive definite, then by Lemma 5.2, we have \(\max _{1\le i\le n}\Vert A(X_i)\Vert =o(n^{1/2})\) \(a.s.\). From (5.1) and standard calculations, when \(n\) is large enough, we can derive that
Thus, \(\max _{1\le i\le n}\Vert \psi _i(\beta )\Vert =o_p(n^{1/2})\). Furthermore, one can obtain that \(\lambda _1=O_p(n^{-1/2})\) by Lemma 5.3(ii) and using the arguments similar to Owen (1990). The proof of Lemma 5.1 is thus completed. \(\square \)
Set \(\lambda _2=(\lambda _{21}^\mathrm{T},\lambda _{22}^\mathrm{T})^\mathrm{T}\), where \(\lambda _{21}\) and \(\lambda _{22}\) are \(r\)-dimensional and \(q\)-dimensional column vectors.
Lemma 5.4
Under the conditions of Theorem 2.3, for a given \(u,\) if \(\alpha (u)\) is the true value of the parameter\(,\) then
where \(\widehat{\xi }_i(\alpha (u))=\bigg (\begin{array}{l} \frac{1}{\sqrt{n}}A(X_i)\\ \frac{1}{\sqrt{nh_2}}\widehat{\zeta }_i(\alpha (u)) \end{array}\bigg )\), \(\Sigma _{AI}(u)=\bigg (\begin{array}{ll} V_{A}&{}0\\ 0&{}\Sigma _1(u) \end{array}\bigg )\) and \(\Sigma _1(u)=\Big \{(\Gamma (u)+\Sigma _{\eta })E(\varepsilon _1^2|U_1=u)+\alpha ^\mathrm{T}(u)\Sigma _{\eta }\alpha (u)\Gamma (u)\Big \}f(u)\int K^2(x)\mathrm{d}x.\)
Proof
Observe that
Note that \(\big \{\big [\varepsilon _i(Z_i+\eta _i)+\eta _i^\mathrm{T}\alpha (U_i)Z_i\big ]K\big (\frac{U_i-u}{h_2}\big ),1\le i\le n\big \}\) is independent and identical distributed with mean zero and
Then we have \(B_{1n}\mathop {\longrightarrow }\limits ^{\mathcal {D}}N(0,\Sigma _1(u))\).
Theorems 2.2 and 2.4, Lemma 5.2 and condition (C3) imply that \(\hat{\beta }-\beta =O_p(n^{-1/2})\), \(\tilde{\alpha }(u)-\alpha (u)=O_p(n^{-1/2})\) and \(\eta _i\eta _i^\mathrm{T}-\Sigma _{\eta }=o(n^{1/s})\) \(a.s.\) Then by some simple calculations we obtain \(B_{in}=o_p(1)\) for \(i=2,\ldots ,7.\) Invoking the Slutsky theorem and (5.3), we get
Note that \(\frac{1}{\sqrt{n}}\sum _{i=1}^nA(X_i)\mathop {\longrightarrow }\limits ^{\mathcal {D}}N(0,V_{A})\) and \(\mathrm{Cov}\big (\frac{1}{\sqrt{n}}\sum _{i=1}^nA(X_i), \frac{1}{\sqrt{nh_2}}\sum _{i=1}^n\hat{\zeta }_i(\alpha (u))\big ) \rightarrow 0\), which together with (5.4) and the central limit theorem leads to Lemma 5.4(i).
Analogously to the proof of Lemma 5.3(ii), we can verify Lemma 5.4(ii) easily. As to Lemma 5.4(iii), we find
From Markov inequality and conditions (C3) and (C5), one can obtain that
which implies that \(D_1=o_p(\sqrt{nh_2})\). Similarly, by \(\hat{\beta }-\beta =O_p(n^{-1/2})\), \(\tilde{\alpha }(u)-\alpha (u)=O_p(n^{-1/2})\) and \(\eta _i\eta _i^\mathrm{T}-\Sigma _{\eta }=o(n^{1/s})\) \(a.s.\), it can be shown that \(J_{in}=o_p(\sqrt{nh_2})\) \(i=2,\ldots ,7\). Therefore we obtain that \(\max _{1\le i\le n}\Vert \widehat{\zeta }_i(\alpha (u))\Vert =o_p((nh_2)^{1/2})\), which together with \(\max _{1\le i\le n}\Vert A(X_i)\Vert =o(n^{1/2})\) \(a.s.\), gives \(\max _{1\le i\le n}\Vert \widehat{\xi }_i(\alpha (u))\Vert =o_p(1)\).
Applying Lemma 5.4(ii) and the proof in Owen (1990) one can derive that \(\lambda _{21}=O_p(n^{-1/2})\) and \(\lambda _{22}=O_p((nh_2)^{-1/2})\), which completes the proof of Lemma 5.4. \(\square \)
Proof of Theorem 2.1
Applying the Taylor expansion to (2.8) and invoking Lemma 5.3, we obtain that
From (2.9), we have
Using Lemma 5.3, we find
Then \(\sum _{i=1}^n[\lambda _1^\mathrm{T} \psi _i(\beta _0)]^2=\sum _{i=1}^n\lambda _1^\mathrm{T} \psi _i(\beta _0)+o_p(1),\) and
Thus,
This together with Lemma 5.3 completes the proof. \(\square \)
Proof of Theorem 2.2
Note that \(\hat{\beta }_\mathrm{ME}\) and \(\hat{\lambda }_1=\lambda (\hat{\beta }_\mathrm{ME})\) satisfy \(H_{1n}(\hat{\beta }_\mathrm{ME},\hat{\lambda }_1)=0\) and \(H_{2n}(\hat{\beta }_\mathrm{ME},\hat{\lambda }_1)=0\), where
Then by expanding \(H_{1n}(\hat{\beta }_\mathrm{ME},\hat{\lambda }_1)=0\) and \(H_{2n}(\hat{\beta }_\mathrm{ME},\hat{\lambda }_1)=0\) at \((\beta _0,0)\), we derive that
where \(\delta _n=\Vert \hat{\beta }_\mathrm{ME}-\beta _0\Vert +\Vert \hat{\lambda }_1\Vert \). Then, we find
where \(\widehat{\Delta }_1=\frac{1}{n}\sum _{i=1}^n(\widetilde{X}_i\widetilde{X}_i^\mathrm{T}-X^\mathrm{T} Q_i^\mathrm{T} \Sigma _{\eta } Q_i X)\). Lemma 5.3 and \(H_{1n}(\beta _0,0)=n^{-1}\sum _{i=1}^n\varphi _i(\beta _0)=O_p(n^{-1/2})\) imply \(\delta _n=O_p(n^{-1/2})\). Therefore,
This together with (5.3), Lemmas 5.1 and Slutsky theorem yields the result of Theorem 2.2. \(\square \)
Theorem 2.3 can be proved by using the same argument used in Theorems 2.1. Theorem 2.4 can be verified by the proof of Theorem 3 in Feng and Xue (2014) and the asymptotic normalities of \(\tilde{\beta }\) and \(\hat{\beta }_\mathrm{ME}\). We omit the details here.
Rights and permissions
About this article
Cite this article
Fan, GL., Xu, HX. & Huang, ZS. Empirical likelihood for semivarying coefficient model with measurement error in the nonparametric part. AStA Adv Stat Anal 100, 21–41 (2016). https://doi.org/10.1007/s10182-015-0247-7
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10182-015-0247-7
Keywords
- Auxiliary information
- Empirical likelihood
- Local bias-corrected
- Measurement error
- Semivarying coefficient model