Abstract
In this paper, we consider the variable selection problem of the generalized random coefficient autoregressive model (GRCA). Instead of parametric likelihood, we use non-parametric empirical likelihood in the information theoretic approach. We propose an empirical likelihood-based Akaike information criterion (AIC) and a Bayesian information criterion (BIC).
Similar content being viewed by others
1 Introduction
Consider the following p-order generalized random coefficient autoregressive model:
where τ denotes the transpose of a matrix or vector, \(\Phi_{t}=( \Phi_{t1},\ldots ,\Phi_{tp})^{\tau }\) is a random coefficient vector, \(Y(t-1)=(Y_{t-1}, \ldots ,Y_{t-p})^{\tau }\), and \(\{\binom{\Phi_{t}}{\varepsilon_{t}}, t=0,\pm 1, \pm 2, \ldots \}\) is a sequence of i.i.d. random vectors with \(E(\Phi_{t})=\phi =(\phi_{1},\ldots ,\phi_{p})\), \(E(\varepsilon _{t})=0\), and \(\operatorname{Var}\binom{\Phi_{t}}{\varepsilon_{t}} = \bigl( {\scriptsize\begin{matrix}{} V_{\phi }& \sigma_{\Phi \varepsilon }\cr \sigma_{\Phi \varepsilon }^{\tau }& \sigma_{\varepsilon }^{2} \end{matrix}}\bigr)\).
As a generalization of the usual autoregressive model, the random coefficient autoregressive (RCAR) model (cf. [1, 2]), the Markovian bilinear model and its generalization, and the random coefficient exponential autoregressive model (cf. [3–5]), model (1) was first introduced by Hwang and Basawa [6]. GRCA has become one of the important models in the nonlinear time series context. In recent years, GRCA has been studied by many authors. For instance, Hwang and Basawa [7] established the local asymptotic normality of a class of generalized random coefficient autoregressive processes. Carrasco and Chen [8] provided the tractable sufficient conditions that simultaneously imply strict stationarity, finiteness of higher-order moments, and β-mixing with geometric decay rates. Zhao and Wang [9] constructed confidence regions for the parameters of model (1) by using an empirical likelihood method. Furthermore, Zhao et al. [10] also considered the problem of testing the constancy of the coefficients in the stationary one-order generalized random coefficient autoregressive model. In this paper, we consider the variable selection problem of the GRCA based on the empirical likelihood method.
Many model selection procedures have been proposed in the statistical literature, including the adjusted \(R^{2}\) (see Theil [11]), the AIC (see Akaike [12]), BIC (see Schwarz [13]), Mallow’S \(C_{p}\) (see Mallows [14]). Other criteria in the literature include Hannan and Quinn’s criterion [15], Geweke and Meese’s criterion [16], Cavanaugh’s Kullback information criterion [17], and the deviance information criterion of Spiegelhalter et al. [18]. Also, Tsay [19], Hurvich and Tsai [20] and Pötscher [21] have studied model selection methods in time series models. Recently, the model selection problem has been extended to moment selection as in Andrews [22], Andrews and Lu [23] and Hong et al. [24]. These model selection methods are concerned with parsimony, as was stressed in Zellner et al. [25], as well as accuracy or power in choosing models.
In this paper, we develop an information theoretic approach to variable selection problem of GRCA. Specifically, instead of parametric likelihood, we use non-parametric empirical likelihood (see Owen [26, 27]) in the information theoretic approach. We propose an empirical likelihood-based Akaike information criterion (EAIC) and a Bayesian information criterion (EBIC).
The paper proceeds as follows. The next section is concerned with the methodology and the main results. Section 3 is devoted to the proofs of the main results.
Throughout the paper, we use the symbols “\(\stackrel{d}{\longrightarrow}\)” and “\(\stackrel{p}{\longrightarrow }\)” to denote convergence in distribution and convergence in probability, respectively. We abbreviate “almost surely” and “independent identical distributed” to “a.s.” and “i.i.d.”, respectively. \(o_{p}(1)\) means a term which converges to zero in probability. \(O_{p}(1)\) means a term which is bounded in probability. Furthermore, the Kronecker product of the matrices A and B is denoted by \(A \otimes B\), and \(\Vert M \Vert \) denotes the \(L_{2}\) norm for vector or matrix M.
2 Methods and main results
In this section, we will first propose the empirical likelihood-based information criteria for choice of a GRCA, then we investigate the asymptotic properties of the new variable selection method.
2.1 Empirical likelihood-based information criteria
Hwang and Basawa [6] derived the conditional least-squares estimator ϕ̂ of ϕ, which is given by
By using the estimating equation of the conditional least-squares estimator, we can obtain the following score function:
where \(G_{t}(\phi )=Y_{t}Y(t-1)-Y(t-1)Y^{\tau }(t-1)\phi \). Following Owen [26], the empirical likelihood statistic for ϕ is defined as
where \(p_{1},\ldots , p_{n}\) are all sets of nonnegative numbers summing to 1. By using the Lagrange multiplier method, let
After simple algebraic calculation, we have
Note that \(\sum_{t=1}^{n}p_{t}=1\) and \(\sum_{t=1}^{n}p _{t}G_{t}(\phi )=0 \). So we have \(\gamma =-n\) and \(p_{t}=\frac{1}{n(1+ \lambda^{\tau }G_{t}(\phi ))}\), which implies that
where λ is the solution of the equation
The definition of \(\tilde{l}(\phi )\) relies on finding a positive \({p_{t}}'s\) such that \(\sum_{t=1}^{n}p_{t}G_{t}(\phi )=0\) for each ϕ. The solution exists if and only if the convex hull of the \(G_{t}(\phi )\), \(t=1, 2, \ldots , n\) contains zero as an inner point. When the model is correct, the solution exists with probability tending to 1 as the sample size \(n\rightarrow \infty \) for ϕ in a neighborhood of \(\phi_{0}\). However, for finite n and at some ϕ value, the equation often does not have a solution in \(p_{t}\). To avoid this problem, we introduce the adjusted empirical likelihood.
Further let \(\bar{G}_{n}=n^{-1}\sum_{t=1}^{n}p_{t}G_{t}(\phi )\) and define \({G}_{n+1}=-a_{n}\bar{G}_{n}\) for some positive constant \(a_{n}\). We adjust the profile empirical log-likelihood ratio function to
with \(\tilde{\lambda }=\tilde{\lambda }(\phi )\) being the solution of
Since 0 always lies on the line connecting \(\bar{G}_{n}\) and \({G}_{n+1}\), the adjusted empirical log-likelihood ratio function is well defined after adding a pseudo-value \({G}_{n+1}\) to the data set. The adjustment is particularly useful so that a numerical program does not crash simply because some undesirable ϕ is assessed.
A full GRCA assumes that \(y_{t}\) relates to \(\Phi_{t}^{\tau }Y(t-1)\) with \(E(\Phi_{t})=\phi \) being unknown parameter of size p. Let s be a subset of \(\{1, 2, \ldots , p\}\), and \(Y^{[s]}(t-1)\) and \(\phi^{[s]}\) be subvectors of \(Y(t-1)\) and ϕ containing entries in positions specified by s. Consider the pth-order GRCA specified by \(E(G_{t}(\phi ))=0\) and a submodel specified by \(E(G^{[s]}_{t}( \phi^{[s]}))=0\), where \(G^{[s]}_{t}(\phi^{[s]})=Y_{t}Y^{[s]}(t-1)-Y ^{[s]}(t-1)(Y^{[s]}(t-1))^{\tau }\phi^{[s]}\). For a given s, let \(G^{[s]}_{t}=Y_{t}Y^{[s]}(t-1)-Y^{[s]}(t-1)(Y^{[s]}(t-1))^{\tau } \phi^{[s]}\), \(\bar{G}_{n}^{[s]}=n^{-1}\sum_{t=1}^{n}G^{[s]} _{t}\) and \(G^{[s]}_{n+1}=-a_{n}\bar{G}^{[s]}_{n}\) for some positive constant \(a_{n}\). The adjusted empirical log-likelihood ratio becomes
with \(\tilde{\lambda }=\tilde{\lambda }(\phi )\) being the solution of
We define the adjusted profile empirical log-likelihood ratio as
The empirical likelihood versions of AIC and BIC are then defined as
where k is the cardinality of s.
After \(l(s)\) is evaluated for all s, we select the model with the minimum EAIC or EBIC value.
2.2 Asymptotic properties
It is well known that under some mild conditions the parametric BIC is consistent for variable selection while the parametric AIC is not. Similarly, we can prove that, when p is constant, EBIC is consistent but EAIC is not.
For purposes of illustration, in what follows, we rewrite the model in the following matrix form (see Hwang and Basawa [6]): let \(U_{t}=(\varepsilon_{t},0,0,\ldots ,0)^{\tau }\) are \(p\times 1\) vectors, \(\tilde{\Phi }_{tj}={\Phi }_{tj}-\phi_{j}\), \(j=1,\ldots ,p\),
Then model (1) can be written as
In order to obtain our theorems, we need the following regularity conditions:
- \(\mathbf{(A_{1})}\) :
-
All the eigenvalues of the matrix \(E(C_{t}\otimes C _{t})+(B\otimes B)\) are less than unity in modulus.
- \(\mathbf{(A_{2})}\) :
-
\(EY_{t}^{6}<\infty \).
Remark 1
As for the condition \(\mathbf{(A_{1})}\) and the sufficient condition for \(E\vert y_{t} \vert ^{2m}<\infty\) (\(m=1, 2, \ldots\)), we refer to Hwang and Basawa [6].
Theorem 2.1
Let \(A=E(G_{t}(\phi_{0})G^{\tau }_{t}(\phi_{0}))\) and \(B=E((\partial G_{t}(\phi )/\partial \phi )|_{\phi =\phi_{0}})\). If \(\mathbf{(A_{1})}\) and \(\mathbf{(A_{2})}\) hold, then there exists a sequence of adjusted empirical likelihood estimates ϕ̃ of ϕ such that
and
where \(U=A^{-1}-A^{-1}B(B^{\tau }A^{-1}B)^{-1}B^{\tau }A^{-1}\).
Note that when a submodel s is a true model, it implies \(\phi_{0} ^{[\bar{s}]}=0\). That is, components of \(\phi_{0}\) not in s are zero. Therefore, \(Y_{t}\) only relates to the variables in positions specified by s. The following theorem shows that when \(\phi_{0}^{[\bar{s}]}=0\) is true, then adjusted empirical log-likelihood ratio statistic has a chi-squared limiting distribution with k fewer degrees of freedom.
Theorem 2.2
Assume that \(\mathbf{(A_{1})}\) and \(\mathbf{(A_{2})}\) hold and \(\phi_{0} ^{[\bar{s}]}=0\) for a submodel s of size k. Then when \(a_{n}=o _{p}(n^{\frac{1}{2}})\), we have \(l(s)\rightarrow \chi^{2}_{p-k}\) in distribution as \(n\rightarrow \infty \).
When the null hypothesis of \(\phi_{0}^{[\bar{s}]}=0\) is not true, the likelihood ratio go to ∞ as \(n\rightarrow \infty \). We state the following theorem in terms of the adjusted empirical likelihood which also applies to the usual empirical likelihood.
Theorem 2.3
Assume that \(\mathbf{(A_{1})}\) and \(\mathbf{(A_{2})}\) hold and \(a_{n}=o_{p}(n ^{\frac{1}{2}})\). Then for any \(\phi \neq \phi_{0}\) such that \(E(G_{t}(\phi ))\neq 0\), \(l(s)\rightarrow \infty \) in probability as \(n\rightarrow \infty \).
The following theorem indicates that, when p is constant, EBIC is consistent but EAIC is not.
Theorem 2.4
Assume that \(\mathbf{(A_{1})}\) and \(\mathbf{(A_{2})}\) hold and if there exists a subset \(s_{0}\) of \(1, 2, \ldots , p\) such that, for any other subset s, \(E(G^{[s]}_{t}(\phi^{[s]}))=0\) for some ϕ if and only if s contains \(s_{0}\). Then, EBIC is consistent and EAIC is not consistent.
3 Proofs of the main results
In order to prove Theorem 2.1, we first present several lemmas.
Lemma 3.1
Assume that \(\mathbf{(A_{1})}\) and \(\mathbf{(A_{2})}\) hold. Then A is positive definite and B has rank p.
Proof
After simple algebra calculation, we have, for any nonzero vector \(c=(c_{1}, \ldots , c_{p})\in R^{P}\),
Note that the conditional distribution of \(Y_{t}\), given \(Y(t-1)\), is not a degenerate distribution, which implies that \(\operatorname{Var}(Y_{t}| Y(t-1))>0 \) a.s. It follows that \((c^{\tau }Y(t-1))^{2}\operatorname{Var}(Y_{t}| Y(t-1)) \geq 0 \) a.s. Therefore, \(c^{\tau }Ac=0\) if and only if \(c^{\tau }Y(t-1)=0\) a.s. Without loss of generality, suppose that the first component \(c_{1}\) of c is 1, so \(Y_{t-1}=-c_{2}Y_{t-2}-\cdots -c_{p}Y_{t-p}\), which is contradictory with the fact that the conditional distribution of \(Y_{t-1}\), given \((Y_{t-2}, \ldots , Y _{t-p})\), is not degenerate. Hence \(c^{\tau }Ac>0\). That is, A is positive definite.
Similarly, we can also prove that B has rank p. The proof of Lemma 3.1 is thus complete. □
Lemma 3.2
Assume that \(\mathbf{(A_{1})}\) and \(\mathbf{(A_{2})}\) hold. Then when \(a_{n}=o(n^{\frac{1}{2}})\), we have
uniformly about \(\phi \in \{\phi | \Vert \phi -\phi_{0} \Vert \leq n^{-\frac{1}{3}}\}\).
Proof
Note that
First, note that
By the ergodic theorem, we have
Further, note that
This, together with (16), proves that
Again by the ergodic theorem, we can prove that
Finally, we prove that
Note that
Similar to the proof of (18), we can show that
In what follows, we consider \(\Vert \frac{1}{n}\sum_{t=1}^{n}G_{t}(\phi_{0}) \Vert \).
Denote the ith component of \(G_{t}(\phi_{0})\) by \(G_{ti}(\phi_{0})\). Then \(\{G_{ti}(\phi_{0}), 1\leq i\leq p\}\) is a stationary ergodic martingale difference sequence with \(E(G_{ti}(\phi_{0}))=0\) and \(E((G_{ti}(\phi_{0}))^{2})<\infty \). By the law of the iterated logarithm of martingale difference sequence, we have, for \(1\leq i \leq p\),
It follows that
Then, by (21) and (22), we have
Therefore
This, together with (18) and (19), proves Lemma 3.2. □
Lemma 3.3
Assume that \(\mathbf{(A_{1})}\) and \(\mathbf{(A_{2})}\) hold. Then when \(a_{n}=o(n^{\frac{1}{2}})\), we have
uniformly about \(\phi \in \{\phi | \Vert \phi -\phi_{0} \Vert \leq n^{-\frac{1}{3}}\}\).
Proof
Note that
From (23), together with \(a_{n}=o(n^{\frac{1}{2}})\), it follows immediately that
The next step in the proof is to show that
By the Fubini theorem, we have, for any positive integer k,
Thus, using the ergodic theorem,
By the Borel–Cantelli lemma, we know that
so that
Take \(k=\frac{1}{m}\), then there exists \(Q_{m}\) with \(P(Q_{m})=0\), such that, for any \(\omega \in Q_{m}^{c}\),
Further, let \(Q=\bigcup_{m=1}^{\infty }Q_{m}\). Then
which, together with the fact that \(P(Q)=0\), implies that
The proof is complete. □
Lemma 3.4
Assume that \(\mathbf{(A_{1})}\) and \(\mathbf{(A_{2})}\) hold. Then when \(a_{n}=o(n^{\frac{1}{2}})\), we have
uniformly about \(\phi \in \{\phi | \Vert \phi -\phi_{0} \Vert \leq n^{-\frac{1}{3}}\}\).
Proof
Write \(\Vert \lambda (\phi ) \Vert =\rho (\phi )\theta (\phi )\), where \(\rho (\phi )>0\) and \(\Vert \theta (\phi ) \Vert =1\). Further let
Then
which implies that
Further, by the ergodic theorem, we have
where \(A=E(G_{t}(\phi_{0})G^{\tau }_{t}(\phi_{0}))\).
Since
which implies that
where \(\sigma_{\min }\) is the smallest eigenvalue and the largest eigenvalue of A. This, together with Lemma 3.1 and (36), proves that
Combined with (23) and Lemma 3.3, this establish (34) and completes the proof. □
Lemma 3.5
Assume that \(\mathbf{(A_{1})}\) and \(\mathbf{(A_{2})}\) hold, and \(a_{n}=o _{(}n^{\frac{1}{2}})\). Then, as \(n\rightarrow \infty \), with probability 1, \(l(\phi )\) attains its minimum value at some point ϕ̃ in the interior of the ball \(\Vert \phi -\phi_{0} \Vert \leq n^{-\frac{1}{3}}\) and ϕ̃ and \(\tilde{\lambda }=\lambda (\tilde{\phi })\) satisfy \(Q_{1,n+1}(\tilde{\phi },\tilde{\lambda })=0\) and \(Q_{2,n+1}( \tilde{\phi },\tilde{\lambda })=0\), where \(Q_{1,n+1}(\phi ,\lambda )\) is defined in (35) and
The proof is similar to the proof of Lemma 1 of Qin and Lawless [28], so we omit the details.
Proof of Theorem 2.1
In what follows, we omit \((\phi ,\lambda )\) in the notation if a function is evaluated at \((\phi_{0}, 0)\). Expanding \(Q_{1,n+1}(\tilde{\phi }, \tilde{\lambda })\), \(Q_{2,n+1}(\tilde{\phi },\tilde{\lambda })\) at \((\phi_{0}, 0)\) leads to
and
where \(\delta_{n}=\Vert \tilde{\phi }-\phi_{0} \Vert ^{2}+\Vert \tilde{\lambda } \Vert ^{2}=O_{p}(n^{-\frac{2}{3}})\).
Note that
and
These, combined with (41) and (42), give
and
Further, applying the central limit theorem to \(Q_{1,n+1}\) and using Slustzky’s theorem, we can prove Theorem 2.1. □
Proof of Theorem 2.2
Let λ̃ be the Lagrange multiplier corresponding to \(\tilde{\phi }^{[s]}\), the maximum point of \(l(\phi^{[s]})\). With this notation, we may write
Note that
This, together with (49), yields
Further note that \(Q_{1,n+1}\) is asymptotic normal with covariance matrix A and \(\{A^{-1}-A^{-1}B(B^{\tau }A^{-1}B)^{-1}B^{\tau }A^{-1} \}A\{A^{-1}-A^{-1}B(B^{\tau }A^{-1}B)^{-1}B^{\tau }A^{-1}\}=\{A^{-1}-A^{-1}B(B^{\tau }A^{-1}B)^{-1}B^{\tau }A^{-1}\}\). Therefore, we have \(\l (s)\rightarrow \chi^{2}(p-k)\) in distribution as \(n\rightarrow \infty \). The proof is complete. □
Proof of Theorem 2.3
Since \(E(G_{t}(\phi ))\neq 0\), it follows that there exists \(\delta >0\), such that
Furthermore, note that \(E(G_{t}^{\tau }(\phi ))^{2}<\infty \). Thus, by a similar method to the proof of (27), we can prove that
Let \(\check{\lambda }=n^{-\frac{2}{3}}(\frac{1}{n}\sum_{t=1} ^{n}G_{t}(\phi ))\log n\). Then
Thus, with probability going to 1, \(1+\check{\lambda }^{\tau }G_{t}( \phi )>0\) for \(i=1, \ldots , n+1\). Using the duality of the maximization problem and (51)–(53), we have
which implies that \(l(s)\rightarrow \infty \) in probability as \(n\rightarrow \infty \). The proof is complete. □
Proof of Theorem 2.4.
First, we consider EAIC. Consider the situation when \(s_{0}\) is empty. Let \(s=\{1\}\) which contains a single covariant. Based on expansion in the proof of Theorem 2.2, we can prove that \(l(s_{0})-l(s)\rightarrow \chi^{2}_{1}\), which implies that \(\lim_{n\rightarrow \infty }P(l(s_{0})-l(s)>2)>0\). Therefore, EAIC is not consistent.
Next, we consider EBIC. Suppose s is a model which does not contain \(s_{0}\). Then, \(E(G^{[s]}_{t}(\phi^{[s]}))\neq 0\) for any \(\phi^{[s]}\). Therefore, we have \(l(s)\geq 2n^{\frac{1}{3}}\delta^{2}\log (n)+o _{P}(1)\). This order implies that
That is, EBIC will not select any model s that does not contain \(s_{0}\).
Furthermore, if s contains \(s_{0}\), and \(k>0\) additional insignificant variables, by Theorem 2.2, we have
which implies that
as \(n\rightarrow \infty \). Thus, the model s will not be selected by EBIC as \(n\rightarrow \infty \). Because p is finite, there are only finite number of scompeting against \(s_{0}\), and each of them has \(o(1)\) probability being selection. So EBIC is consistent. The proof is complete. □
4 Conclusions
It should be pointed out that variable selection has always been an important problem for our statistician. Many variable selection methods have been proposed in the statistical literature. But for the variable selection method of GRCA, so far it has not been provided by statistician. In this paper, instead of parametric likelihood, we further propose an Akaike information criterion (EAIC) and a Bayesian information criterion (EBIC) for the variable selection problem of GRCA based on the empirical likelihood method. Moreover, we also prove that under some mild conditions the parametric EBIC is consistent, while the parametric EAIC is not when p is constant.
References
Nicholls, D.F., Quinn, B.G.: Random Coefficient Autoregressive Models: An Introduction. Springer, New York (1982)
Tong, H.: Nonlinear Time Series. Oxford University Press, Oxford (1990)
Tong, H.: A note on a Markov bilinear stochastic process in discrete time. J. Time Ser. Anal. 2, 279–284 (1981)
Feigin, P.D., Tweedie, R.L.: Random coefficient autoregressive processes: a Markov chain analysis of stationarity and finiteness of moments. J. Time Ser. Anal. 6, 1–14 (1985)
Hwang, S.Y., Basawa, I.V.: Asymptotic optimal inference for a class of nonlinear time series models. Stoch. Process. Appl. 46, 91–113 (1993)
Hwang, S.Y., Basawa, I.V.: Parameter estimation for generalized random coefficient autoregressive processes. J. Stat. Plan. Inference 68, 323–327 (1998)
Hwang, S.Y., Basawa, I.V.: The local asymptotic normality of a class of generalized random coefficient autoregressive processes. Stat. Probab. Lett. 34, 165–170 (1997)
Carrasco, M., Chen, X.: β-Mixing and moment properties of RCA models with application to \(\operatorname{GARCH}(p, q)\). C. R. Acad. Sci., Sér. 1 Math. 331, 85–90 (2000).
Zhao, Z.W., Wang, D.H.: Statistical inference for generalized random coefficient autoregressive model. Math. Comput. Model. 56, 152–166 (2012)
Zhao, Z.W., Wang, D.H., Peng, C.X.: Coefficient constancy test in generalized random coefficient autoregressive model. Appl. Math. Comput. 219, 10283–10292 (2013)
Theil, H.: Economic Forecasts and Policy. North-Holland, Amsterdam (1961)
Akaike, H.: A new look at the statistical model identification. IEEE Trans. Autom. Control 19, 716–723 (1974)
Schwarz, G.: Estimating the dimension of a model. Ann. Stat. 6, 461–464 (1978)
Mallows, C.L.: Some comments on \(C_{p}\). Technometrics 15, 661–675 (1973)
Hannan, E.J., Quinn, B.G.: The determination of the order of an autoregression. J. R. Stat. Soc., Ser. B, Stat. Methodol. 41, 190–195 (1979)
Geweke, J., Meese, R.: Estimating regression models of finite but unknown order. Int. Econ. Rev. 16, 55–70 (1981)
Cavanaugh, J.E.: A large-sample model selection criterion based on kull-backs symmetric divergence. Stat. Probab. Lett. 42, 333–343 (1999)
Spiegelhalter, D.J., Best, N.G., Carlin, B.P., Linde, A.V.D.: Bayesian measures of model complexity and fit. J. R. Stat. Soc., Ser. B, Stat. Methodol. 64, 583–639 (2002)
Tsay, R.S.: Order selection in nonstationary autoregressive models. Ann. Stat. 12, 1151–1596 (1984)
Hurvich, C.M., Tsai, C.L.: Regression and time series model selection in small samples. Biometrika 76, 297–307 (1989)
Pötscher, B.M.: Model selection under nonstationarity: autoregressive models and stochastic linear regression models. Ann. Stat. 17, 1257–1274 (1989)
Andrews, D.W.K.: Consistent moment selection procedures for generalized method of moments estimation. Econometrica 67, 543–564 (1999)
Andrews, D.W.K., Lu, B.: Consistent model and moment selection criteria for GMM estimation with applications to dynamic panel models. J. Econom. 101, 123–164 (2001)
Hong, H., Preston, B., Shum, M.: Generalized empirical likelihood-based model selection criteria for moment condition models. Econ. Theory 19, 923–943 (2003)
Zellner, A., Keuzenkamp, H.A., Mcaleer, M.: Simplicity, Inference and Modelling: Keeping It Sophisticatedly Simple. Cambridge University Press, Cambridge (2001)
Owen, A.B.: Empirical likelihood ratio confidence intervals for a single functional. Biometrika 75, 237–249 (1988)
Owen, A.B.: Empirical Likelihood. Chapman and Hall, New York (2001)
Qin, J., Lawless, J.: Empirical likelihood and general estimating equations. Ann. Stat. 22, 300–325 (1994)
Acknowledgements
This work is supported by National Natural Science Foundation of China (No. 11571138, 11671054, 11301137, 11271155, 11371168, J1310022, 11501241), the National Social Science fund of China (16BTJ020), Science and Technology Research Program of Education Department in Jilin Province for the 12th Five-Year Plan (440020031139). “Thirteenth Five-Year Plan” Science and Technology Research Project of the Education of Jilin Province (Grant No. 2016103) and Jilin Province Natural Science Foundation (20130101066JC, 20130522102JH, 20150520053JH).
Author information
Authors and Affiliations
Contributions
All the authors read and approved the final manuscript.
Corresponding author
Ethics declarations
Competing interests
The authors declare that they have no competing interests.
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.
About this article
Cite this article
Zhao, Z., Liu, Y. & Peng, C. Variable selection in generalized random coefficient autoregressive models. J Inequal Appl 2018, 82 (2018). https://doi.org/10.1186/s13660-018-1680-4
Received:
Accepted:
Published:
DOI: https://doi.org/10.1186/s13660-018-1680-4