Abstract
In this paper, we consider the penalized estimation procedure for Poisson autoregressive model with sparse parameter structure. We study the theoretical properties of penalized conditional maximum likelihood (PCML) with several different penalties. We show that the penalized estimators perform as well as the true model was known. We establish the oracle properties of PCML estimators. Some simulation studies are conducted to verify the proposed procedure. A real data example is also provided.
Similar content being viewed by others
References
Al-Osh MA, Alzaid AA (1987) First-order integer-valued autoregressive (INAR (1)) process. J Time Ser Anal 8(3):261–275
Al-Osh MA, Alzaid AA (1988) Integer-valued moving average (INMA) process. Stat Pap 29(1):281–300
Brillinger DR (2001) Time series: data analysis and theory. SIAM
Davis RA, Dunsmuir WTM, Streett SB (2003) Observation-driven models for Poisson counts. Biometrika 90(4):777–790
Dicker L, Huang B, Lin X (2013) Variable selection and estimation with the seamless-L0 penalty. Stat Sin:929–962
Doukhan P, Fokianos K, Tjøtheim D (2012) On weak dependence conditions for Poisson autoregressions. Stat Prob Lett 82:942–48
Fan J, Li R (2001) Variable selection via nonconcave penalized likelihood and its oracle properties. J Am Stat Assoc 96(456):1348–1360
Fan J, Lv J (2010) A selective overview of variable selection in high dimensional feature space. Stat Sin 20:101–148
Fan J, Peng H (2004) On nonconcave penalized likelihood with diverging number of parameters. Ann Stat 32:928–961
Ferland R, Latour A, Oraichi D (2006) Integer-valued GARCH process. J Time Ser Anal 27(6):923–942
Fokianos K, Rahbek A, Tjøstheim D (2009) Poisson autoregression. J Am Stat Assoc 104(488):1430–1439
Hall P, Heyde CC (2014) Martingale limit theory and its application. Academic press, New York
Kedem B, Fokianos K (2005) Regression models for time series analysis. Wiley, New York
Khoo WC, Ong SH, Biswas A (2017) Modeling time series of counts with a new class of INAR (1) model. Stat Pap 58:393–416
Nardi Y, Rinaldo A (2011) Autoregressive process modeling via the LASSO procedure. J Multivar Anal 102(3):528–549
Steutel FW, Van Harn K (1979) Discrete analogues of self-decomposability and stability. Ann Prob:893–899
Wang H, Li G, Tsai CL (2007a) Regression coefficient and autoregressive order shrinkage and selection via the LASSO. J R Stat Soc 69(1):63–78
Wang H, Li R, Tsai CL (2007b) Tuning parameter selectors for the smoothly clipped absolute deviation method. Biometrika 94(3):553–568
Yang K, Wang D, Jia B, Li H (2016) An integer-valued threshold autoregressive process based on negative binomial thinning. Stat Pap. doi:10.1007/s00362-016-0808-1
Yoon YJ, Park C, Lee T (2013) Penalized regression models with autoregressive error terms. J Stat Comput Simul 83(9):1756–1772
Zhang CH (2010) Nearly unbiased variable selection under minimax concave penalty. Ann Stat:894–942
Zhang H, Wang D, Zhu F (2010) Inference for INAR (p) processes with signed generalized power series thinning operator. J Stat Plan Inference 140(3):667–683
Zheng H, Basawa IV, Datta S (2006) Inference for pth-order random coefficient integer-valued autoregressive processes. J Time Ser Anal 27(3):411–440
Zhu F, Wang D (2011) Estimation and testing for a Poisson autoregressive model. Metrika 73(2):211–230
Zou H (2006) The adaptive LASSO and its oracle properties. J Am Stat Assoc 101(476):1418–1429
Acknowledgements
We thank the Editor and two reviewers for their valuable suggestions and comments which greatly improved the article. This work is supported by National Natural Science Foundation of China (No. 11271155, 11371168, J1310022, 11571138, 11501241, 11571051, 11301137, 11301212 and 11401146), National Social Science Foundation of China (16BTJ020), Science and Technology Research Program of Education Department in Jilin Province for the 12th Five-Year Plan (440020031139) and Jilin Province Natural Science Foundation (20150520053JH).
Author information
Authors and Affiliations
Corresponding author
Appendix
Appendix
To prove Theorem 1, we need the following lemma.
Lemma A.1
Under condition (C.1), as \(n\rightarrow \infty \) we have
where \(B({\varvec{\theta }}_{\varvec{0}})=\sum ^{n}_{t=1}\frac{\partial l_{n}({\varvec{\theta }}_{\varvec{0}})}{\partial {\varvec{\theta }}}\); the Fisher information matrix \({\varvec{\Sigma }}({\varvec{\theta }_{\varvec{0}}})=E\left( \frac{{\varvec{Y}}_{t}{\varvec{Y}}_{t}^{\mathrm {T}}}{\gamma _{t}}\right) \) with \({\varvec{Y}}_{t}=(1,X_{t-p},\ldots ,X_{t-1})^{\mathrm {T}}\).
Proof of Lemma A.1
Let
Through some calculation, we can derive that
which implies that \(\{T_{n1},\mathscr {F}_{n},n\ge 1\}\) is a martingale with \(\mathscr {F}_{n}=\sigma \left( X_{n},X_{n-1},\ldots ,X_{0}\right) \). By \(E|X_{t}|^{4}<\infty \), the strict stationarity of \(\{X_{t}\}\), and the ergodic theorem, we obtain that
Using the martingale central limit theorem (Hall and Heyde 2014), we get that
Similarly, we can prove \(\{T_{ni},\mathscr {F}_{n},n\ge 1\}\), \(i=2,\ldots ,p+1\) is a martingale and
For any \(\mathbf c =(c_{1},\ldots ,c_{p+1})^{\mathrm {T}}\in \mathbb {R}^{p+1}\backslash (0,\ldots ,0)^{\mathrm {T}}\), we get
Thus, by the Cramer-Wold device,
This end this proof. \(\square \)
Proof of Theorem 1
Let \(\beta _{n}=(n^{-1/2}+a_{n})\), following Fan and Li (2001), we need to show that for any \(\varepsilon >0\), there exists a constant d, such that
which implies that there exists a local maximum in the ball \(\{\varvec{\theta _{0}}+\beta _{n}{} \mathbf u :\Vert \mathbf u \Vert \le d\}\) with probability at least \(1-\varepsilon \), then there exists a local maximizer with \(\Vert \varvec{\hat{\theta }}-\varvec{\theta _{0}}\Vert =O_{p}(\beta _{n})\). Note that
By Taylor series expansion, we obtain
where
From Lemma A.1, we know that \(n^{-1/2}B(\varvec{\theta _{0}})=O_{p}(1)\), then \(A_{1}=O_{p}(n^{1/2}\beta _{n})=O_{p}(n\beta ^{2}_{n})\). By ergodicity, we get \(A_{2}= -n\beta _{n}^{2}{} \mathbf u ^{\mathrm {T}}\varvec{\Sigma }(\varvec{\theta _{0}})\mathbf u \), as \(n\rightarrow \infty \). From conditions (C.2) and (C.3), we have \(A_{3}\) is bounded by \(\sqrt{s}\beta _{n}a_{n}\Vert \mathbf u \Vert +n\beta _{n}^{2}b_{n}\Vert \mathbf u \Vert ^{2}.\) By choosing a sufficient large d, both \(A_{1}\) and \(A_{3}\) are dominated by \(A_{2}\). The proof is completed. \(\square \)
Proof of Lemma 1
We need to prove that with probability tending to one, as \(n\rightarrow \infty \) for any \(\varvec{\theta _{1}}\) satisfying \(\Vert \varvec{\theta _{1}}-\varvec{\theta _{10}}\Vert =O_{p}(n^{-1/2})\) and for some small \(\epsilon _{n}=\eta n^{-1/2}\) and \(j=s+1,\ldots ,p+1\)
To show (9), by Taylor’s expansion,
From Lemma A.1, we know that \(\frac{\partial L_{n}(\varvec{\theta _{0}})}{\partial \theta _{j}}=O_{p}(n^{1/2})\). By law of large numbers, strict stationarity and \(\Vert \varvec{\theta _{1}}-\varvec{\theta _{10}}\Vert =O_{p}(n^{-1/2})\), we have
Thus, \(\frac{\partial Q_{n}(\varvec{\theta })}{\partial \theta _{j}}=n\lambda _{n}\left\{ O_{p}(n^{-1/2}/\lambda _{n})-\lambda _{n}^{-1}\dot{P}_{\lambda _{n}}(|\theta _{j}|)\text {sgn}(\theta _{j})\right\} \). Since \(n^{-1/2}/\lambda _{n}\rightarrow 0\) and \(\lambda _{n}^{-1}\dot{P}_{\lambda _{n}}(|\theta _{j}|)>0\) as \(n\rightarrow \infty \). The sign of (11) is dominated by that of \(\theta _{j}\). Hence, (10) follows. This completes the proof. \(\square \)
Proof of Theorem 2
Part (\(i\)) holds by Lemma 1. We only need to prove (\(ii\)). From part (\(i\)), we know that \(\varvec{\hat{\theta }_{2}}=\mathbf 0 \) with probability tending to 1. Thus, there exists a root-n consistent local maximum estimator \(\varvec{\hat{\theta }_{1}}\) satisfies the following equation
By the Taylor expansion, we have
This indicates
where \(B^{s}(\varvec{\theta }_{\varvec{0}})=\sum ^{n}_{t=1}\frac{1}{\gamma _{t}}\left( X_{t}-\gamma _{t}\right) {\varvec{Y}}^{s}_{t}\) and \({\varvec{Y}}^{s}_{t}=(1,X_{t-p},\ldots ,X_{t-p+s-2})^{\mathrm {T}}\). From the Slutskys theorem and the martingale central limit theorem, we complete the proof. \(\square \)
Rights and permissions
About this article
Cite this article
Wang, X., Wang, D. & Zhang, H. Poisson autoregressive process modeling via the penalized conditional maximum likelihood procedure. Stat Papers 61, 245–260 (2020). https://doi.org/10.1007/s00362-017-0938-0
Received:
Revised:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00362-017-0938-0