Abstract
This paper studies goodness-of-fit tests for the bivariate Poisson distribution. Specifically, we propose and study several Cramér–von Mises type tests based on the empirical probability generating function. They are consistent against fixed alternatives for adequate choices of the weight function involved in their definition. They are also able to detect local alternatives converging to the null at a certain rate. The bootstrap can be used to consistently estimate the null distribution of the test statistics. A simulation study investigates the goodness of the bootstrap approximation and compares their powers for finite sample sizes. Extensions for testing goodness-of-fit for the multivariate Poisson distribution are also discussed.
Similar content being viewed by others
References
Baringhaus L, Gürtler N, Henze N (2000) Weighted integral test statistics and components of smooth tests of fit. Aust N Z J Stat 42:179–192
Baringhaus L, Henze N (1992) A goodness of fit test for the Poisson distribution based on the empirical generating function. Stat Prob Lett 13:269–274
Berkhout P, Plug E (2004) A bivariate Poisson count data model using conditional probabilities. Stat Neerlandica 58(3):349–364
Crockett NG (1979) A quick test of fit of a bivariate distribution. In: McNeil D (ed) Interactive statistics. North-Holland, Amsterdam, pp 185–191
Feuerverger A (1989) On the empirical saddlepoint approximation. Biometrika 76:457–464
Gillings DB (1974) Some further results for bivariate generalizations of the Neyman Type A distribution. Biometrics 30:619–628
Gregory G (1977) Large sample theory for \(U\)-statistics and tests of fit. Ann Stat 5:110–123
Gürtler N, Henze N (2000) Recent and classical goodness-of-fit tests for the Poisson distribution. J Stat Plan Inference 90:207–225
Hamdan MA (1972) Estimation in the truncated bivariate Poisson distribution. Technometrics 14:37–45
Ho LL, Singer JM (2001) Generalizad least squares methods for bivariate Poisson regression. Commun Stat Theory Methods 30(2):263–277
Holgate P (1964) Estimation for the bivariate Poisson distribution. Biometrika 51:241–245
Holgate P (1966) Bivariate generalizations of Neyman’s Type A distribution. Biometrika 53:241–245
Johnson NL, Kotz S, Balakrishnan N (1997) Discrete multivariate distributions. Wiley, New York
Karlis D, Ntzoufras I (2005) Bivariate Poisson and diagonal inflated bivariate Poisson regression models in R. J Stat Softw 14(10):1–36
Karlis D, Tsiamyrtzis P (2008) Exact Bayesian modeling for bivariate Poisson data and extensions. Stat Comput 18:27–40
Kocherlakota S, Kocherlakota K (1992) Bivariate discrete distributions. Marcel Dekker Inc, New York
Kundu S, Majumdar S, Mukherjee K (2000) Central limits theorems revisited. Stat Probab Lett 47:265–275
Lehmann EL, Romano JP (2005) Testing statistical hypothesis. Springer, Berlin
Loukas S, Kemp CD (1986) The index of dispersion test for the bivariate Poisson distribution. Biometrics 42:941–948
Meintanis S, Swanepoel J (2007) Bootstrap goodness-of-fit tests with estimated parameters based on empirirical transforms. Stat Probab Lett 77:1004–1013
Papageorgiou H, Loukas S (1988) Conditional even point estimation for bivariate discrete distributions. Commun Stat Theory Methods 17:3403–3412
Rayner JCW, Best DJ (1995) Smooth tests for the bivariate Poisson distribution. Aust N Z J Stat 37:233–245
Rueda R, Pérez-Abreu V, O’Reilly F (1991) Goodness of fit for the Poisson distribution based on the probability generating function. Commun Stat Theory Methods A20:3093–3110
Serfling RJ (1980) Approximation theorems of mathematical statistics. Wiley, New York
Acknowledgments
The authors would like to thank Professor Norbert Henze and an anonymous referee for their careful reading and comments which greatly improved the paper. F. Novoa-Muñoz also wishes to thank his institution, University of Bío-Bío (Chile) and the scholarship given by the Chilean Ministry of Education through the Superior Education MECE Program 2, which make his doctoral studies possible.
Author information
Authors and Affiliations
Corresponding author
Appendix
Appendix
1.1 Proofs
Here we prove the results in Sects. 2–5, 7. Since the mathematical machinery employed in the proofs is quite standard, we just sketch the proofs. A detailed derivation of the results can be obtained from the authors upon request.
In order to prove Proposition 1, we first give a preliminary lemma. For an arbitrary set \(S, \partial S\) and \(int S\) denote the sets of boundary points and interior points of \(S\), respectively.
Lemma 1
Let \(\{f_n\}_{n\ge 1}\) and \(f\) be a collection of nondecreasing real functions defined on \(R=[b_1,c_1]\times [b_2,c_2]\times \cdots \times [b_d,c_d] \subseteq \mathbb {R}^d\), with \(-\infty <b_j\le c_j<\infty , j=1,2,\ldots ,d\). Let \(D=D_1\cup D_2 \cup D_3\), where \(D_1\) is the vertex set of \(R, D_2\) is a dense set in \(\partial R\) and \(D_3\) is a dense set in \(int R\). If \(f\) is continuous on \(R\) and
then
Proof
Let \(\varepsilon >0\) be arbitrary but fixed. Because \(f\) is a uniformly continuous function on \(R, \exists \delta =\delta (\varepsilon )>0\), such that \(\forall x,y\in R\) with \(\Vert x-y\Vert <\delta \), we have that \(|f(x)-f(y)|<\varepsilon \). Let
with \(\Vert u_i-v_i\Vert < \delta \), where \(u_i=(u_{1i},u_{2i},\ldots ,u_{di}) \in D\) and \(v_i=(v_{1i},v_{2i},\ldots ,v_{di}) \in D, 1\le i \le M\), such that \(R=\bigcup _{i=1}^M H_i.\) Let \(\Delta _n=\max _{x\in V} \left| f_n(x)-f(x)\right| \), where \(V=\{u_i,v_i: i=1,2,\ldots ,M\}\). From (14), it follows that \(\Delta _n \rightarrow 0\). Now, for each \(x\in R\), there exists \(i\in \{1,2,\ldots ,M\}\), such that \(x\in H_i\), and thus \(u_i\le x\le v_i\). Since \(f_n\) and \(f\) are nondecreasing, we get
Analogously, \(f_n(x)-f(x)\ge -\Delta _n-\varepsilon \). Since \(\varepsilon >0\) is arbitrary, these two inequalities imply the result. \(\square \)
Proof of Proposition 1
Let \(D\) be a countable dense set in \(R\) according to Lemma 1. By the SLLN, there exists a set \(A_D \in \mathcal {A}\) such that \(P(A_D)=1\) and for all \(\omega \in A_D\): \(g_n(u,\omega )\ \mathop {\longrightarrow } \ g(u), \forall u \in D\). Since \(X_i\ge 0, 1\le i \le d\), we have that \(g_n(u,\omega )\) and \(g(u)\) are nondecreasing functions. Now (2) follows from Lemma 1. The proof of (3) follows similar steps. \(\square \)
Proof of Proposition 2
Let \((X_1,X_2)\) be a random vector and let \(g(u_1,u_2)=E\left( u_1^{X_1} u_2^{X_2}\right) \) be its pgf. Then, from the first equation in (6)
Integrating (15) over \(u_1\), we obtain
where \(\phi (u_2)=\phi _1(u_2)+\theta _1+\theta _3(u_2-1)\). Proceeding similarly, from the second equation in (6) we get
Thus, necessarily \(\varphi (u_1)=\theta _1(u_1-1)\) and \(\phi (u_2)=\theta _2(u_2-1)\), in other words, the BDP is the only solution of (6) \(\square \)
Proof of Theorem 1
By definition, \(R_{n,w}(\hat{{\theta }}_n)= ||G_n(\hat{{\theta }}_n)||_{\mathcal {H}}^2\). Note that
with
To prove (7), we apply a Taylor expansion of \(G(\varvec{X}_i,\hat{\theta }_n; {u})\) around \(\hat{\theta }_n=\theta \). Now following the lines in the proof of Theorem 3.1 in Meintanis and Swanepoel (2007) we get (7). Observe that
where, \(h(\varvec{x},\varvec{y};\theta )\) is defined in (10) and satisfies \(h(\varvec{x},\varvec{y};\theta )= h(\varvec{y},\varvec{x};\theta ), E_{\theta }\left\{ h^2(\varvec{X}_1,\right. \left. \varvec{X}_2;\theta )\right\} < \infty , E_{\theta } \left\{ |h(\varvec{X}_1,\varvec{X}_1;\theta )|\right\} < \infty \) and \(E_{\theta }\{h(\varvec{X}_1,\varvec{X}_2;\theta )\}=0\). Thus, from Theorem 6.4.1.B in Serfling (1980),
where \(\chi ^2_{11},\chi ^2_{12},\ldots \) and the set \(\{\lambda _j\}\) are as defined in the statement of the Theorem. In particular, \(\Vert R_{n}\Vert _{_\mathcal {H}}^{2}=O_P(1)\), which implies (8). \(\square \)
Proof of Theorem 2
By definition \(R_{n,w}^*(\hat{\theta }_n^*)= \Vert G_{n}^*(\hat{\theta }_n^*)\Vert _{\mathcal {H}}^{2}\), with
and \(G(\varvec{X};\theta ; u)\) defined in (16). Following similar steps to those given in the proof of Theorem 1 it can be seen that \(R_{n,w}^*(\hat{\theta }_n^*)= \Vert R_{n}^*\Vert _{\mathcal {H}}^{2}+r^*_{n}\) with \(r_{n}^*=o_{P_*}(1)\) a.s., where \(R_{n}^*(u)\) is defined as \(R_{n}(u)\) with \(\varvec{X}_i\) and \(\theta \) replaced by \(\varvec{X}^*_i\) and \(\hat{\theta }_n\), respectively. Now, the result follows by applying Theorem 1.1 in Kundu et al. (2000) to \(R_{n}^*\) and the continuous mapping theorem. \(\square \)
Proof of Theorem 3
We only present the proof of part (b), since the proof of part (a) is quite similar. By definition,
We have
\(\forall u \in [0,1]^2\), where \(r_0=\sup _{u\in [0,1]^2}\left| g_n(u)- g(u)\right| \) and \(r_i=\sup _{u\in [0,1]^2}\left| \frac{\partial }{\partial u_i}g_n(u)-\right. \left. \frac{\partial }{\partial u_i}g(u)\right| , i=1,2\). Since \(r_i=o(1), i=0,1,2, |\theta _i-\hat{\theta }_{in}| =o(1), i=1,2,3\), and (5), we conclude that \(\frac{1}{n}S_{n,w}(\hat{\theta }_n)=\Vert E_{1}\Vert ^2_{\mathcal {H}}+ \Vert E_{2}\Vert ^2_{\mathcal {H}}+o(1)\), which shows the result because \(\Vert E_{1}\Vert ^2_{\mathcal {H}}+ \Vert E_{2}\Vert ^2_{\mathcal {H}}=\sigma (g;\theta )\). \(\square \)
Proof of Theorem 4
Let \(A_n \subseteq \mathbb {N}_0^2\) so that \(P_{\theta }(A_n)\rightarrow 0\), then it can be easily seen that \(P_n(A_n)\rightarrow 0\), and thus \(P_n\) is contiguous to \(P_{\theta }\). Next we show that, under \(P_n, \sqrt{n}(\hat{\theta }_n-\theta )\) converges in distribution to a normal law. With this aim, let \(\varvec{X}_1, \varvec{X}_2, \ldots , \varvec{X}_n\) be iid with common pmf \(P_{\theta }\) and let
Let \(Z_{n,i}=b_n(\varvec{X}_i), 1\le i \le n\). Then, by a Taylor expansion,
where \(\varrho _n=\varrho _n(Z_{n,1},Z_{n,2}, \ldots , Z_{n,n})\),
with \(a_{in}=a_{in}(Z_{n,i})\)
for some \(0<\alpha _i<1, 1\le i \le n\). Next, we study the limit of each term in the right side of (17).
Let \(\varepsilon \) be a positive constant. Let \(M_n=\sqrt{n}/2\) and \(\tilde{\varrho }_n=\varrho _n(Z_{n,1}I\{|Z_{n,1}|\le M_n\}, \ldots , Z_{n,n}I\{|Z_{n,n}|\le M_n\})\). Note that \(\tilde{a}_{in}=a_{in}(Z_{n,i}I\{|Z_{n,i}|\le M_n\})\ge 0.5, 1\le i \le n\), and therefore
which implies
We also have
From (18) and (19), we conclude
From Assumption 3, the central limit theorem for triangular arrays implies
with \(\sigma ^2=E_{\theta }\left\{ b(X_1,X_2)^2\right\} <\infty \).
From Assumption 3 we also have \(E_{\theta }\left( \frac{1}{n}\sum _{i=1}^n Z_{n,i}^2\right) \rightarrow \sigma ^2\) and \(Var_{\theta }\left( \frac{1}{n}\sum _{i=1}^n\right. \left. Z_{n,i}^2\right) = \frac{1}{n}Var_{\theta }(Z_{n,i}^2)\le \frac{1}{n}\sup _n E_{\theta }\left\{ b_n(X_1,X_2)^4\right\} \rightarrow 0\), which imply
Now, from Assumption 1, (17) and (20)–(22), the sequence \((\sqrt{n}(\hat{\theta }_n-\theta ), l_n)\) converges in law to a 4-dimensional normal variate when the data comes from \(P_{\theta }\). Thus, from Le Cam’s Third Lemma (Corollary 12.3.2 in Lehmann and Romano 2005), we have that, when the data come from \(P_n, \sqrt{n}(\hat{\theta }_n-\theta )\) converges in distribution to a normal law, which implies that it is bounded in probability. Now, following similar steps as those given in the proof of Theorem 1, it can be shown that (7) also holds when the data come from \(P_n\), with \(P_n(|r_n|>\varepsilon )\rightarrow 0, \forall \varepsilon >0\). Finally, applying Theorem 2.3 in Gregory (1977), we obtain that when the data have pmf \(P_n\)
and thus the result. \(\square \)
Proof of Proposition 3
The proof follows is quite similar to that of Proposition 2, so we omit it. \(\square \)
1.2 Expressions for statistics \(\varvec{T}, \varvec{I_B}\) and \(\varvec{NI_B}\) and critical regions
We have
where \(S^2_{X_i}\) is the sample variance, \(i=1,2, S^2_{X_1X_2}\) is the sample covariance, \(r\) is the sample correlation coefficient and \(\chi ^2_{k, \alpha }\), for \(0<\alpha <1\) and \(k\in \mathbb {N}\), denotes the upper \(\alpha \)-percentile of the \(\chi ^2\) distribution with \(k\) degrees of freedom.
1.3 Computation of the test statistic \(R_{n,w}(\hat{\theta }_n)\)
For \(w\) as in (12), for \(a_1, a_2\in \mathbb {N}_0\), we get
To compute \(R_{n,w}(\hat{\theta }_n)\) we can calculate the integrals on the right hand side of (23). Nevertheless, for computational reasons, the following representation (which is obtained by interchanging the order of integration)
turned out to be more appropriate, where \(p(i, j;\theta )=P_{\theta }(X_1=i, X_2=j)\) and \(p_n(i,j)\) is the relative frequency of the pair \((i,j)\). Besides, (24) is less restrictive, since it allows \(a_1>-1\), and \(a_2>-1\). A truncation of the four infinite series at \(M+15\) yielded sufficiently precise values of the statistic and a good performance of the corresponding subroutine, where \(\displaystyle M\!=\!\max \{ X_{1(n)},X_{2(n)}\}, X_{k(n)}\!=\!\max \nolimits _{1\le i\le n}X_{ki}, k=1,2\).
1.4 Computation of the test statistic \(S_{n,w}(\hat{\theta }_n)\)
When the weight function is chosen as in (12), for some \(a=(a_1, a_2)\) with \(a_i>-1, i=1,2\), we have
with
\(k=1,2, 1\le i,j \le n\), where \( B_{rs}=\{X_{rs}\ge 1\}, r=1,2, 1\le s \le n\).
Rights and permissions
About this article
Cite this article
Novoa-Muñoz, F., Jiménez-Gamero, M.D. Testing for the bivariate Poisson distribution. Metrika 77, 771–793 (2014). https://doi.org/10.1007/s00184-013-0464-6
Received:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00184-013-0464-6