Abstract
This article develops estimators for certain population characteristics using a judgment post stratified (JPS) sample. The paper first constructs a conditional JPS sample with a reduced set size K by conditioning on the ranks of the measured observations of the original JPS sample of set size \(H \ge K\). The paper shows that the estimators of the population mean, median and distribution function based on this conditional JPS sample are consistent and have limiting normal distributions. It is shown that the proposed estimators, unlike the ratio and regression estimators, where they require a strong linearity assumption, only need a monotonic relationship between the response and auxiliary variable. For moderate sample sizes, the paper provides a bootstrap distribution to draw statistical inference. A small-scale simulation study shows that the proposed estimators based on a reduced set JPS sample perform better than the corresponding estimators based on a regular JPS sample.
Similar content being viewed by others
References
Chen, M., Ahn, S., Wang, X., Lim, J. (2014). Generalized isotonized mean estimators for judgment post-stratification with multiple rankers. Journal of Agricultural, Biological, and Environmental Statistics, 19, 405–418.
Dastbaravarde, A., Arghami, N. R., Sarmad, M. (2016). Some theoretical results concerning non-prametric estimation by using a judsgment post-stratification sample with perfect ranking. Communications in Statistics, Theory and Methods, 45, 2182–2203.
Dell, T. R., Clutter, J. L. (1972). Ranked set sampling theory with order statistics backround. Biometrics, 28, 545–555.
Du, J., MacEachern, S. N. (2008). Judgement post-stratification for designed experiments. Biometrics, 64, 345–354.
Frey, J. (2012). Constrained nonparametric estimation of the mean and the CDF using ranked-set sampling with a covariate. Annals of the Institute of Statistical Mathematics, 64, 439–456.
Frey, J., Feeman, T. G. (2012). An improved mean estimator for judgment post-stratification. Computational Statistics and Data Analysis, 56, 418–426.
Frey, J., Feeman, T. G. (2013). Variance estimation using judgment post-stratification. Annals of the Institute of Statistical Mathematics, 65, 551–569.
Frey, J., Ozturk, O. (2011). Constrained estimation using judgment post-stratification. Annals of the Institute of Statistical Mathematics, 63, 769–789.
Hollander, M., Wolfe, D. A., Chicken, E. (2014). Nonparametric Statistical Methods (3rd ed.). New York: Wiley.
MacEachern, S. N., Stasny, E. A., Wolfe, D. A. (2004). Judgment post-stratification with imprecise rankings. Biometrics, 60, 207–215.
Ozturk, O. (2012). Quantile inference based on partially rank-ordered set samples. Journal of Statistical Planning and Inference, 142, 2116–2127.
Ozturk, O. (2013). Combining multi-observer information in partially rank-ordered judgment post-stratified and ranked set samples. The Canadian Journal of Statistics, 41, 304–324.
Ozturk, O. (2014). Statistical inference for population quantiles and variance in judgment post-stratified samples. Computational Statistics and Data Analysis, 77, 188–205.
Ozturk, O. (2015). Distribution free two-sample methods for judgment post-stratified data. Statistica Sinica, 25, 1691–1712.
Stokes, S. L., Wang, X., Chen, M. (2007). Judgment post-stratification with multiple rankers. Journal of Statistical Theory and Applications, 6, 344–359.
Wang, X., Stokes, S. L., Lim, J., Chen, M. (2006). Concomitants of multivariate order statistics with application to judgment post-stratification. Journal of the American Statistical Association, 101, 1693–1704.
Wang, X., Lim, J., Stokes, S. L. (2008). A nonparametric mean estimator for judgment post-stratified data. Biometrics, 64, 355–363.
Wang, X., Wang, K., Lim, J. (2012). Isotonized CDF estimation from judgment post-stratification data with empty strata. Biometrics, 68, 194–202.
Wolfe, D. A. (2012). Ranked set sampling: its relevance and impact on statistical inference. ISRN Probability and Statistics. doi:10.5402/2012/568385.
Acknowledgments
The author thanks Professor Jesse Frey who provided the sketch of the proof of the unbiasedness of the estimator in Theorem 1, and the Editor, Associate Editor and two anonymous referees for their helpful comments to improve the presentation of the paper.
Author information
Authors and Affiliations
Corresponding author
Appendix
Appendix
Proof of Theorem 1
For the proof of (i), from Eqs. (2) and (3) and from the law of large numbers, we write
and
It is now easy to observe that
For the proof of (ii), without loss of generality assume that H is an even integer and \(b=0\). We consider
In a JPS sample, since \(E(N_{h})= E(N_{H+1-h})\), the expected weights also preserve the equality \(E(w_h)=E(w_{H+1-h})\). By using the assumption of the theorem and the symmetry of the expected weights, we write
This completes the proof. \(\square \)
Proof of Lemma (1)
\(\bar{\varvec{T}}_{g}(b)\) is the mean of independent random vectors. Hence, central limit theorem gives the asymptotic normality. The expression \(E(\varvec{\bar{T}}_g)\) follows from Eqs. (2) and (3). We now compute the variance and covariances. For \(\sigma _{r,s}\), \(1 \le r,s \le K\), we use the conditional covariance given \(R_i\)
In the above equation the first term reduces to
In a similar approach, we have
The second term in the right side of the above expression follows from Eq. (2). Combining expressions A and B, we obtain
For \(\tau _{r,s}\), \(1 \le r \le k\) and \(1\le s \le K-1\) we compute
Finally, the \( \gamma _{r,s}\), \(1 \le r,s \le K-1\), follows from
The proof is completed. \(\square \)
Proof of Theorem 2
Let Q be a transformation from \(\varvec{\bar{T}}_g(b)\) to \(\hat{\theta }_g(b)\)
where \(\bar{T}_{g,k}(b)\) is the k th component of vector \(\varvec{\bar{T}}_g(b)\) and \(\bar{T}_{g,2K}(b)=1-\sum _{k=1}^{K-1} \bar{T}_{g,K+k}(b)\). The second equality in the above equation follows from the fact that \(I_k/d_n\) converges in probability to 1 / K for \(k=1,\ldots ,K\). It is clear that \(Q(E(\varvec{\bar{T}}_g(b)))=E(g(X-b))\). For notational convenience, let \(\varvec{\mu }_T=E(\varvec{\bar{T}}_g(b))\). By using a Taylor expansion of \(Q(\varvec{\bar{T}}_g(b))\) around \(\varvec{\mu }_T\) we write
where \(\varvec{L}_T\) is a \(2K-1\) dimensional partial derivative vector of \(Q(\varvec{\mu }_T)\),
and \(\mu _T(r)\) is the r th component of vector \(\mu _T\). Let \(\varvec{L}^\top =(\varvec{L}_1^\top ,\varvec{L}_2^\top )\) with \(\varvec{L}_1^\top =(K,\ldots ,K)\) and \(\varvec{L}_2=K(Eg(X_{[K:K]}-b)-Eg(X_{[1:K]}-b),\ldots , Eg(X_{[K:K]}-b)-Eg(X_{[K-1:K]}-b))\). It is then easy to see that \(\sqrt{n}\left( Q(\varvec{\bar{T}}_g(b))-E(g(X-b))\right) \) converges to a normal distribution with mean zero and variance
By using Lemma 1, we write
Let \(d_s=Eg(X_{[K:K]}-b)-Eg(X_{[s:K]}-b)\), \(s=1,\ldots , K-1\).
For the last expression, we consider
This completes the proof. \(\square \)
Proof of Corollary 1
We first rewrite \(\sigma ^2_{\theta :K}\) as
We need to show \(A_d\) is non-negative. We first consider
We re-write expression \(B_d\) as
Use of equalities \(\sum _{s=1}^K a_{s:K|h}=1\) and \(\sum _{h=1}^H a_{s:K|h}=H/K\) in the first term of the above equation reduces \(B_d\) to \(2A_d\)
This completes the proof. \(\square \)
Proof of Theorem 3
Let \(U_n(b/\sqrt{n})=\left\{ S_n(\eta _p+b/\sqrt{n})-S_n(\eta _p))/(b/\sqrt{n}\right\} \). By using conditional expectation given rank vector \(\varvec{R}\), we obtain
For large n, we can write
It is now easy to observe that \(E(U_n(b/\sqrt{n})\) has a limit at \(-bf(\eta _p)\)
We now show that the variance of \(U_n(b/\sqrt{n})\) converges to zero as n goes to infinity. Without loss of generality, assume that \(b>0\) and \(\eta _p=0\). In this case, \(U_n(b/\sqrt{n})\) can be written as
The variance of \(U_n(b/\sqrt{n})\) can be written as
The expression \(D_n\) is given by
In the above expression, the covariances are all finite and the double sum in the first term converges to zero as n gets large. Hence, \(D_n\) has a limit at 0. In a similar fashion, \(G_n\) can be written as
Since the expectation in the above expression has a finite limit, \(G_n\) converges to zero as n goes to infinity. We then conclude that variance \(U_n(b/\sqrt{n})\) goes to zero as n approaches to infinity. This establishes the point-wise convergence
The uniform convergence follows from the monotonicity of \(S_n(b)\). \(\square \)
Proof of Theorem 4
We first observe that sample size vector \(\varvec{N}^{\top }= (N_1,\ldots , N_H)\) has a multinomial distribution with parameters n and \((1/H,\ldots , 1/H)\). The probability that \(W_n(q)\) equals K is
Let
For a fixed j, we also define \(B_{j,H-K,q}\) to be the event that each one of the j judgment classes in set \(B_{H-K,q}\) has exactly q observations. Note that by the definition of the set \(B_{H-K,q}\), \(B_{j,H-K,q}= B_{H-K,H-K,q}\) when \(q=0\). With this new notation we write
It is clear that
and
Let
where \(A_{h:K,q}\) is the event that \(N_{h} > q\) for \(h \le K\). By using DeMorgan’s law, we write
We now evaluate the following conditional probability
where \(T_i= \sum _{y=1}^i r_y\). With some algebra, the above equation simplifies to
We complete the proof by putting this expression in (11). \(\square \)
About this article
Cite this article
Ozturk, O. Statistical inference with empty strata in judgment post stratified samples. Ann Inst Stat Math 69, 1029–1057 (2017). https://doi.org/10.1007/s10463-016-0572-y
Received:
Revised:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10463-016-0572-y