Abstract
With the advent of web-based technology, online testing is becoming a mainstream mode in large-scale educational assessments. Most online tests are administered continuously in a testing window, which may post test security problems because examinees who take the test earlier may share information with those who take the test later. Researchers have proposed various statistical indices to assess the test security, and one most often used index is the average test-overlap rate, which was further generalized to the item pooling index (Chang & Zhang, 2002, 2003). These indices, however, are all defined as the means (that is, the expected proportion of common items among examinees) and they were originally proposed for computerized adaptive testing (CAT). Recently, multistage testing (MST) has become a popular alternative to CAT. The unique features of MST make it important to report not only the mean, but also the standard deviation (SD) of test overlap rate, as we advocate in this paper. The standard deviation of test overlap rate adds important information to the test security profile, because for the same mean, a large SD reflects that certain groups of examinees share more common items than other groups. In this study, we analytically derived the lower bounds of the SD under MST, with the results under CAT as a benchmark. It is shown that when the mean overlap rate is the same between MST and CAT, the SD of test overlap tends to be larger in MST. A simulation study was conducted to provide empirical evidence. We also compared the security of MST under the single-pool versus the multiple-pool designs; both analytical and simulation studies show that the non-overlapping multiple-pool design will slightly increase the security risk.
Similar content being viewed by others
Notes
Suppose we can assemble \(\prod_{t=1}^{T}\prod_{t=1}^{S_{t}}P_{it}\) all possible panels from the item bank by fully mixing-and-matching all possible parallel forms, then in the nonpanel assumption, the probability that one form is assigned to a test taker is \(\frac{1}{N_{t}}\), whereas with a typical panel structure, the probability that the same form is assigned to a test taker is \(\frac{1}{P_{it}}\frac{P_{it}}{N_{t}}=\frac{1}{N_{t}}\), where \(\frac{1}{P_{it}}\) represents the probability that a randomly chosen panel contains that particular form, and \(\frac{P_{it}}{N_{t}}\) denotes the probability that the form will be chosen at stage t. Apparently, the two probabilities are the same.
References
Ariel, A., Veldkamp, B.P., & van der Linden, W.J. (2004). Constructing rotating item pools for constrained adaptive testing. Journal of Educational Measurement, 41, 345–359.
Barrada, J.R., Olea, J., & Abad, F.J. (2008). Rotating item banks versus restriction of maximum exposure rates in computerized adaptive testing. The Spanish Journal of Psychology, 11, 618–625.
Breithaupt, K., & Hare, D.R. (2007). Automated simultaneous assembly of multistage testlets for a high-stakes licensing examination. Educational and Psychological Measurement, 67(1), 5–20.
Chang, H.-H. (2004). Understanding computerized adaptive testing: from Robbins-Monro to Lord and beyond. In D. Kaplan (Ed.), The Sage handbook of quantitative methodology for the social sciences (pp. 117–133). Thousand Oaks: Sage.
Chang, H., Wang, S., & Ying, Z. (1997). Three dimensional visulization of item/test information. Paper presented at the annual meeting of American Educational Research Association, Chicago, IL.
Chang, H.-H., & Ying, Z. (1999). Alpha-stratified multistage computerized adaptive testing. Applied Psychological Measurement, 23, 211–222.
Chang, H.-H., & Zhang, J. (2002). Hypergeometric family and item overlap rates in computerized adaptive testing. Psychometrika, 67, 387–398.
Chang, H., & Zhang, J. (2003, April). Assessing CAT security breaches by the item pooling index—to compromise a CAT item bank, how many thieves are needed? Paper presented at the annual meeting of the National Council on Measurement in Education, Chicago.
Cheng, Y., & Chang, H. (2009). The maximum priority index method for severely constrained item selection in computerized adaptive testing. British Journal of Mathematical & Statistical Psychology, 62, 369–383.
Chen, S.Y., Ankenmann, R.D., & Spray, J.A. (2003). The relationship between item exposure and test overlap in computerized adaptive testing. Journal of Educational Measurement, 40, 129–145.
Davey, T., & Nering, M. (2002). Controlling item exposure and maintaining item security. In C. Mills, M.T. Potenza, J.J. Fremer, & W.C. Ward (Eds.), Computer-based testing: building the foundation for future assessments. Mahwah: Lawrence Erlbaum Associates.
Dean, V., & Martineau, J. (2012). A state perspective on enhancing assessment and accountability systems through systematic implementation of technology. In R.W. Lissitz & H. Jiao (Eds.), Computers and their impact on state assessment: recent history and predictions for the future (pp. 55–77). Charlotte: Information Age Publisher.
Finkelman, M., Nering, M.L., & Roussos, L.A. (2009). A conditional exposure control method for multidimensional adaptive testing. Journal of Educational Measurement, 46(1), 84–103.
Hendrickson, A. (2007). An NCME instructional module on multistage testing. Educational Measurement, Issues and Practice, 26, 44–52.
Kim, H., & Plake, B. (1993, April). Monte Carlo simulation comparison of two-stage testing and computerized adaptive testing. Paper presented at the meeting of the National Council on Measurement in Education, Atlanta, GA.
Lim, E. (2010). The effectiveness of using multiple item pools to increase test security in computerized adaptive testing. Unpublished doctoral thesis, University of Illinois at Urbana-Champaign.
Luecht, R.M., & Nungester, R.J. (1998). Some practical examples of computer-adaptive sequential testing. Journal of Educational Measurement, 35(3), 229–249.
Mills, C.N., & Steffen, M. (2000). The GRE computer adaptive test: operational issues. In W.J. van der Linden & C.A.W. Glas (Eds.), Computerized adaptive testing: theory and practice (pp. 75–99). Dordrecht: Kluwer.
Stocking, M.L., & Lewis, C. (1998). Controlling item exposure conditional on ability in computerized adaptive testing. Journal of Educational and Behavioral Statistics, 23, 57–75.
Wang, C., & Chang, H.-H. (2008, June). Continuous a-stratification index in computerized item selection. Paper presented at the annual meeting of the Psychometric Society, Durham, NH.
Way, W.D. (1998). Protecting the integrity of computerized testing item pools. Educational Measurement, Issues and Practice, 17, 17–27.
Way, W., Zara, A., & Leahy, J. (1996, April). Modifying the NCLEX TM CAT item selection algorithm to improve item exposure. Paper presented at the Annual Meeting of the American Educational Research Association, New York, NY.
Yi, Q., Zhang, J., & Chang, H.-H. (2008). Severity of organized item theft in computerized adaptive testing: a simulation study. Applied Psychological Measurement, 32(7), 543–558.
Zhang, J., Chang, H.-H., & Yi, Q. (2012). Comparing single-pool and multiple-pool designs regarding test security in computerized testing. Behavior Research Methods, 44, 742–752.
Author information
Authors and Affiliations
Corresponding author
Appendix
Appendix
1.1 A.1 Proof of Theorem 1
(1) MST The average overlap rate for MST when the random route rule is used is as follows:
where m f is the number of times a form is administered, l is the number of items within each form, p is the number of examinees in total, and \(r_{f} = \frac{m_{f}}{p}\) is the form exposure rate.
According to this derivation, first, the number of items within each form (i.e., the test length) will not affect the test security.
To decrease the average test overlap rate, the objective is to minimize \(\frac{1}{T}\sum_{t = 1}^{T} \sum_{f = 1}^{N_{t}} r_{f}^{2}\) while subject to the constraint that \(\bar{r}_{f} = \frac{1}{N_{t}}\).
Notice that
The conclusion is: when random selection is considered, only N t affects the results, that is, the number of alternative forms in each stage. When more informative routing rule is considered, for fixed N t , the value of P it and S t will influence the estimation precision.
(2) CAT When the CAT item pool is composed of all the items in the MST panels, and assume the item selection is purely random, the average overlap rate would be (Chen et al. 2003):
1.2 A.2 Proof for Theorem 2
(1) CAT The probability that two randomly selected examinees have x same items in common is
where S is the total number of items in the bank, L is the test length.
The variance of the overlap rate is simply the variance of the hypergeometric distribution, expressed as
(2) MST First, \(S = \sum_{t = 1}^{T} N_{t}l\), and \(\sum_{t = 1}^{T} l = Tl = L\). The probably that two randomly selected examinees have one form in common at stage t is
The probability that two randomly selected examinees have x forms in common during the entire test is
where P represents permutation, Ω x is the set of forms (stages) that are in common. The variance of the pair-wise overlap rate is \({\mathop{\operatorname{var}}} ( \frac{xl}{L} ) = \frac{1}{T^{2}}{\mathop{\operatorname{var}}} (x)\), where as \({\mathop{\operatorname{var}}} (x)\) can be derived easily for the given probability mass function in (A.3). The final closed form of the variance is \(\sum_{t = 1}^{T} \frac{N_{t} - 1}{N_{t}^{2}}\).
(3) Proof of the inequality
To prove \(\sigma_{\mathrm{CAT}}^{2} \le \sigma_{\mathrm{MST}}^{2}\), we need to show
because
According to Cauchy inequality, \(\sum_{t = 1}^{T} N_{t} \sum_{t = 1}^{T} \frac{1}{N_{t}} \ge T^{2}\), therefore,
Because
Combining (A.4) and (A.5), the inequality holds.
1.3 A.3 Proof of Theorem 3
(1) CAT For CAT, the mean and variance of the overlap will change accordingly by plugging in the new item bank size \(\sum_{t = 1}^{T} (N_{t}l - n_{t})\).
(2) MST Let x t denotes the number of common items between two randomly picked examinees at stage t.
Let X denote the number of common items between two randomly selected examinees, then
It is straightforward to compute that
Because x t ’s are independent across different stages,
Rights and permissions
About this article
Cite this article
Wang, C., Zheng, Y. & Chang, HH. Does Standard Deviation Matter? Using “Standard Deviation” to Quantify Security of Multistage Testing. Psychometrika 79, 154–174 (2014). https://doi.org/10.1007/s11336-013-9356-y
Received:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11336-013-9356-y