A Gibbs Sampler for the (Extended) Marginal Rasch Model

In their seminal work on characterizing the manifest probabilities of latent trait models, Cressie and Holland give a theoretically important characterization of the marginal Rasch model. Because their representation of the marginal Rasch model does not involve any latent trait, nor any specific distribution of a latent trait, it opens up the possibility for constructing a Markov chain - Monte Carlo method for Bayesian inference for the marginal Rasch model that does not rely on data augmentation. Such an approach would be highly efficient as its computational cost does not depend on the number of respondents, which makes it suitable for large-scale educational measurement. In this paper, such an approach will be developed and its operating characteristics illustrated with simulated data.


Introduction
Over the last two decades, Markov chain -Monte Carlo (MCMC) approaches to Bayesian inference for item response theory (IRT) models have become increasingly popular. Most applications follow the data augmentation Gibbs (DA-Gibbs) approach of Albert (1992) (see also, Albert & Chib, 1993) for the normal ogive model. The work of Albert (1992) has been extended in many directions, see for instance Maris and Maris (2002), Fox and Glas (2001), Béguin and Glas (2001), and many others.
Data augmentation provides a very powerful tool to simplify sampling from distributions that are otherwise intractable. However, the tractability comes at a prize in terms of both the autocorrelation and the computational cost of every step in the resulting Markov chain, which limits its usefulness for large-scale applications.
The approach advocated by Albert (1992) involves two layers of augmented data. First, for every person an unobserved ability is introduced, and second, for every item response a normally distributed variable is introduced. Johnson and Junker (2003) propose to use a Metropolis-within-Gibbs algorithm to remove one layer of augmented data from the problem.
In this paper, a different approach will be developed that does not use data augmentation at all, and hence will give a Markov chain with lower autocorrelation, whilst at the same time producing tractable full conditional distributions. Moreover, as will become apparent later on, the computational cost for every iteration of the algorithm is independent of the number of persons.
where x i equals one for correct and zero for incorrect responses, δ i is the difficulty of item i, and θ denotes ability.
Recognizing that is proportional to the posterior distribution of ability for someone who answers all items incorrectly, with as proportionality constant the (marginal) probability to answer all items incorrectly (P(0)), we may express the marginal Rasch model as follows: where b i = exp(−δ i ) and x + denotes the sum score. The theoretical significance of Eq. 2, which corresponds to Equation 13 from Cressie and Holland (1983), is that it clearly shows that one cannot infer the full population distribution from the marginal Rasch model. However, theoretically, important this result is, we will treat Eq. 2 as a characterization of the marginal Rasch model that is useful for constructing a Gibbs sampler for Bayesian inference.
With some further change of notation we finally obtain the following characterization of the marginal Rasch model: As it stands, the marginal Rasch model, as written in Eq. 3, need, without further constraints, not even represent a probability distribution. A constraint which suffices to ensure that Eq. 3 represents a probability distribution (i.e. x P(x) = 1) is the following: in which the γ s function denotes the elementary symmetric function 2 of order s of the vector b. Imposing the constraint in Eq. 4 we obtain the following expression for the marginal Rasch model from which we readily see that it does indeed represent a probability distribution for all (nonnegative) values of its parameters. Additional insight in the structure of the marginal Rasch model derives from considering some of its properties. We focus on properties that are not only theoretically but also practically significant. First, from the distribution in Eq. 5 we readily find the following factorization The elementary symmetric function of order s of the vector b is defined as where the sum runs over all response patterns x that yield the sum score s.

862
PSYCHOMETRIKA which gives the conditional likelihood distribution that is also used in conditional maximumlikelihood estimation for the Rasch model (Andersen, 1973) and the score distribution. Observe that the factorization shows that the observed score distribution is the sufficient statistic for λ.
Observe that the parameters b, λ and b, π are one-one transformations of each other. The last expression is due to Tjur (1982), and is called the extended Rasch model by Cressie and Holland (1983). We see directly from Eq. 6 that the conditional maximum-likelihood estimates of the item difficulty parameters (Andersen, 1973) are equivalent to their maximum-likelihood estimates under an extended Rasch model. As a consequence, Bayesian inferences for the parameters of the extended Rasch model can be perceived as the Bayesian analogue of conditional maximumlikelihood estimation. Second, we consider the marginal and conditional distributions corresponding to Eq. 5. In particular, we consider the distribution of x without item n (which we denote by x (n) ): where the last equality follows from the following recursive property of elementary symmetric functions (Verhelst, Glas, & van der Sluis, 1984): and shows that X (n) is also a marginal Rasch model. We readily obtain the distribution of X n conditionally on the remaining n − 1 responses: We find that this conditional distribution only depends on the remaining n − 1 responses via the raw score x (n) + , and it is independent of the values of the remaining item parameters b (n) . That is, expression 9 gives an analytical expression for the item-rest regression function, which may be used for evaluating the fit of the marginal Rasch model. Third, in rewriting Eq. 2 as Eq. 3, we actually did more than just change the parametrization. Specifically, the model in Eq. 3 reduces to the model in Eq. 2 if, and only if, the λ s parameters represent a sequence of moments. To appreciate the kind of constraints this implies, we consider λ 1 and λ 2 . From the fact that the variance of a random variable is non-negative, we readily obtain that In its most general form, these inequality constraints can be formulated as follows (Shohat & Tamarkin, 1943): After introducing a Gibbs sampler for the extended Rasch model in Eq. 3, in the next Section, we consider how the additional constraints implied by the marginal Rasch model in Eq. 2 can be incorporated in the algorithm. In a more restricted setting, Theorem 3 of Hessen (2011) provides the constraints needed for the extended Rasch model to be equivalent to a marginal Rasch model in which the latent variable is normally distributed. Fourth, even if all the moment constraints are met, the λ s parameters are not very easy to interpret, as they correspond to a sequence of moments corresponding to the posterior distribution of ability for a person who fails all the items. For that reason we introduce a more natural parametrization. Specifically, from the Dutch identity (Holland, 1990) applied to the marginal Rasch model, we immediately obtain which is recognized as the posterior expectation of ability for different scores. Observe that the posterior expectation of ability for a person who answers all questions correctly cannot be estimated. As we find later, this new parametrization is also useful when considering the moment constraints implied by the marginal Rasch model. In terms of the item parameters b and the EAP parameters τ , the marginal Rasch model can be expressed as follows: Fifth, a further consequence of the Dutch identity is that we can obtain not only the EAP estimators for ability, but also more generally 864 PSYCHOMETRIKA We proceed to show how this fact can be used in combination with an algorithm to sample from the posterior distribution of the parameters of the marginal Rasch model to obtain estimates of both the posterior mean and variance of ability, taking into account the uncertainty regarding the parameters of the marginal Rasch model. Using Eq. 11, we obtain that from which we directly obtain (for s = 0, . . . , n − 1) which can be directly estimated (using Monte Carlo integration) with a sample from the posterior distribution of . Similarly, we can estimate the posterior variance of ability (for s = 0, . . . , n−2) where V(exp( )|X + = s, b, λ) is estimated as follows: The first term on the right-hand side of Eq. 12 reflects uncertainty due to the fact that the parameters b and λ are not known, whereas the second term reflects uncertainty due to finite test length. Specifically, as the number of persons tends to infinity, the first term in Eq. 12 tends to zero. The second term, however, tends to zero as the number of items tends to infinity. For some, inferences regarding exp(θ ) rather than regarding θ directly may seem inconvenient. Particularly, since the posterior distribution of exp(θ ) converges to its asymptotic (in the number of items) normal limit at a slower rate than does the posterior distribution of θ . Hence, the posterior mean and variance of exp(θ ) need not give a good summary of the posterior distribution. Using Corollary 1 from Holland (1990), we may for scores for which the posterior distribution of θ can be considered to be approximately normal, use the relation between moments of the log-normal distribution, and the mean and variance of the corresponding normal distribution to obtain approximations to the posterior mean and variances of θ (denoted below with μ s and σ 2 s ): and Finally, in the field of educational surveys (such as PISA, TIMMS, ESLC, etc.), the purpose of the study is to relate ability to student (or school, or system) characteristics. We shortly consider how such research could, in principle, be based on the marginal Rasch model. In typical applications, the relation between student responses and other student characteristics (e.g. gender) runs through ability. That is Y (the student characteristics) and the student responses X are independent conditionally on ability. Typically, the distribution of ability conditionally on Y is modelled as a normal regression model. Theorem 1. If Y⊥ ⊥ X|θ and X⊥ ⊥ |X + , then also Y⊥ ⊥ X|X + Proof. The conditions of the Theorem imply the following joint distribution: Theorem 1 shows that under the assumptions of independence between X and Y conditionally on θ , and of sufficiency of the sum score, all information on the relation between Y and X is contained in the distribution of Y conditionally on X + , which is (at least in principle) directly observable (to any desired degree of accuracy). Observe that Theorem 1 holds true for every element of Y in isolation, which implies that we may model main effects of student characteristics with an appropriate item-rest regression function (with the item relating to an element of Y, and the rest to X + ). Observe furthermore that, using Bayes theorem, we may equally well estimate the distribution of X + conditionally on an element from Y.

A Gibbs Sampler
Looking at the likelihood function in Eq. 5, we readily see that the parameters are not identifiable from X. Specifically, using the following well-known relation for elementary symmetric functions γ s (cb) = c s γ s (b) (Verhelst et al., 1984), we obtain with λ * s = λ s /c s . This type of non-identifiability can easily be resolved with a constraint on one of the item parameters b i = 1 (which we assume to be the first one, without loss of generality). Observe, however, that changing the identifying constraint also changes the values of λ. Observe, that the λ s may all be multiplied with the same constant, without changing the distribution. This additional type of non-identifiability can easily be resolved by constraining one of the λ s parameters to a constant.
In order to construct an algorithm for sampling from the posterior distribution of b and λ corresponding to Eq. 5, a prior distribution needs to specified. We consider a simple prior distribution for the parameters which give rise to tractable full conditional distributions for each of the parameters. The prior we consider is the following: Assuming that none of the items is answered (in)correctly by all students, and that every score occurs at least once, we can specify an improper uniform prior distribution of b and λ by choosing all α i and β s to be equal to one: that still yields a proper posterior distribution. Using this prior, the posterior distribution is the following: where x +i refers to the number of persons that make item i correct, m s refers to the number of persons that obtain a sum score equal to s, and m denotes the number of persons.
The distribution in Eq. 14 is not very tractable. Specifically, it is not immediately clear how to generate iid draws from it. We show that using a Gibbs sampler (Geman & Geman, 1984;Gelfand & Smith, 1990;Casella & George, 1992) we obtain full conditional distributions that are each easy to sample from. In that way, we can generate a Markov chain for which the posterior distribution in Eq. 14 is the unique invariant distribution.

Full Conditional Distribution for b i
The full conditional distribution for an item parameter b i is proportional to In order to see how a sample from the full conditional distribution in Eq. 15 may be generated, we use the recursive property of elementary symmetric functions in Eq. 8 which shows that elementary symmetric functions are linear in each of their arguments.
Using the result in Eq. 8 allows us to rewrite the full conditional distribution in Eq. 15 as follows: where c is a constant depending only on all other parameters: With a transformation of variables we obtain the following expression which is readily seen to be a beta distribution.
That is, if we generate y from a beta (x +i + α i , m − x +i − α i ) distribution, then the following transformation of y (being the inverse to the transformation in Eq. 17) gives us a draw from the full conditional distribution in Eq. 15. Formally, the distribution in Eq. 15 classifies as a scaled Beta prime distribution.

Full Conditional Distribution for λ s
The full conditional distribution for an element of λ is readily seen to be the following: As we found when considering the full conditional distribution for the item parameters, we see that the denominator in Eq. 19 is linear in λ t , such that we obtain where now the constant (with respect to λ t ) c equals We see that the full conditional distributions for both the b i and the λ s parameters belong to the same family of distributions.

Simulation Results
In this section we present some simulation results to illustrate the operating characteristics of our new Gibbs sampler. We focus on two aspects. First, we evaluate the autocorrelation in the Markov chain, which drives convergence. Second, we evaluate the computational burden. In Appendix an illustrative implementation of our Gibbs sampler is given in R (R Development Core Team, 2011). This code was used to generate the simulation results presented below. Observe that when n becomes large, significant computational advantages can be obtained by coding (parts of) the algorithm in a compiled language (e.g. C++, Fortran, Pascal). All simulations were run on a Lenovo X200s laptop with an Intel Core2 Duo CPU with a clock speed of 2.13 GHz and 2 gigabytes of memory running Windows 7 Enterprise.

Autocorrelation and Convergence
Convergence of Markov chains is driven by the autocorrelation structure of the chain. In this simulation study we evaluate the autocorrelation as a function of lag, and convergence of the Gibbs sampler. A Markov chain is converged in iteration t if the cumulative distribution function (CDF) at iteration t and t + 1 coincide. For a 30 item test, with true item difficulties uniformly distributed between −2 and 2, and 100,000 persons drawn from a standard normal distribution, 5000 replications of the Gibbs sampler were run for 50 iterations each, with starting values uniformly distributed between 0 and 1 for b, and λ. These 5000 Markov chains allow us to estimate the autocorrelation between any two iterations, and to evaluate the distribution of every parameter at every iteration. Figure 1 shows the empirical CDF (ECDF) after 49 and 50 iterations for one item parameter (b 2 ) and one of the λ parameters (λ 10 ). It is clear from Figure 1 that after only 50 iterations the Markov chain is converged. Figure 2 gives the autocorrelation for lag 0 to 50, after discarding the first 49 iterations. It is clear from Figure 2 that except for the lag 1 autocorrelation, autocorrelation is negligible.
We conclude that our Markov chain comes close to generating an independent and identically distributed sample from the posterior distribution, with virtually no autocorrelation whatsoever.

Computational Complexity
An algorithm for which the computational cost does not depend on the number of persons has in principle great advantages over algorithms for which the computational cost increases with the number of persons. For instance, we can guarantee that for some sample size m * our algorithm will outperform any particular competitor for which the computational cost increases with sample size. However, it is only practically relevant if m * is some modest number. Clearly, if m * equals 10 9 there is little need for our algorithm. Moreover, the question remains whether our algorithm is feasible for realistic sample sizes. For instance, if for 30 items and 10 5 persons, one iteration takes a week, our algorithm may be more feasible than competitors, but still not feasible.
To evaluate the feasibility of the algorithm, the average time for one iteration for tests with a different number of items, and 100,000 persons, is given in Figure 3 (left panel).
The average time per iteration appears to increase as a quadratic function of the number of items. The largest cost per iteration is in the repeated evaluation of elementary symmetric functions, the computational complexity of which is quadratic in the number of items. To illustrate the computational gain from coding the algorithm in a compiled language, we compare the naive R implementation that is in Appendix with a C implementation of both the full conditional distribution for b and λ that is called from within R using a dynamic link library. The right panel of Figure 3 gives results on the computational time per iteration for tests consisting of different numbers of items. We see that even for a test consisting of 200 items, we can do roughly 150 iterations per second, regardless of the number of students. Comparing the right with the left-hand panel in Figure 3 shows the dramatic improvement that results from implementing key parts of the algorithm in C (or Fortran, etc.).
Finally, for comparison, the DA-Gibbs sampler of Albert (1992), or the Metropolis-within-Gibbs sampler of Johnson and Junker (2003) have a computational cost that increases as a linear function of both the number of items and persons. For the DA-Gibbs sampler we illustrate the average time for one iteration, for a C implementation, with different numbers of items and 100,000 persons, in Figure 4. We see in Figure 4 that the average time per iteration increases as a linear function of the number of items, and is considerably larger than the average times for our new algorithm when implemented in C.

Conclusion
The combination of low autocorrelation that implies a low number of burn in iterations to reach convergence of the Markov chain, and a small number of iterations after convergence on which inferences will be based, together with a cost per iteration that only depends on the number of items (such that for a test of 200 items we can do 9000 iterations a minute), make our Gibbs sampler extremely feasible, even for very large-scale applications.

Extensions
The approach taken to estimate the parameters of the marginal Rasch model can easily be generalized in various directions. To illustrate its flexibility, we consider dealing with incomplete designs, dealing with polytomous responses, dealing with multidimensional Rasch models and incorporating moment constraints. As will become clear, all of these generalizations can be combined with each other without losing the desirable characteristics of the simple algorithm presented above.

Incomplete Designs
The first problem we tackle is to show how the marginal Rasch model works out for data collected with a non-equivalent groups anchor test (NEAT) design. We consider the simplest NEAT design explicitly, but all results carry over immediately to more complicated designs.
Consider two groups of students, from possibly different populations, taking a test that consists of an anchor (we use x to denote responses on the anchor, and b for its item parameters), and a unique set of items (we use y and z for responses on the unique sets, and c and d for their parameters). Applying our representation for the marginal Rasch model we obtain the following two distributions: It is immediately clear that for the parameters c, d, λ and η, we obtain the same full conditional distributions as before. For the anchor items, the full conditional becomes the following: c, d, λ, η, x, y, z; α, β) which can be rewritten to the following general form: c, d, λ, η, x, y, z; α, β) ∝ b which classifies as a rational distribution.
With a further transformation of variables used c, d, λ, η, x, y, z; It is readily found that the natural logarithm of this distribution is concave and has linear tails and a single mode: c, d, λ, η, x, y, z; α, β)) Since the distribution is log-concave, we may use the adaptive rejection sampler from Gilks and Wild (1992). As an alternative, we propose a Metropolis sampler with a proposal distribution that closely matches the full conditional distribution. As a proposal distribution we consider the following distribution: the logarithm of which has linear tails with the same slope, which is recognized to be of the same form as the full conditional distribution for b i found with a complete design (i.e. Eq. 16 with a transformation of variables). We propose to choose the parameter c in such a way that the derivative of the logarithm of the proposal distribution with respect to δ i matches the value found for the target full conditional distribution, at its current value in the Markov chain. This proposal distribution closely matches the target full conditional distribution, as is illustrated in Figure 5 (left panel), which ensures that the resulting Metropolis-within-Gibbs algorithm will converge rapidly to its invariant distribution. For comparison, the right panel in Figure 5 gives the outer and inner hull for an adaptive rejection sampler based on three support points. Based on this comparison, we expect our Metropolis algorithm to outperform the adaptive rejection sampler, although either algorithm will work.

Multidimensional Model
A second generalization we want to consider is a situation where we have two tests measuring different constructs administered to a group of students. That is, we consider the following marginal likelihood: Using the same approach as taken for the marginal Rasch model we obtain the following representation: which may be reparameterized to We readily find that all full conditionals will be of the same general form as those for the marginal Rasch model.

Polytomous Responses
As a generalization of the Rasch model for polytomous items we consider a special case of the Nominal Response Model (Bock, 1972) namely one with a fixed scoring rule. The Gibbs sampler for this model will be developed along the same lines as that for the Rasch model.
Consider an item i with J i + 1 response alternatives j = 0, . . . , J i ; one of which is chosen. Let X pi denote the response alternative and for practical reasons we also consider the dummy coded variables Y i j = 1 if category j was chosen and Y i j = 0 otherwise. The category response function of the NRM is given by where a i0 = δ i0 = 0 for identification. We assume that the parameters a i j are known integer constant and the NRM specializes to an exponential family model in which y ++ = i j a i j y i j = i a i,x pi is a sufficient statistic for θ p . Among others the One Parameter Logistic Model (OPLM: Verhelst & Glas, 1995) and the partial credit model (e.g. Masters, 1982) are special cases that satisfy these additional constraints.
A derivation of the Gibbs sampler for this model proceeds along the same lines as before. First, with i, j as a shorthand notation for the product i J i j=1 where b i j = exp(−δ i j ) and X = 0 denotes a response pattern where zero credit was earned on each of the items. Thus, we obtain the following parametrization of the marginal model: where 874 PSYCHOMETRIKA are the elementary functions which satisfy the recursion Note that all formulae specialize to those for the dichotomous Rasch model when J i = 1 for all i, and a i j = 1 for j = 1. Using the recursive property of the elementary symmetric functions, it follows that the denominator in the expression for P(x) is linear in individual parameters which means that the Gibbs sampler for the polytomous model will be similar to the one for the Rasch model. The difference is in the normalizing constants for the full conditional distributions.

Parameter Constraints
As observed above, the extended Rasch model reduces to the marginal Rasch model if, and only if, certain constraints on the λ s parameters are met. Here we consider how parameter constraints can be incorporated in the Gibbs sampler. We focus on two different types of constraints. On the one hand we consider imposing some of the moment constraints on the λ s parameters.
On the other hand we show how to constrain the λ s parameters such that the model reproduces moments of the score distribution, rather than the complete score distribution.
Before, we found that λ 2 − λ 2 1 ≥ 0, is one (and probably the simplest) of the moment constraints. However, all constraints are formulated as a function of a set of λ s parameters that needs to be non-negative. Hence, the marginal Rasch model corresponds to an extended Rasch model with particular inequality constraints on the λ s parameters.
In contrast to maximum-likelihood-based inference, Bayesian MCMC algorithms are particularly well suited for incorporating inequality constraints between parameters for the purpose of parameter estimation. Before illustrating this, we first recast the moment constraints in a different form, which is important for educational measurement purposes.
Using Eq. 11, we obtain from the non-negativity of the (posterior) variance (for every score) that which we can equivalently express as This expression is important, as it implies that the τ parameters are a monotone function of the score, which is the minimal constraint on the extended Rasch model needed for educational measurement purposes. We now consider how the Gibbs sampler can be adapted, to incorporate the inequality constraints in Eq. 28. In a Bayesian framework, inequality constraints are introduced through the prior distribution. Specifically, we obtain the following prior distribution for the λ parameters: With this prior distribution, the full conditional distribution for, say, λ 2 becomes We find that all that is needed is an algorithm for sampling from a double-truncated scaled Beta prime distribution, which is a fully tractable problem. The extended Rasch model is an exponential family model with as sufficient statistics the observed number of students answering each item correct, and the observed score distribution. If we impose a log-polynomial constraint on the λ s parameters: α j s j we effectively replace the entire score distribution as sufficient statistics with the first J non-central moments of the score distribution. This effectively smooths the observed score distribution.

Discussion
The algorithm proposed in this paper provides a flexible, robust, and highly efficient approach to Bayesian inference for the marginal Rasch model.
As opposed to maximum-likelihood estimation, our Bayesian approach (a) allows for accounting for all sources of uncertainty in the model parameters (especially in the posterior expectation of ability), (b) does not need computation and inversion of the information matrix (both of which are computationally expensive) and (c) allows for imposing moment constraints. This last point allows for considering models that are more restrictive than the extended Rasch model, yet less restrictive than the typical marginal Rasch model (i.e. assuming a normal distribution for ability).
The various generalizations we considered (incomplete data, polytomous responses, multidimensional marginal Rasch models, moment constraints) demonstrate the flexibility of our approach. The efficiency of our approach derives from the fact that no form of data augmentation is used. This not only is highly beneficial in terms of the resulting autocorrelation of the Markov chain, but also in terms of the computational cost. To be explicit, the computational cost is independent of the number of respondents, which makes our approach ideally suitable for largescale educational measurement applications involving hundreds of thousands of respondents. The efficiency derives from our starting point, the closed form representation of the marginal Rasch model from Cressie and Holland (1983), that removes the need for any form of data augmentation. Because no assumptions need to be made regarding the distribution of ability, our approach is robust compared to other approaches that do rely on such assumptions. To wit, without assumptions there can also be no wrong assumptions, and hence no bias that may result from them. Because we in fact set up a Markov chain for the extended marginal Rasch model, we do not even have to assume that a distribution exists. The extended marginal Rasch model is a proper statistical model in its own right.
The alternative parametrization of the extended Rasch model in terms of the posterior expectations corresponding to the different scores (τ s ) shows that the least assumption we would want to add to the model, in most educational measurement contexts, is that the sequence τ s is nondecreasing in s. This assumption ensures that all the item-rest regression functions are nondecreasing, which is what we would expect from a test intended to measure a single construct. This additional assumption is easily imposed and/or tested in a Bayesian framework.
This last remark being true, it is still worthwhile not only from a theoretical, but also from a practical, point of view to keep the distinction between the proper marginal Rasch model and the extended marginal Rasch model in mind. Much of the power of latent trait models such as the marginal Rasch model derives from the fact that a complex multivariate distribution may be reduced to a single (latent) variable, the relation of which with all sorts of other variables (both as explained and as explanatory) is an important field of research. Keeping the distinction between the proper and extended marginal Rasch model in mind, we can have two distinct meanings.
First, we may impose on the algorithm for the extended marginal Rasch model, the proper constraints to ensure that the parameters correspond to the marginal Rasch model. The simplest approach involves imposing the inequality constraints from the reduced moment problem via the prior distribution, as we illustrated. This approach is easily implemented and only requires an efficient algorithm for sampling from a truncated beta distribution.
Second, we may want to test the fit of the proper marginal Rasch model against the extended marginal Rasch model. That is, we want to test the hypothesis λ ∈ (where indicates the subset of the parameter space consistent with the reduced moment problem). This takes the form of testing a set of inequality constraints. In principle, this can be accomplished using Bayes factors or via evaluating the posterior probability of . As this topic deserves attention in its own right, and its details extend well beyond the scope of this paper, we leave this as a topic for future research.
We perceive the use of our approach as being part of a plug-and-play divide-and-conquer approach to statistical inference for the Rasch model. The algorithm developed in this paper allows us to evaluate the fit of the marginal Rasch model, and allows for sound statistical inference on the item parameters, without the need for modelling the distribution of a latent trait. In a second step, after having concluded that the marginal Rasch model fits the data, we can start modelling the latent trait distribution. This topic will not be developed further in this paper and is also left for future research. Considering the representation of the marginal Rasch model in Eq. 6, this entails setting up a parametric model for the score distribution (π). Such a model is useful for the purpose of relating the latent trait to explanatory variables (e.g. for latent regression). Combining draws from the posterior distribution of the item parameters (integrating out the λ parameters), with draws from the posterior distribution of population specific parameters (in a parametric family of population distributions), conditionally on the item parameters, allows for the construction of simple and robust plug-and-play algorithms for survey research.