Personality assessments that rely on respondent self-report have been widely used for personnel selection. Such assessments typically adopt single-statement formats, such as Likert-type items, where respondents are presented with one statement at a time and are required to choose one among several alternatives (e.g., agree/disagree). However, especially for high-stakes testing, this format is vulnerable to faking and other types of response biases, such as central tendency, acquiescence, socially desirable responding, halo effects, leniency, and impression management (Brown & Maydeu-Olivares, 2011; Cheung & Chan, 2002; Morrison & Bies, 1991). To address these concerns, one alternative is multidimensional forced-choice (MFC) item formats (Brown & Maydeu-Olivares, 2011). Instead of evaluating each statement separately, respondents are presented with blocks consisting of two or more similarly attractive statements, in which each statement is assumed to measure only one personality trait. Respondents are required to make comparative judgments, choosing between statements according to the extent to which the statements describe their preferences or behavior (Brown & Maydeu-Olivares, 2013). While comparative judgments may reduce response biases, the MFC item formats have also met controversy (Brown & Maydeu-Olivares, 2013; Walton et al., 2020). One commonly cited problem is that the traditional scoring approaches of MFC items produce ipsative data, that is, the total score of a test is constant for all respondents. Ipsative scoring distorts individual profiles (i.e., it is impossible to achieve all high or all low scores), and creates challenges in estimating construct validity, criterion-related validity, and reliability (Brown & Maydeu-Olivares, 2013; Dueber et al., 2019). To address such issues, a series of MFC item response theory (IRT) models have been proposed (e.g., Andrich, 1995; Brown & Maydeu-Olivares, 2011; Morillo et al., 2016; Stark et al., 2005; Wang et al., 2017; Zinnes & Griggs, 1974) to model comparative responses generated via forced-choice items. For example, Stark et al. (2005) developed the multi-unidimensional pairwise-preference (MUPP) model for blocks only containing two statements, and Brown and Maydeu-Olivares (2011) developed the Thurstonian IRT (TIRT) model, which can model blocks with more than two statements.

Recently, the integration of MFC item formats and computerized adaptive testing (MFC-CAT) has gained increasing attention as studies demonstrate great advantages, such as reducing testing time, obtaining more information with a shorter test, and improving measurement accuracy (e.g., Joo et al., 2020; Stark et al., 2012). A few studies explored adaptive testing of personality using forced-choice IRT models, but most of them have focused exclusively on ideal-point models. For example, Borman et al. (2001) compared a unidimensional forced-choice CAT with other CAT rating scales in terms of reliability, validity, and accuracy of performance ratings. Stark et al. (2012) implemented simulation studies based on the MUPP model (Stark et al., 2005), where they examined the effects of dimensionality, test length, inter-trait correlations, and other test design specifications on latent trait estimation accuracy in nonadaptive and adaptive situations. Since then, most studies and applications for MFC-CAT have used pairwise preference forced-choice items, and these studies have shown more efficient trait estimation than nonadaptive tests of an equal length (e.g., Aon Hewitt, 2015; Drasgow et al., 2012; Stark et al., 2012, 2014). To explore the benefits of MFC-CAT with more than two statements in a block, Joo et al. (2020) compared the accuracy of latent trait estimation with MFC pair, triplet, and tetrad tests using adaptive item selection based on the GGUM-RANK (generalized graded unfolding-RANK) model (Hontangas et al., 2015; Joo et al., 2018).

While the above studies all used ideal-point models, another group of IRT models developed for MFC items is dominance models (Wang et al., 2017), such as Maydeu-Olivares and Brown’s (2010) TIRT model, Wang et al.’s (2017) Rasch ipsative model (RIM), and a polytomous extension of RIM (Qiu & Wang, 2016). We chose to focus on the TIRT model in this study. The TIRT model can be used to model a variety of forced-choice scales and has demonstrated efficacy in accommodating many combinations of traits and block sizes, which makes it widely applicable to many existing forced-choice questionnaires, such as the Survey of Interpersonal Values (Gordon, 1976), the Customer Contact Styles Questionnaire (SHL, 1997), and the Occupational Personality Questionnaire (SHL, 2006) (Brown & Maydeu-Olivares, 2011). Therefore, developing an adaptive testing approach based on the TIRT model presents a promising gateway towards further applications of MFC-CAT in personality tests that saves substantial cost of test administration.

In an adaptive test, the method used to select items from the item pool for each test-taker adaptively as the test progresses exerts a significant influence on measurement accuracy, test validity, and uniformity of item pool usage. Among the existing item selection methods for CAT, a group of methods developed for single-statement multidimensional CAT (MCAT; e.g., Chang & Ying, 1996; Mulder & van der Linden, 2009, 2010; Segall, 1996; Veldkamp & van der Linden, 2002) provides the foundation for this study, because MFC items measure multidimensional latent traits.

Among studies of item selection methods for single-statement MCAT, Mulder and van der Linden (2009) compared several methods based on the Fisher information (FI) and found that the estimation accuracy of the A-optimality method was slightly better than that of the D-optimality method, and the E-optimality method was the most unstable method. Although the FI-based item selection methods have achieved great popularity, several problems need to be addressed. For example, one assumption of the FI-based item selection methods is that the estimated trait levels are close to their true values, which is often violated at an early stage of CAT when few items have been administered, namely the attenuation paradox issue (e.g., Chang & Ying, 1996; Wang & Chang, 2011). When items with high FI are selected to match inaccurate trait estimates, the adaptive test loses efficiency and item exposure rates become uneven (Chang & Ying, 1996; Lin, 2012). As a global information index, the Kullback–Leibler (KL) information (Chang & Ying, 1996) has been proposed as an alternative to the FI to be used for CAT item selection. Veldkamp and van der Linden (2002) extended the KL information index (KL index, KI) method to multidimensional scenarios and proposed the posterior expectation KL information method (the KB method), and illustrated that the KL-based item selection methods performed better in estimation accuracy than the FI-based item selection methods.

Note that research on item selection methods for single-statement MCAT so far has mainly concentrated on single-statement items. Although several studies have explored multidimensional forced-choice IRT (MFC-IRT) under nonadaptive testing (e.g., Brown & Maydeu-Olivares, 2011; Hontangas et al., 2015; Joo et al., 2018; Stark et al., 2005; Wang et al., 2017), only two studies so far discussed adaptive item selection methods in MFC-CAT contexts (Joo et al., 2020; Stark et al., 2012). Stark et al. (2012) conducted four simulation studies to explore the effects of test length, dimensionality, inter-trait correlations, and the advantages of adaptive item selection on the accuracy and precision of latent trait estimates for pairwise preference testing. Joo et al. (2020) conducted simulations of MFC-CAT with pair, triplet, and tetrad formats using the FI-based item selection methods, specifically the A-optimality method for MFC items (MFC-A-optimality). In contrast to FI-based item selection item selection methods, methods based on the KL information have not been studied in the MFC-CAT contexts. Hence, this article focuses on the extension and application of item selection methods based on the KL information for MFC-CAT.

To achieve the above goals and provide a foundation for MFC-CAT research using KL-based item selection methods, this article is organized as follows: First, a brief summary of the TIRT model is presented. Second, we provide an introduction of the FI-based item selection methods that have been used in MFC-CAT contexts and present the proposed extension of the proposed KL-based item selection methods from single-statement MCAT to MFC-CAT. Third, we describe two Monte Carlo simulation studies to explore statistical properties and feasibility of these methods in MFC-CAT. We also discuss how test length, dimensionality, and inter-trait correlation affect the estimation accuracy and uniformity of item pool usage of MFC-CAT. Next, we present a simulation study based on real data using the item pool of the Big-Five factor marker questionnaire with forced-choice items to examine the empirical efficiency of the proposed item selection methods in a personality assessment application. We compare the latent trait estimation accuracy and uniformity of item pool usage of the new methods with the existing methods. Finally, we discuss limitations and recommendations.

TIRT

Thurstone (1927) proposed the law of comparative judgment to describe comparative choices made between statements in a forced-choice item block. This law assumes that each of the two statements (i.e., \({i}\) and m) in a block elicits a corresponding utility (i.e., ti and tm). A respondent prefers to choose the statement with the larger utility. Let \({\mathcal{Y}}_l\) denote the observed binary outcome and \({\mathcal{Y}}_l^{\ast }\) denote the unobserved difference of utilities for a pairwise comparison, l = {i, m}, within a forced-choice item block.

$${\mathcal{Y}}_l=\left\{\begin{array}{c}1,\kern0.75em if\ statement\ i\ is\ preferred\ to\ statement\ m,\\ {}0,\kern0.75em if\ statement\ m\ is\ preferred\ to\ statement\ i.\end{array}\right.$$
(1)
$${\mathcal{Y}}_l^{\ast }={t}_i-{t}_m.$$
(2)

Then, Thurstone’s (1927) law can be written as the relationship between the observed binary outcome \({\mathcal{Y}}_l\) and the unobserved difference of utilities \({\mathcal{Y}}_l^{\ast }\):

$${\mathcal{Y}}_l=\left\{\begin{array}{c}1,\kern0.75em if\ {\mathcal{Y}}_l^{\ast}\ge 0,\\ {}0,\kern0.75em if\ {\mathcal{Y}}_l^{\ast }<0.\end{array}\right.$$
(3)

Based on Thurstone’s (1927) law of comparative judgment, Brown & Maydeu-Olivares (2010, 2011) developed the TIRT model, which can be used to model a variety of forced-choice scales and has demonstrated efficacy in accommodating many combinations of traits and block sizes. When comparing statement \(\mathit{i}\) measuring latent trait ηa and statement m measuring the latent trait ηb, the item characteristic function (ICF) of the binary outcome \({\mathcal{Y}}_l\) can be described as

$$P\left({\mathcal{Y}}_l=1\left|{\eta}_a,{\eta}_b\right.\right)=\varPhi \left(\frac{-{\gamma}_l+{\uplambda}_i{\eta}_a-{\uplambda}_m{\eta}_b}{\sqrt{\psi_i^2+{\psi}_m^2}}\right),$$
(4)

where Φ(x) denotes the cumulative probability function of the standard normal distribution evaluated at x, γl is the threshold parameter for binary outcome γl, λi and λm are the statements’ factor loadings, and  \({\psi}_i^2\) and \({\psi}_m^2\) denote the statements’ uniqueness.

Now, let

$${\alpha}_l=\frac{-{\gamma}_l}{\sqrt{\psi_i^2+{\psi}_m^2}},{\beta}_i=\frac{\uplambda_i}{\sqrt{\psi_i^2+{\psi}_m^2}},{\beta}_m=\frac{\uplambda_m}{\sqrt{\psi_i^2+{\psi}_m^2}},$$
(5)

then the TIRT model (defined by Eq. 4) can be written in an intercept/slope form as

$$P\left({\mathcal{Y}}_l=1\left|{\eta}_a,{\eta}_b\right.\right)=\varPhi \left({\alpha}_l+{\beta}_i{\eta}_a-{\beta}_m{\eta}_b\right),$$
(6)

where αl is the intercept parameter for binary outcome γl, and βi and βm are the slope parameters for statement i and statement m, respectively.

To help readers better understand the TIRT model and facilitate computations, in this study, we replaced the cumulative probability function of the standard normal distribution in the TIRT model with a logistic function by referring to the processing method adopted by Morillo et al. (2016):

$$P\left({\mathcal{Y}}_l=1\left|{\eta}_a,{\eta}_b\right.\right)=\frac{1}{1+\exp \left[-\left({\beta}_i{\eta}_a-{\beta}_m{\eta}_b+{\alpha}_l\right)\right]}.$$
(7)

Note that if a forced-choice item block contains more than two statements, there exist more than one pairwise comparison (e.g., three pairwise comparisons for a block with three statements). For example, the comparisons between three statements A, B, and C for a block can be presented as follows.

Three-statement block

Pairwise comparisons

A

B

C

A vs. B

B vs. C

A vs. C

  1. Note. A vs. B = the statement A is compared with the statement B; B vs. C = the statement A is compared with the statement B; A vs. C = the statement A is compared with the statement B.

Extension of item selection methods from MCAT to MFC-CAT

In order to facilitate the presentation, several notations will be introduced here. d denotes the measured trait dimensions in tests, \(z\in\left\{1,\cdots,d\right\}\) denotes the component of latent trait vector η (η is a d-dimensional vector of latent traits), R represents the item pool, Sk − 1 denotes the set of the first k − 1 administered blocks, Uk − 1 denotes the response vector of the k − 1 administered blocks, jk denotes the block administered as the kth block in the test, and Rk denotes the set of blocks remaining in the item pool after the (k − 1)th block is administered.

Under the framework of single-statement MCAT, a group of item selection methods have been developed (e.g., Chang & Ying, 1996; Mulder & van der Linden, 2009, 2010; Segall, 1996; Veldkamp & van der Linden, 2002). At present, only the MFC-A-optimality method, which is based on the FI, has been applied to MFC-CAT (Joo et al., 2020). The FI-based item selection methods assume that the intermediate trait estimates are close to their true values, which is often violated at the beginning of CAT due to few items having been administered (Mulder & van der Linden, 2009; Segall, 1996). One alternative to the FI to be used for CAT item selection is the global KL information (Chang & Ying, 1996), which is a measure of discrepancy between two probability distributions. It does not require that the estimated latent trait, \(\hat{\boldsymbol{\eta}}\), be close to the true value, η, and it is more robust than FI against early-stage estimation instability (Lima Passos et al., 2007). Several studies have demonstrated that the performance of KL-based item selection methods is more stable, efficient, and precise in terms of trait estimation, especially at an early stage of CAT or for a short CAT (Chang & Ying, 1996; Veldkamp & van der Linden, 2002; Wang et al., 2011). Therefore, with this study, we propose an extension of KL-based item selection methods from the single-statement MCAT context to the MFC-CAT context. We then explore whether the properties of the KL-based item selection methods continue to hold true in the MFC-CAT context. In the following sections, we first describe the FI-based item selection methods for MFC-CAT and then introduce the proposed KL-based item selection methods for MFC-CAT.

FI-based item selection methods for MFC-CAT

Under the framework of MFC-CAT, the FI is given as a matrix. With the TIRT model employed, the FI matrix for Block j can be defined as

$${\displaystyle \begin{array}{c}{I}_j^{\ast}\left(\boldsymbol{\eta} \right)={P}_j\left(\boldsymbol{\eta} \right){Q}_j\left(\boldsymbol{\eta} \right)\left[\begin{array}{c}{\beta}_{1j}^2\kern0.5em {\beta}_{1j}{\beta}_{2j}\kern0.5em \begin{array}{cc}\cdots & {\beta}_{1j}{\beta}_{dj} \end{array}\\ {}\begin{array}{ccc}{\beta}_{2j}{\beta}_{1j}& {\beta}_{2j}^2& \begin{array}{cc}\cdots & {\beta}_{2j}{\beta}_{dj} \end{array}\end{array}\\ {}\begin{array}{c}\begin{array}{ccc}\vdots & \vdots & \begin{array}{cc}\ddots & \vdots \end{array}\end{array}\\ {}\begin{array}{ccc}{\beta}_{dj}{\beta}_{1j}& {\beta}_{dj}{\beta}_{2j}& \begin{array}{cc}\cdots & {\beta}_{dj}^2\end{array}\end{array}\end{array}\end{array}\right]\\ {}=\left[\begin{array}{c}\begin{array}{ccc}{\beta}_{1j}^2{P}_j\left({\eta}_a,{\eta}_b\right){Q}_j\left({\eta}_a,{\eta}_b\right)& {\beta}_{1j}{\beta}_{2j}{P}_j\left({\eta}_a,{\eta}_b\right){Q}_j\left({\eta}_a,{\eta}_b\right)& \begin{array}{cc}\cdots & {\beta}_{1j}{\beta}_{dj}{P}_j\left({\eta}_a,{\eta}_b\right){Q}_j\left({\eta}_a,{\eta}_b\right)\end{array}\end{array}\\ {}\begin{array}{ccc}{\beta}_{2j}{\beta}_{1j}{P}_j\left({\eta}_a,{\eta}_b\right){Q}_j\left({\eta}_a,{\eta}_b\right)& {\beta}_{2j}^2{P}_j\left({\eta}_a,{\eta}_b\right){Q}_j\left({\eta}_a,{\eta}_b\right)& \begin{array}{cc}\cdots & {\beta}_{2j}{\beta}_{dj}{P}_j\left({\eta}_a,{\eta}_b\right){Q}_j\left({\eta}_a,{\eta}_b\right)\end{array}\end{array}\\ {}\begin{array}{c}\begin{array}{ccc}\vdots & \vdots & \begin{array}{cc}\ddots & \vdots \end{array}\end{array}\\ {}\begin{array}{ccc}{\beta}_{dj}{\beta}_{1j}{P}_j\left({\eta}_a,{\eta}_b\right){Q}_j\left({\eta}_a,{\eta}_b\right)& {\beta}_{dj}{\beta}_{2j}{P}_j\left({\eta}_a,{\eta}_b\right){Q}_j\left({\eta}_a,{\eta}_b\right)& \begin{array}{cc}\cdots & {\beta}_{dj}^2{P}_j\left({\eta}_a,{\eta}_b\right){Q}_j\left({\eta}_a,{\eta}_b\right)\end{array}\end{array}\end{array}\end{array}\right]\end{array}}$$
(8)

where d denotes the number of dimensions measured by the test. \(P_j (\eta{_a}, \eta_{b})\) denotes the probability of preferring the first statement measuring trait ηa over the second statement measuring trait ηb in a pairwise comparison, which is the shorthand notation for \(P\left({\mathcal{Y}}_l=1\left|{\eta}_a,{\eta}_b\right.\right)\) in Eq. 7, and \({Q}_j(\eta_{a},\eta_{b})=1{-}{P}_{j}(\eta_{a},\eta_{b})\). Note that a single pair block only involves statements pertaining to two of the d dimensions, and hence the information matrix has only four nonzero elements, and all other elements equal 0. Likewise, a single triplet block only involves three of the d dimensions, and the information matrix has only nine nonzero elements. Also note that different blocks have different nonzero entries depending on the dimensions measured by the block respectively. However, these information matrices can be summed up across different blocks as in Eqs. 9 and 10 below because they share the same structure.

Under the conditional independence assumption of the responses given η, the information matrix of a test is equal to the sum of the block information matrices. Therefore, the FI matrix of the test can be expressed as

$${I}^{\ast}\left(\boldsymbol{\eta} \right)=\sum_{j=1}^J{I}_j^{\ast}\left(\boldsymbol{\eta} \right).$$
(9)

Then, the FI matrix of a set of Sk − 1 blocks could be computed by

$${I}_{S_{k-1}}^{\ast}\left(\boldsymbol{\eta} \right)=\sum_{j\in {S}_{k-1}}{I}_j^{\ast}\left(\boldsymbol{\eta} \right).$$
(10)

Based on the FI, three popular optimality methods, namely the D-optimality method, the A-optimality method, and the E-optimality method, have been developed for single-statement MCAT (Mulder & van der Linden, 2009). The MFC-A-optimality method has been used in previous MFC-CAT studies but without being expressed with formulation (Joo et al., 2020). Mulder and van der Linden (2009) found that E-optimality lacks robustness in applications with sparse data. Therefore, we present the formulas for MFC-A-optimality and MFC-D-optimality as the following.

The D-optimality method for MFC items (MFC-D-optimality)

The MFC-D-optimality method seeks to select the next item that maximizes the determinant of the information matrix, and this method can be expressed as

$${j}_k=\mathit{\arg}\underset{j\in {R}_k}{\min}\left\{\det \left[{I}_{S_{k-1}}^{\ast}\left({\hat{\boldsymbol{\eta}}}_{k-1}\right)+{I}_{j}^{\ast}\left({\hat{\boldsymbol{\eta}}}_{\boldsymbol{k}-\textbf{1}}\right)\right]\right\},$$
(11)

where \(\left[{I}_{S_{k-1}}^{\ast}\left({\hat{\boldsymbol{\eta}}}_{\boldsymbol{k}-\textbf{1}}\right)+{I}_j^{\ast}\left({\hat{\boldsymbol{\eta}}}_{\boldsymbol{k}-\textbf{1}}\right)\right]\) denotes the sum of the information matrix after the k − 1 blocks already administered and the information matrix for candidate block j.

The A-optimality method for MFC items (MFC-A-optimality)

This method seeks to select the next block that minimizes the sum of the (asymptotic) sampling variances of the trait estimators, which is equivalent to minimizing the trace of the inverse of the information matrix. Its formulation is

$${\displaystyle \begin{array}{c}{j}_k=\mathit{\arg}\ \underset{j\in {R}_k}{\min}\left\{ trace\left[{\left({I}_{S_{k-1}}^{\ast}\left({\hat{\boldsymbol{\eta}}}_{k-1}\right)+{I}_j^{\ast}\left({\hat{\boldsymbol{\eta}}}_{k-1}\right)\right)}^{-1}\right]\right\}\\ {}=\arg\ \underset{j\in {R}_k}{\max}\left\{\frac{\det \Big[{I}_{S_{k-1}}^{\ast}\left({\hat{\boldsymbol{\eta}}}_{k-1}\right)+{I}_j^{\ast}\left({\hat{\boldsymbol{\eta}}}_{k-1}\right)}{\sum\limits_{\mathit{z}=1}^d\det \left(\right[{I}_{S_{k-1}}^{\ast}\left({\hat{\boldsymbol{\eta}}}_{k-1}\right)+{I}_j^{\ast }{\left({\hat{\boldsymbol{\eta}}}_{k-1}\right)}_{\left[\mathit{z},\mathit{z}\right]}}\right\},\end{array}}$$
(12)

where \({\hat{\boldsymbol{\eta}}}_{k-1}\) denotes the trait estimator after the first k − 1 blocks are administrated, and \({\left[{I}_{S_{k-1}}^{\ast}\left({\hat{\boldsymbol{\eta}}}_{k-1}\right)+{I}_j^{\ast}\left({\hat{\boldsymbol{\eta}}}_{k-1}\right)\right]}_{\left[\mathit{z},\mathit{z}\right]}\) is the submatrix after deleting the \(\mathit{z}\)th row and column of the information matrix \(\left[{I}_{S_{k-1}}^{\ast}\left({\hat{\boldsymbol{\eta}}}_{k-1}\right)+{I}_j^{\ast}\left({\hat{\boldsymbol{\eta}}}_{k-1}\right)\right]\).

The proposed extension of KL-based item selection methods for MFC-CAT

Several adaptive selection methods based on KL information have been developed for single-statement MCAT (Chang & Ying, 1996; Mulder & van der Linden, 2010; Veldkamp & van der Linden, 2002; Wang & Chang, 2011), such as the KL index (KI) method, posterior expected KL information method (KB), and the KL distance between subsequent posteriors (KLP) method. To adapt the above KL-based item selection methods to MFC-CAT, we propose to modify the classical KL information as

$$K{L}_j^{\ast}\left(\hat{\boldsymbol{\eta}}\parallel \boldsymbol{\eta} \right)=\sum_{c=1}^{C_j}{L}_{cj}\left(\hat{\boldsymbol{\eta}}\right)\log \left[\frac{L_{cj}\left(\hat{\boldsymbol{\eta}}\right)}{L_{cj}\left(\boldsymbol{\eta} \right)}\right],$$
(13)

where η and \(\hat{\boldsymbol{\eta}}\) denote the unknown and estimated latent trait vectors, respectively; j denotes the jth block, Cj is the number of possible scoring patterns for Block j (e.g., a block with three statements, such as A, B, and C, has six possible scoring patterns; see Table 1); c(c = 1, 2, ⋯, Cj) indicates the cth scoring pattern.

Table 1 All possible scoring patterns in a block with three statements

L cj(η) and \({L}_{cj}\left(\hat{\boldsymbol{\eta}}\right)\) refer to the block response probability, namely the likelihood of pairwise comparison response probability, for latent traits η and \(\hat{\boldsymbol{\eta}}\), respectively, given the cth scoring patterns of Block j. The expression of Lcj(η) and \({L}_{cj}\left(\hat{\boldsymbol{\eta}}\right)\) are respectively given by

$${L}_{cj}\left(\boldsymbol{\eta} \right)=\prod_{\alpha =1}^{K_{j-1}}\prod_{b=a+1}^{K_j}{P}_j{\left({\eta}_a,{\eta}_b\right)}^{{\mathcal{Y}}_l}{\left[1-{P}_j\left({\eta}_a,{\eta}_b\right)\right]}^{\left(1-{\mathcal{Y}}_l\right)},$$
(14)

and

$${L}_{cj}\left(\hat{\boldsymbol{\eta}}\right)=\prod_{a=1}^{K_{j-1}}\prod_{b=a+1}^{K_j}{P}_j{\left({\eta}_a,{\eta}_b\right)}^{{\mathcal{Y}}_l}{\left[1-{P}_j\left({\eta}_a,{\eta}_b\right)\right]}^{\left(1-{\mathcal{Y}}_l\right)}$$
(15)

where Kj denotes the number of statements in Block j, \({\mathcal{Y}}_l\) is defined in Eq. 1, and Pj(ηa, ηb) is defined in Eq. 7. Brown and Maydeu-Olivares (2011) proposed that effects of ignoring these dependencies on the latent trait estimates have been shown to be negligible in applications involving a single ranking task, and they are likely to be even smaller in forced-choice questionnaires where blocks are smaller and there are fewer local dependencies per item. So, throughout this article, we will use the simplifying assumption that the ICFs for the binary outcomes are locally independent.

The proposed extension of the KI method for MFC items (MFC-KI)

The KL information as shown in Eq. 13 is a function of the true trait η, but the true trait value is unknown. Therefore, Chang and Ying (1996) proposed to calculate the KL index (KI) by integrating the estimated trait \(\hat{\boldsymbol{\eta}}\). The extended KI item selection method for MFC items (MFC-KI) can be defined as

$${j}_k=\mathit{\arg}\ \underset{j\in {R}_k}{\max}\left\{{KI}_j\left({\hat{\boldsymbol{\eta}}}_{k-1}\right)\right\}=\mathit{\arg}\ \underset{j\in {R}_k}{\max}\left\{\int_{{\hat{\boldsymbol{\eta}}}_k-{\delta}_{k-1}}^{\hat{\boldsymbol{\eta}}+{\delta}_{k-1}}{KL}_j^{\ast}\left({\hat{\boldsymbol{\eta}}}_{k-1}\parallel \boldsymbol{\eta} \right)\partial \boldsymbol{\eta} \right\},$$
(16)

where \({\delta}_k=d\sqrt{k-1}\) determines the size of the area on which the average is calculated, d usually takes a value of 3 (Chang & Ying, 1996; Veldkamp & van der Linden, 2002), and k − 1 denotes the number of blocks that have been administered.

The MFC-KI method selects the blocks with the largest KI value among the remaining blocks Rk in the item pool.

The proposed extension of the KB method for MFC items (MFC-KB)

By weighting KL through the posterior distribution of latent trait η, Veldkamp and van der Linden (2002) proposed a Bayesian version of the KI method, that is, the multidimensional posterior expected KL information method (KB). Under the framework of MFC-CAT, the expression of the KB method for MFC items (MFC-KB) can be written as

$${\displaystyle \begin{array}{c}{j}_k=\mathit{\arg}\underset{j\in {R}_k}{\max }{K}_j^B\left({\hat{\boldsymbol{\eta}}}_{k-1}\right)=\mathit{\arg}\underset{j\in {R}_k}{\max }{\int}_{\eta }{KL}_j^{\ast}\left({\hat{\boldsymbol{\eta}}}_{k-1}\parallel \boldsymbol{\eta} \right){\pi}_{k-1}\left(\boldsymbol{\eta} \parallel {\boldsymbol{U}}_{k-1}\right)\partial \boldsymbol{\eta} \\ {}=\mathit{\arg}\underset{j\in {R}_k}{\max }{\int}_{\eta}\left\{\sum\limits_{c=1}^{2^{K_j}}{L}_{cj}\left({u}_{jk}\left|{\hat{\boldsymbol{\eta}}}_{k-1}\right.\right)\log \left[\frac{L_{cj}\left({u}_{jk}\left|{\hat{\boldsymbol{\eta}}}_{k-1}\right.\right)}{L_{cj}\left({u}_{jk}\left|\boldsymbol{\eta} \right.\right)}\right]\right\}{\pi}_{k-1}\left(\boldsymbol{\eta} \left|{\boldsymbol{U}}_{k-1}\right.\right)\partial \boldsymbol{\eta}, \end{array}}$$
(17)

where Lcj(ujk|η) and \({L}_{cj}\left({u}_{jk}\left|{\hat{\boldsymbol{\eta}}}_{k-1}\right.\right)\) denote the response probability for η and \({\hat{\boldsymbol{\eta}}}_{k-1}\) when selecting Block j as the kth administrated block of the test with the response score ujk(ujk = 0, 1), respectively. πk − 1(η|Uk − 1) indicates the posterior distribution for η after k − 1 blocks have been administrated:

$${\pi}_{k-1}=\left(\boldsymbol{\eta} \left|{\boldsymbol{U}}_{k-1}\right.\right)=\frac{g\left(\boldsymbol{\eta} \right)L\left({\boldsymbol{U}}_{k-1}\left|\boldsymbol{\eta} \right.\right)}{\int g\left(\boldsymbol{\eta} \right)L\left({\boldsymbol{U}}_{k-1}\left|\boldsymbol{\eta} \right.\right)\partial \boldsymbol{\eta}},$$
(18)

where g(η) denotes a prior distribution for η, Uk − 1 denotes the response vector of the k − 1 administered blocks, and L(Uk − 1|η) denotes the likelihood associated with response vector Uk − 1.

The proposed extension of the KLP method for MFC items (MFC-KLP)

An item should be selected to maximize the divergence between the posterior distributions of η. One of the possible responses to the candidate item would move the posterior distribution of η toward the respondents’ true trait, and the other would move it away from the respondents’ true trait, and then this level of divergence between the response distributions generated by two different trait levels can be formalized by the KL information, i.e., the KLP (Mulder & van der Linden, 2010). The KLP method selects the item with the maximum expected KLP distributions πk − 1(η|Uk − 1) and πk(η|Uk − 1, ujk) (Mulder & van der Linden, 2010; Tu et al., 2018). Under the framework of MFC-CAT, the expression of the KLP method for MFC items (MFC-KLP) can be defined as

$${\displaystyle \begin{array}{c}{j}_k=\mathit{\arg}\underset{j\in {R}_k}{\max } KL{P}_j=\mathit{\arg}\underset{j\in {R}_k}{\max}\sum_{c=1}^{Cj}{\int}_{\eta }{L}_{cj}\left({u}_{jk}\left|{\boldsymbol{U}}_{k-1}\right.\right) KL\left({\pi}_{k-1}\left(\boldsymbol{\eta} \left|{\boldsymbol{U}}_{k-1}\right.\right)\parallel {\pi}_k\left(\boldsymbol{\eta} \left|{\boldsymbol{U}}_{k-1},{u}_{jk}\right.\right)\right)\partial \boldsymbol{\eta} \\ {}=\mathit{\arg}\underset{j\in {R}_k}{\max}\sum_{c=1}^{C_j}{\int}_{\eta }{L}_{cj}\left({u}_{jk}\left|{\boldsymbol{U}}_{k-1}\right.\right){\pi}_{k-1}\left(\boldsymbol{\eta} \left|{\boldsymbol{U}}_{k-1}\right.\right)\log \left[\frac{\pi_{k-1}\left(\boldsymbol{\eta} \left|{\boldsymbol{U}}_{k-1}\right.\right)}{\pi_k\left(\boldsymbol{\eta} \left|{\boldsymbol{U}}_{k-1},{u}_{jk}\right.\right)}\right],\end{array}}$$
(19)

where the predictive probability and posterior distribution of the kth candidate block after k − 1 blocks have been administrated can be defined as follows

$${L}_{cj}\left({u}_{jk}\left|{\boldsymbol{U}}_{k-1}\right.\right)={\int}_{\eta }{L}_{cj}\left({u}_{jk}\left|\boldsymbol{\eta} \right.\right){\pi}_{k-1}\left(\boldsymbol{\eta} \left|{\boldsymbol{U}}_{k-1}\right.\right)\partial \boldsymbol{\eta},$$
(20)
$${\pi}_k\left(\boldsymbol{\eta} \left|{\boldsymbol{U}}_{k-\textbf{1}},{u}_{jk}\right.\right)=\frac{g\left(\boldsymbol{\eta} \right)L\left({\boldsymbol{U}}_{k-1},{u}_{jk}\left|\boldsymbol{\eta} \right.\right)}{\int g\left(\boldsymbol{\eta} \right)L\left({\boldsymbol{U}}_{k-1},{u}_{jk}\left|\boldsymbol{\eta} \right.\right)\partial \boldsymbol{\eta}},$$
(21)

where L(Uk − 1, ujk|η) = L(Uk − 1|η)Lcj(ujk|η) denotes the likelihood of the kth candidate block after k − 1 blocks have been administrated.

The R codes of the proposed MFC-KI, MFC-KB, and MFC-KLP methods can be found at https://osf.io/bmg8r/.

Simulation studies

Two Monte Carlo simulation studies and a simulation based on real data were conducted to evaluate the proposed KL-based item selection methods for MFC-CAT. Study 1 and study 2 compared the performance of the newly developed KL-based item selection methods against the existing FI-based item selection methods in terms of trait estimation accuracy and uniformity of item pool usage in three-dimensional and five-dimensional MFC-CAT scenarios, respectively. Finally, the simulation based on real data (the Big-Five factor marker questionnaire response data) further investigated the feasibility of the proposed KL-based item selection methods in real MFC-CAT testing situations.

Simulation study 1

Simulation design

In this study, we were focused on the triplet test, where three latent trait dimensions (d=3) were measured in a block consisting of three statements, because it is more common in block matching. An item pool containing 100 triplet blocks were pre-assembled following methods used by Joo et al. (2020). Specifically, Joo et al. (2020) found that the percentage of unidimensional blocks had little influence on GGUM-RANK scoring. Therefore, we only considered the case that each statement in each block measures different traits from different dimensions. Item responses were simulated based on the TIRT model. The slope parameters β and the intercept parameters α were randomly sampled from a lognormal distribution and a normal distribution respectively. To compare item selection methods under a variety of test scenarios, we varied the correlations between traits (inter-trait correlations) at 0 and 0.5, and varied the test length at 5, 10, and 15 blocks. To simulate data for this study, 500 true latent trait vectors were randomly generated from a multivariate standard normal distribution with the abovementioned inter-trait correlations.

In sum, there were 5 (item selection method: MFC-A-optimality, MFC-D-optimality, MFC-KI, MFC-KB, and MFC-KLP) × 2 (inter-trait correlation: 0, 0.5) × 3 (test length: 5, 10, 15) = 30 simulation conditions. For each condition, 20 replications were performed. This study used the expected a posteriori (EAP; Bock & Mislevy, 1982) estimation for latent trait estimation, in which the trait prior distribution was set to a multivariate standard normal distribution. Gauss-Hermite numerical integration (Glas, 1992) was used for the parameter estimation and the integration was taken over the range of trait [−3, +3]. All simulation code was written in R.

Evaluation criteria

The performance of each method was evaluated by trait estimation accuracy and uniformity of item pool usage. In this study, the indices to evaluate trait estimation accuracy were bias (BIAS), root mean squared error (RMSE) and the correlation between the generating and estimated traits (CORR), while the index to evaluate uniformity of item pool usage included chi-square (x2).

The three trait estimation accuracy indices were computed as follows:

$${BIAS}_d=\frac{1}{N}\sum_{n=1}^N\left({\hat{\eta}}_{nd}-{\eta}_{nd}\right),$$
(22)
$${RMSE}_d=\sqrt{\frac{1}{N}\sum_{n-1}^N{\left({\hat{\eta}}_{nd}-{\eta}_{nd}\right)}^2,}$$
(23)
$${CORR}_d=\frac{\sum\limits_{n=1}^N\left({\eta}_{nd}-{\overline{\eta}}_d\right)\left({\hat{\eta}}_{nd}-{\overline{\hat{\eta}}}_d\right)}{S_{\eta_d}{S}_{{\hat{\eta}}_d}},$$
(24)

where N is the total number of respondents in the test, n denotes the nth respondent, and ηnd and \({\hat{\eta}}_{nd}\) are the true traits and the estimated traits of respondent n respectively. \(\overline{\eta}_d\) and \({S}_{\eta_d}\) are the mean value and standard deviation of the true traits of all respondents, while \({\overline{\hat\eta}_d}\) and \({S}_{{\hat{\eta}}_d}\) are the mean value and standard deviation of the estimated traits respectively. The smaller the BIAS and RMSE values and the larger the CORR values, the higher the trait estimation accuracy.

The x2 index is employed to measure the overall exposure and it is defined as

$${x}^2=\sum_{j=1}^J\frac{{\left[{ER}_j-E\left({ER}_j\right)\right]}^2}{E\left({ER}_j\right)},$$
(25)

where ERj = fi/N is the exposure rate of block j, fj is the number of times that block j is selected. E(ERj) = T/J is the expected exposure rate of block j, T is the test length, and J is the number of blocks in item pool (Chang & Ying, 1999). The smaller the x2 is, the more evenly the whole item pool is used.

Results of study 1

Trait estimation accuracy

The trait estimation accuracies of the five compared item selection methods (MFC-A-optimality, MFC-D-optimality, MFC-KI, MFC-KB, and MFC-KLP) under different inter-trait correlations and test lengths in the three-dimensional MFC-CAT scenarios are presented in Table 2. As shown, all average RMSEs ranged from 0.308 to 0.582, all CORRs ranged from 0.803 to 0.951, and all biases were around zero, which indicates that the trait estimation accuracy of MFC-CAT adaptive methods was relatively high across all three-dimensional conditions. Except for the MFC-KI method, all the other methods achieved satisfactory estimation accuracy, which demonstrates their applicability to MFC-CAT. Note that: (1) Among the existing FI-based item selection methods, MFC-A-optimality was comparable to MFC-D-optimality with slightly higher estimation accuracy of the latter. (2) Among the proposed KL-based item selection methods, MFC-KI performed noticeably worse than the other two methods, as it rendered the largest RMSE and BIAS, and the smallest CORR. (3) Among the five item selection methods, MFC-KB performed similarly to MFC-KLP with higher trait estimation accuracy, which demonstrates that the proposed methods based on the KL information outperformed the existing methods based on the FI, especially when the test is short. These results are in line with the original expectations of this study.

Table 2 Trait estimation accuracy of the five compared item selection methods for three-dimensional MFC-CAT

Other factors held constant, the inter-trait correlations have a non-negligible influence on the trait estimation accuracy of MFC-CATs implemented in this study. The RMSEs increase and the CORRs decrease as the inter-trait correlations increase. In other words, the trait estimation accuracy of all methods reduces considerably as the inter-trait correlations increase, which is consistent with the results of Brown and Maydeu-Olivares’s (2011) and Bürkner et al.’s (2019) study. For example, the average RMSEs of the MFC-KB method ranged from 0.308 to 0.474 when the inter-trait correlation was 0, and took higher values ranging from 0.315 to 0.515 when the inter-trait correlation was 0.5 (see Table 2).

By contrast, the RMSEs of all methods decrease and the CORRs increase as the test length increases. It was evident that the estimation accuracy gradually improves as the test length increases. According to the results, the estimation accuracy of the 15-block tests performed better than the 5-block or 10-block tests. For example, in the conditions where the inter-trait correlation was 0, the average RMSEs of all methods for the 15-block tests ranged from 0.308 to 0.354, versus 0.474 to 0.531 for the 5-block tests.

Similarly, in the same condition, the CORRs of the 15-block tests ranged from 0.935 to 0.951, as opposed to 0.844 to 0.881 for the 5-block tests. As the test length increases, the difference of estimation accuracy between the proposed item selection methods and the existing MFC methods narrowed down. In sum, the proposed KL-based MFC-KB and MFC-KLP methods performed better than the FI-based item selection methods in terms of trait estimation accuracy, especially when the test is short (or equivalently, at an early stage of MFC-CATs). However, the performance of the MFC-KI method needs to be further improved with lower trait estimation accuracy. The same pattern was consistently observed from other indices, as well.

Uniformity of item pool usage

Item exposure control is an important component in CAT design and operation, especially for high-stake tests. Stocking and Lewis (1998) pointed out that in order to reduce the cost of item pool development, adaptive selection methods should also maximize the utilization of the item pool. Table 3 shows the results of the x2 values. The results demonstrated that the proposed MFC-KB method rendered the lowest x2 values across five methods. Namely, MFC-KB outperformed the existing FI-based MFC item selection methods in terms of uniformity of the item pool. The MFC-KLP method promoted greater utilization of the item pool and produced smaller x2 values at the early stage. However, similar to the performance of estimation accuracy, MFC-KI performed the worst in item pool usage. For the FI-based item selection methods, MFC-A-optimality outperformed MFC-D-optimality in uniformity of item pool usage, though the former’s accuracy was slightly worse than the latter. For example, the x2 values of MFC-D-optimality was as high as 40.605 compared with the largest x2 values of 39.313 when MFC-A-optimality was used. Overall, the use of the item pool was relatively more even when the KL-based item selection methods were used than when the FI-based item selection methods were used.

Table 3 The x2 values of the five compared item selection methods for three-dimensional MFC-CAT

Simulation study 2

Simulation design

Simulation study 1 mainly discussed the feasibility of all item selection methods under the three-dimensional MFC-CAT scenarios. In practice, however, MFC tests may need to measure more than three dimensions, namely higher-dimensional tests (e.g., TAPAS; Drasgow et al., 2012; Stark et al., 2014). Hence, study 2 intends to further explore the performance of all methods in relatively higher-dimensional (i.e., five-dimensional) MFC-CAT scenarios. At the same time, the performance of each method in the five-dimensional conditions is compared against study 1.

The simulation design of study 2 was mostly the same as study 1, except for the following aspects: first, five latent trait dimensions (d=5) were measured for triplet tests in this study. Furthermore, the number of MFC blocks administered were changed from 5, 10, and 15 blocks to 10, 15, and 20 blocks, respectively. In total, there were 5 (item selection method: MFC-A-optimality, MFC-D-optimality, MFC-KI, MFC-KB, and MFC-KLP) × 2 (inter-trait correlation: 0, 0.5) × 3 (test length: 10, 15, 20) = 30 simulation conditions. For each condition, 20 replications were conducted. EAP estimation and Gauss-Hermite numerical integration were again utilized for trait estimation with the R program. Study 2 used the same evaluation criteria as study 1.

Results of study 2

Trait estimation accuracy

For five-dimensional MFC-CATs, the RMSEs, biases, and CORRs of the five item selection methods are presented in Table 4. Overall, the average biases of all methods under various conditions were between [−0.014, 0.001]. The average RMSEs of each method under various conditions were between [0.341, 0.566], and the mean CORRs of each method were still acceptable, between [0.822, 0.936]. Therefore, the trait estimation accuracy was acceptable, which indicates that the proposed methods are also applicable to MFC-CATs under the higher-dimensional conditions.

Table 4 Trait estimation accuracy of the five compared item selection methods for five-dimensional MFC-CAT

Compared with the three-dimensional study (simulation study 1), the estimation accuracy of all methods, especially MFC-KI, decreased significantly with the increase of dimensionality. As can be seen from Table 2 and Table 4, under the three-dimensional conditions, except for MFC-KI, which generated the lowest estimation accuracy, the estimation accuracies of MFC-A-optimality and MFC-D-optimality were relatively high, and the estimation accuracies of MFC-KB and MFC-KLP were higher than the other item selection methods. Similar to the three-dimensional study, in the case of five dimensions, MFC-KB and MFC-KLP have similar performance. Among the five item selection methods, those two item selection methods have a higher estimation accuracy and a greater accuracy improvement over the other item selection methods. Moreover, the estimation accuracy of MFC-KI slightly improved, while MFC-A-optimality performed the worst. In conclusion, under both the three-dimensional and five-dimensional conditions, the proposed MFC-KB and MFC-KLP methods not only had high estimation accuracy, but also were notably better than the existing FI-based item selection methods, while MFC-KI did not perform as well as the others.

The influence of the inter-trait correlations on trait estimation of item selection methods varied by the level of correlations. Other factors held constant, the trait estimation accuracy of the five methods decreases as the inter-trait correlations increase, which is consistent with study 1. Moreover, this performance pattern was more obvious in the five-dimensional conditions. For example, in the conditions in which the dimension correlation was 0 (see Table 4), the average RMSEs for MFC-KB ranged from 0.341 to 0.427, versus 0.370 to 0.484 in which the inter-trait correlation was set to 0.5. The same pattern was consistently observed from other indices as well.

The test length also has a non-negligible impact on the estimation accuracy of methods in five-dimensional simulation. As expected, as the length of the MFC-CAT test increases, the estimation accuracy of all methods gradually improves. For example, in the conditions in which the inter-trait correlation was 0, the average RMSEs of all methods with 20-block tests ranged from 0.340 to 0.392, versus 0.427 to 0.485 for 10-block tests. This may be because the more blocks administrated in the tests, the more information was provided. Compared with the three-dimensional MFC-CAT, this trend was more notable in five-dimensional tests. When the test length increases from 10 blocks to 15 blocks, or from 15 blocks to 20 blocks, the estimation accuracy of each method significantly improved.

To confirm that our observed result patterns are also statistically significant, we performed a three-way factorial ANOVA on the RMSE outcomes, and the results are presented in Table 5. Although the two-way interactions are significant, based on Keppel and Wickens (2004), because these interaction effects are all noticeably smaller than the main effects as indicated by the smaller F values, it is meaningful to interpret the main effects as reflecting the general trends in the data. The main effect of the item selection method on RMSE was significant (F (4, 180) = 297.3, p < .001, η2 = 0.888). Multiple comparisons revealed that the KL-based methods evoked smaller RMSE than those of the FI-based methods (all p < .001). The main effect of the correlation between traits on RMSE was significant (F (1, 180) =1005.492, p < .001, η2 = 0.870). Multiple comparisons revealed that the 0 inter-trait correlation evoked smaller RMSE than those in the 0.5 inter-trait correlation condition (all p < .001). The main effect of the test length on RMSE was significant (F (2, 180) = 1694.602, p < .001, η2 = 0.958). Multiple comparisons revealed that the 10 block and 15 block conditions evoked smaller RMSE than those in the 5 block condition (all p < .001).

Table 5 Main effects of item selection method, inter-trait correlation, and test length on RMSE

Uniformity of item pool usage

The x2 values of each method are shown in Table 6. In the five-dimensional MFC-CAT, except for MFC-KI, the x2 values of all methods were relatively small. MFC-A-optimality had the most uniform exposure, while it had the lowest estimation accuracy. For the FI-based item selection methods, the higher the estimation accuracy was, the more uneven the utilization of the item pool. Among the KL-based item selection methods, MFC-KI has a relatively uneven item pool usage, while MFC-KB and MFC-KLP had more even item pool usage. On the whole, the results indicated that the uniformity of item pool usage of the proposed KL-based item selection methods also better performed in five-dimensional study.

Table 6 The x2 values of the five compared item selection methods for five-dimensional MFC-CAT

A simulation based on real data

The first two simulation studies provide evidence for the feasibility and effectiveness of the proposed KL-based item selection methods to measure various numbers of dimensions. The third simulation evaluates the proposed methods in real testing situations. This study used the Big-Five factor marker questionnaire with forced-choice items (Bunji & Okada, 2020), which measures five traits with 25 blocks, each block containing two statements measuring different traits. Based on the response data from 499 subjects provided by Bunji and Okada (2020), the Markov Chain Monte Carlo (MCMC) method was used to estimate the correlation matrix and item parameters (see Table 7), which were used as the true and generating correlation matrix and item parameters in this simulation. The real data can be found at https://osf.io/x92a3/.

Table 7 The correlation matrices of the Big-Five factor marker questionnaire

For this study, five trait dimensions were measured, and the test length was fixed to 10, 15 and 20 blocks. A total of 1000 true latent trait vectors were randomly generated from a multivariate standard normal distribution with the correlation matrices of the NEO-PIR shown in Table 7. In sum, there were 5 (item selection method: MFC-A-optimality, MFC-D-optimality, MFC-KI, MFC-KB, and MFC-KLP) × 3 (test length: 10, 15, 20) = 15 simulation conditions. For each condition, 20 replications were conducted. EAP estimation and Gauss-Hermite numerical integration were utilized for trait estimation with the R program.

For the trait estimation accuracy evaluation, the RMSEs of each dimension are presented next (BIAS and CORR are omitted for this study as previous studies revealed similar patterns as RMSE). For item exposure, the x2 index was computed.

Results

Table 8 summarizes the RMSEs and x2 values of study 3. It is evident that the estimation accuracy and uniformity of item pool usage of five item selection methods were acceptable in real testing situations. Compared with the five-dimensional MFC-CAT simulation in study 2, the estimated RMSEs of each method were relatively high. This may be because the quality of blocks in the item pool was relatively low, and the inter-trait correlations in the real correlation matrix were relatively high. The performance pattern of five methods in real testing situations was similar with that in the previous two simulation studies. For example, the average RMSEs of MFC-KB ranged from 0.723 to 0.772, which performed better than the FI-based methods. As shown in Table 8, MFC-KB yielded the smallest RMSEs, while MFC-A-optimality produced the largest RMSEs. In general, the estimation accuracies of the KL-based item selection methods exceed that of the FI-based item selection methods in real testing situations.

Table 8 The results of the five compared item selection methods for MFC-CAT based on real data

The performance pattern of the five methods in terms of uniformity of item pool usage was also similar to the first two simulation studies. Among the five methods, the item pool usage of the KL-based item selection methods is relatively even with lower x2 values, which outperformed the FI-based item selection methods. However, MFC-KI still had the worst performance.

In summary, from the perspective of the estimation accuracy and uniformity of item pool usage, MFC-KB performed the best and the proposed KL-based item selection methods generally outperformed the existing FI-based item selection methods under the circumstance of the practical NEO-PIR item pool.

Summary and discussion

MFC-CAT is a promising new research area that has gained more and more attention given that it integrates MFC personality assessment with CAT. Compared with traditional tests, MFC-CAT not only greatly reduces test time, but also eliminates response bias, thus improving test efficiency and estimation accuracy. Currently, studies on MFC-CATs were mainly focused on the FI-based item selection methods using the GGUM-RANK model (e.g., Joo et al., 2020). However, studies found that the KL-based item selection methods can be an alternative to address the issue of attenuation paradox of FI-based item selection methods (Chang & Ying, 1996; Veldkamp & van der Linden, 2002). Moreover, the TIRT model is a promising alternative model for MFC-CAT as it was widely used to model a variety of forced-choice scales and has demonstrated efficacy in accommodating many combinations of traits and block sizes (Brown & Maydeu-Olivares, 2011, 2013).

Therefore, this study constructs the MFC-CAT procedures based on the TIRT model and proposes the MFC-KI, MFC-KB, and MFC-KLP item selection methods based on the KL information for MFC-CAT. The results from three simulation studies confirmed that the proposed KL-based item selection methods outperformed the existing FI-based item selection methods, especially when the test is short (or equivalently, at an early stage of the CAT), generating greater trait estimation accuracies and utilization of the item pool. These findings are encouraging for applications of MFC-CAT to noncognitive personality evaluation in talent assessment.

More specifically, two Monte Carlo simulations and a simulation based on real data were conducted under three-dimensional, five-dimensional, and real testing settings. In these simulations, we manipulated several factors, including the number of dimensions, the inter-trait correlations, and the test length. The findings are summarized as the following.

First, the trait estimation accuracy and uniformity of item pool usage of all proposed item selection methods were acceptable. Among the five compared methods, the proposed MFC-KB and MFC-KLP performed best and comparably in terms of estimation accuracy and uniformity of item pool usage. By using the posterior distribution, these two item selection methods extract more information from the respondents (Mulder & van der Linden, 2010; Veldkamp & van der Linden, 2002), resulting in more precise trait estimation than the other methods. Except for MFC-KI, which performed the worst among all five compared methods and resulted in lower trait estimation accuracy and relatively higher utilization of the item pool. It is consistent with previous studies in single-statement MCAT (e.g., Tu et al., 2018). This may be because MFC-KI prefers blocks with high discrimination parameters in both dimensions, while blocks with larger KI do not necessarily provide higher power to discriminate η from \(\hat{\boldsymbol{\eta}}\). For example, a block j satisfying \(\sum\limits_{d=1}^p{\alpha}_{jd}\left({\hat{\eta}}_d-{\eta}_d\right)=0\) may has high KI, but it does not actually provide discrimination power with respect to η and \(\hat{\boldsymbol{\eta}}\) as \(KL\left(\hat{\boldsymbol{\eta}}\parallel \boldsymbol{\eta} \right)=0\) (Tu et al., 2018; Wang & Chang, 2011).

Second, the influence of the inter-trait correlations, test lengths, and dimensionality on various item selection methods for MFC-CAT was examined. We found that the lower the inter-trait correlations, the higher the estimation accuracy and the utilization of the item pool. These findings are consistent with similar studies (Brown & Maydeu-Olivares, 2011; Bürkner et al., 2019). The reason may be that, in forced-choice tests, as the correlation between the traits measured in each block increases, the uncertainty of the participants’ responses increases, thus reducing the trait estimation accuracy. Similarly, consistent with the previous MFC-CAT studies (Bürkner et al., 2019; Joo et al., 2020), the more test items, the higher the estimation accuracy. From three to five dimensions, the performance pattern of the five MFC-CAT item selection methods as varying by inter-trait correlations and test lengths stays the same.

Lastly, a simulation based on real data was conducted to evaluate the proposed KL-based item selection methods in a practical setting. Results show that acceptable trait estimation accuracy (in terms of RMSEs) and acceptable uniformity of item pool usage (in terms of x2 values) can also be rendered in a practical application of the proposed methods in MFC-CAT.

In sum, simulation results show that the proposed KL-based item selection methods are all viable to the MFC-CAT, and MFC-KB and MFC-KLP are the best choices recommended.

The simulation studies conducted in this research are by no means exhaustive. This article represents a crucial step in the research of MFC-CAT by exploring CAT procedures and item selection methods applicable to forced-choice items based on the TIRT model. For future studies, it is interesting to investigate other adaptive methods for MFC-CAT. The item selection methods used in this paper are extended from the single-statement MCAT. New and more efficient methods and algorithms may be explored for MFC-CAT. To make MFC-CAT more applicable in real work contexts, it is necessary to discuss the nonstatistical factors, such as item exposure control, content constraints, and so on. Moreover, in order to further verify the practical applicability of the proposed methods, real empirical research is needed. Last but not least, while the MFC-CAT simulations in this study are fixed-length tests, future research can be conducted to explore termination strategies in variable-length MFC-CAT, which may further shorten the test length and improve the efficiency and fairness of the test.