Literature DB >> 36046415

Estimating three- and four-parameter MIRT models with importance-weighted sampling enhanced variational auto-encoder.

Tianci Liu1, Chun Wang2, Gongjun Xu1.   

Abstract

Multidimensional Item Response Theory (MIRT) is widely used in educational and psychological assessment and evaluation. With the increasing size of modern assessment data, many existing estimation methods become computationally demanding and hence they are not scalable to big data, especially for the multidimensional three-parameter and four-parameter logistic models (i.e., M3PL and M4PL). To address this issue, we propose an importance-weighted sampling enhanced Variational Autoencoder (VAE) approach for the estimation of M3PL and M4PL. The key idea is to adopt a variational inference procedure in machine learning literature to approximate the intractable marginal likelihood, and further use importance-weighted samples to boost the trained VAE with a better log-likelihood approximation. Simulation studies are conducted to demonstrate the computational efficiency and scalability of the new algorithm in comparison to the popular alternative algorithms, i.e., Monte Carlo EM and Metropolis-Hastings Robbins-Monro methods. The good performance of the proposed method is also illustrated by a NAEP multistage testing data set.
Copyright © 2022 Liu, Wang and Xu.

Entities:  

Keywords:  Monte Carlo (MC) algorithm; Multidimensional Item Response Theory (MIRT); estimation; four parameter item response theory; variational auto encoder (VAE)

Year:  2022        PMID: 36046415      PMCID: PMC9421264          DOI: 10.3389/fpsyg.2022.935419

Source DB:  PubMed          Journal:  Front Psychol        ISSN: 1664-1078


1. Introduction

Item response theory (IRT) has been widely used for the evaluation and assessment of education and psychology test data. The most commonly used IRT is the 2-parameter logistic model (2PL), which is based on a logistic model for dichotomous responses and assigns a scalar factor score for each respondent. After observing its success, flexibility beyond the 2PL model has also been pursued for decades. Notably, McDonald (1967) suggested that the lower and upper asymptote in 2PL can be freed up from fixed 0 and 1, respectively. Estimating a different lower asymptote for each item results in the so-called 3PL model, which has been quite useful for multiple-choice items where guessing is possible; but little empirical evidence was found to support that estimating upper asymptote was beneficial as well; therefore, it was widely believed that the 4PL model was only of theoretical interest and there was no compelling reason for practitioners to use it (Barton and Lord, 1981; Hambleton and Swaminathan, 1985). Until the 2000s, researchers started revisiting the 4PL model and demonstrated the rationale of introducing upper asymptote parameters after observing early signs of its importance (Reise and Waller, 2003; Loken and Rulison, 2010; Waller and Reise, 2010; Yen et al., 2012). Waller and Feuerstahler (2017) took a step further and conducted a comprehensive study of 4PL model on a variety of real and synthetic data. In their experiment, the 4PL model achieved promising accuracy on medium to large data. However, despite these existing studies and estimation methods (e.g., Ogasawara, 2002; Waller and Feuerstahler, 2017; Meng et al., 2020), difficulties of parameter estimation in 3PL and 4PL models still remain, especially when data sizes are large and the latent factors exhibit a multidimensional or even high-dimensional structure. Multidimensional IRT (MIRT) models are a family of models where the latent trait is no longer assumed to be unidimensional. By allowing latent factors to exhibit multidimensional structures, 2PL, 3PL, and 4PL models are turned into the multidimensional 2PL (M2PL), 3PL (M3PL), and 4PL (M4PL) models, respectively. Compared with IRT models, MIRT models are capable to model each individual's multiple latent traits simultaneously and are usually favored by large scale and complex real data, thereof (Reckase, 2009). In this article, we study the general MIRT models with a special focus on M3PL and M4PL models. Specifically, assume that there are N individuals who respond to J items independently with binary response Y, for i = 1, …, N and j = 1, …, J. The M3PL model assumes that this response from the i-th individual to the j-th item is modeled by the following item response function (IRF). where is a K-dimensional vector of item discrimination (loading) parameters for the j-th item; b is referred to as the item easiness parameters. −b/||||2 is sometimes termed as item difficulty (Cho et al., 2021); c ≥ 0 is known as the lower asymptote of the j-th item and measures the probability of guessing j-th item correctly when is of negative infinity. Moreover, is a K-dimensional latent variable denoting the ability of i-th respondent, which is assumed to have a standard K-dimensional Gaussian distribution in IRT literature. Further generalizing M3PL, the M4PL model has an IRF of where additional d ≤ 1 is referred to as the upper asymptote parameter, which is the maximum probability of answering the j-th item correctly when goes to infinity. Intuitively, 1 − d can be treated as the slipping probability that an individual who is able to answer the item correctly but miss it accidentally. For both M3PL and M4PL models, we denote model parameters A = {, j = 1, …, J}, = {b, j = 1, …, J}, = {c, j = 1, …, J}, = {d, j = 1, …, J}; and for M3PL model d = 1, j = 1, …, J and M = {A, , , } is the collection of all model parameters. Under the typical local independence assumption, the marginal log-likelihood of M is given by where p() is the probability density function of a standard K-dimensional Gaussian distribution. Due to the latent variable structure, the K dimensional integrals involved in (3) makes maximization of the log-likelihood function with respect to M intractable. Direct numerical approximations of the integrals were proposed, including the Gauss-Hermite quadrature (Bock and Aitkin, 1981) and Laplace approximation (Tierney and Kadane, 1986; Lindstrom and Bates, 1988; Wolfinger and O'connell, 1993). However, these methods usually fail to handle complicated MIRT model, especially when the dimension K of latent factors grows: Gauss-Hermite quadrature quickly becomes computationally expensive in a high-dimensional setting; the Laplace approximation, though being efficient in computation, often performs less accurately when K increases or when the likelihood function is skewed. Monte Carlo (MC) simulations have also been applied to obtain numerical approximations for MIRT, such as Monte Carlo expectation-maximization (MCEM, McCulloch, 1997), stochastic expectation-maximization (StEM, von Davier and Sinharay, 2010), and Metropolis-Hastings Robbins-Monro (MHRM, Cai, 2010a,b). Nevertheless, MC based methods need drawing samples from posterior distributions, which could be computationally demanding as well. Recently, Zhang et al. (2020) improved StEM for item factor analysis, but its stochastic E-step involves an adaptive rejection-based Gibbs sampler and may still be time consuming. All methods discussed above can be seen as variants of the marginal maximum likelihood (MML) estimator proposed in Bock and Aitkin (1981), where latent are considered as random variables and are integrated out. Chen et al. (2019) instead studied the constraint joint maximum likelihood estimator (CJMLE) by treating as fixed effect parameters in order to achieve higher speeds. Unfortunately, many existing studies focusing on the M2PL model cannot be applied to M3(4)PL models easily: for MHRM, commercial software FlexMIRT (Chung and Houts, 2020) does not support M4PL, and for M3PL, MHRM is known to suffer from a lower convergence rate (Cho et al., 2021) than M2PL; for CJMLE, the authors only derived methods for M2PL and which not support M3(4)PL models. In general, computationally efficient estimation methods for M3(4)PL models are still under explored. Variational approaches stem from the machine learning literature, which maximizes a tractable lower bound of the log-likelihood rather than maximizing the log-likelihood directly. They have been applied to fitting IRT models in recent years (Rijmen and Jeon, 2013; Natesan et al., 2016; Hui et al., 2017; Jeon et al., 2017). More recently, these variational methods also established a variety of successes on more complicated MIRT (Curi et al., 2019; Wu et al., 2020; Cho et al., 2021) and graded response models (Urban and Bauer, 2021). Notably, variational autoencoder (VAE), deep learning based variational method, and its variation, importance weighted autoencoder (IWAE), are shown to be effective in parameters estimation and achieve performances competitive to traditional techniques at much faster speeds (Curi et al., 2019; Wu et al., 2020; Urban and Bauer, 2021). In this article, we investigate the VAE method for the more challenging M3(4)PL models with possible missing data. Extending explorations from Urban and Bauer (2021), we propose a new training strategy for VAE by enhancing it with the objective function of IWAE. As revealed in Section 2.2, although IWAE is computationally more expensive than VAE, our mixing training method inherits both the speed advantage of VAE and the better performance of IWAE. We also pay great attention to several practical issues and challenges in model training and propose corresponding methods/tricks to solve them, which allows our model to handle missing data and have better numeric robustness. Compared with the existing estimation approaches, such as MCEM and MHRM, our method succeeds in achieving comparable or better accuracy in parameter estimation and exhibits a much faster speed. Moreover, our method converges under M3(4)PL models within constant fitting times on different sizes, comparable to what Urban and Bauer (2021) found in the M2PL model, which is a key advantage of VAE based estimation over traditional methods. The rest of this article is organized as follows. Section 2 covers our new training strategy of VAE based estimation, which is named as Importance-Weighted sampling enhanced VAE (IWVAE); to make the section self-contained, we also provide an overview of VAE and IWAE; important tricks for handling missing data as well as improving numerical stability are also introduced. Section 3 provides a large-scale simulation study where IWVAE shows consistently competitive performances to MHRM and MCEM methods across different sample sizes, item structures, and asymptotic regimes. Section 4 compares three methods on a real data set from a multistage testing design. We end up this article with final discussion and remarks in Section 5.

2. Methods

We start with a brief overview of variational inference and how it helps tackle maximizing likelihoods whose exact forms are unavailable. We then introduce a gradient based model from deep learning called variational autoencoder (VAE), along with its generalization importance weighted autoencoder (IWAE). Given the importance and popularity of multilayer perceptron (MLP) in machine learning, which provides an efficient way of parameterizing and implementing VAE and IWAE, we include a concise introduction of MLP and reveal its ability to handle missing entries which are ubiquitous in large datasets. We end up this section with our new proposed mixing training method of VAE, importance-weighted sampling enhanced VAE (IWVAE). IWVAE uses both VAE and IWAE's objective functions and enjoys both benefits of them.

2.1. A review of variational inference and variational autoencoder

Since the integration in Equation (3) does not admit a closed form solution, we need a tractable objective function to approximate it, and variational inference (VI) is a machine learning technique to achieve this (Bishop, 2006; Blei et al., 2017). There are two equivalent ways to setup the VI objective. The first one aims to find the best approximation of the posterior of latent variable given and M, which results in a lower bound of Equation (3). Additionally, the second one directly derives the bound using Jensen Inequality. We start with the first derivation as it better clarifies the connection between VI and the expectation maximization (EM) algorithm (Dempster et al., 1977). The second derivation is revisited in Section 2.2 when we introduce a new tighter lower bound. Let Θ = (1, …, ) denote the collection of all latent variables. The best approximation of posterior p(Θ ∣ Y; M), which we refer to as q(Θ), is obtained by finding a candidate from some simple and tractable variational distribution families such that the Kullback-Leibler (KL) divergence DKL[q(Θ)||p(Θ∣Y; M)] is minimized. One common variational family is the factorized distribution , where the subscript ik in q(θ) is to emphasize that different dimensions can follow different distributions, or follow the same distribution but have different parameters. For instance, we can choose the popular Gaussian distribution for each q(θ), equivalently, we have q() to follow a K-dimensional diagonal Gaussian distribution. If one intends to characterize the dependence structure among different dimensions of , we may choose the factorized distribution family , with q() following a K-dimensional Gaussian distribution. Under this setting, the optimal variational approximation q*(Θ) is given by (Blei et al., 2017). Note that log p(Y ∣ M) is independent of Θ, it is easy to obtain the optimization objective and following decomposition Since DKL[q(Θ)||p(Θ ∣ Y; M)] is non-negative, the decomposition reveals the fact that minimizing Equation (5) is equivalent to maximizing a lower bound of the marginal log-likelihood, which is known as evidence lower bound (ELBO). Remark 1. The derivation of VI above has a close connection to the EM algorithm. Using the decomposition from Bishop ( where q(Θ) is an arbitrary distribution that includes the variational distribution families. And the first term is precisely the ELBO. In the EM algorithm, is maximized with respect to q(Θ) and M in an iterative way. In the E-step, the maximization is over q(Θ), which requires a closed-form solution: the true posterior of Θ given Y and fixed . By doing so the second KL divergence disappears and . Since the right hand side does not depend on q(Θ), the ELBO takes equality thereof. In M-step, the M is optimized to maximize the by fixing q(Θ). By repeating two steps the EM algorithm is guaranteed to converge to a local optimum of log p(Y ∣ M). The main difference between VI on IRT and EM algorithm is that because p(Θ ∣ Y; M) is intractable, we cannot obtain the analytic update of q(Θ) in each step, as a result, plain EM algorithm does not scale up well to the high-dimensional MIRT model. VI, on the other hand, finds a tractable approximation in its “E-step” and consequently, it always optimizes a strict lower bound. In general, another philosophical difference between VI and EM is that unknown parameters in VI are usually treated as latent variables as well, refer to Bishop ( and latent variable Θ, but this is not necessary, refer to Wu et al. ( was also treated as latent variables as well and modeled together with Θ. Evidence lower bound derived in Equation (5) is a global lower bound of the marginal log-likelihood of all observations. Given the local independence assumption, we can obtain a tighter lower bound by constructing each individual a corresponding local lower bound. Deriving local lower bounds indicate finding q() such that DKL[q()||p( ∣ ; M)] is minimized. This is called local variational methods, we recommend Chapter 10.4 of Bishop (2006) for more detailed explanations, and Cho et al. (2021) for its successful implementations on M2PL and M3PL models. However, despite the success of Cho et al. (2021), in general, the local variational method is computationally expensive on large scale data. One alternate to handle this challenge is the amortized variational inference (AVI). To characterize , where denotes the diagonal of the covariance matrix, AVI assumes that depend on through a function F(·) parameterized by , formally Henceforth, we denote q() as q( ∣ ). In practice, F can be flexible and expressed by a deep neural network. One of its most famous applications on AVI is the variational autoencoder (VAE) proposed by Kingma and Welling (2014). VAE uses two neural networks together to maximize the ELBO bound: F is termed as inference or encoder network (please refer to Section 2.3.1 for the specification of F); the other generative or decoder network learns the generative process of given , where this process in MIRT is essentially estimating model parameters M. In VAE, and M are learned through stochastic gradient descents. Following Kingma and Welling (2014) and Urban and Bauer (2021), we give a brief review here. Note the ELBO for the i-th individual is given by The gradient ∇ ELBO can be estimated readily with S Monte Carlo samples for s = 1, …, S as . However, gradient ∇ELBO cannot be obtained in the same way, as in general ∇ and 𝔼 cannot be switched. To solve this problem, Kingma and Welling (2014) reparameterized as follows where ⊙ means the element-wise multiplications. By transforming the integration over q( ∣ ) to p(), we have Then, the first term can be estimated with Monte Carlo samples , and the second term can be computed effectively by observing that the KL divergence between q( ∣ ) and p() has an analytic form (Kingma and Welling, 2014). The gradient can be computed readily through the chain rule thereof. For more details, please refer to Kingma and Welling (2014) and Urban and Bauer (2021).

2.2. Importance weighted variational inference

Since the ELBO is a lower bound of the marginal likelihood that we want to maximize, a tighter ELBO is appealing as the true likelihood can be approximated more accurately. It is known that the tightness of the ELBO is coupled with the expressiveness of the variational family and limited expressivity can negatively affect the learned models, and there have been many works on reducing the gap between ELBO and marginal log-likelihood (Burda et al., 2016; Kingma et al., 2016; Kingma and Welling, 2019). Some studies aimed to extend the capacity of the variational family, and techniques including normalizing flows have been applied (Kingma et al., 2016; Papamakarios et al., 2021). Burda et al. (2016) introduced a new importance-weighted ELBO (IW-ELBO) which alleviated the coupling without changing the variational families. To better illustrate the connection between IW-ELBO bound and ELBO, we start with the second derivation of ELBO via Jensen Inequality. The above derivation can be generalized as follows Equation (13) is known as IW-ELBO where . When q is reparameterizable, Monte Carlo estimates of IW-ELBO and its gradient are given by where S and R are corresponding numbers of Monte Carlo samples and importance-weighted samples. Replacing ELBO with IW-ELBO in VAE leads to IWAE, which is a generalization of the VAE, as indicated by observing that IW-ELBO will reduce to ELBO for R = 1. Notably, IW-ELBO increases in R and converges to log p( ∣ M) as R → ∞ under mild conditions (Burda et al., 2016). However, Rainforth et al. (2018) showed that using more important samples is not always helpful. The authors introduced the signal-to-noise ratio (SNR) of an estimator δ as the ratio between the absolute value of its expectation and its SD, i.e., SNR(δ) ≜ |𝔼(δ)|/σ(δ). Then they show the below orders (rewritten with our notations) In words, for any given S, increasing R makes gradient estimates of parameters in the inference network noisier. Despite the fact that the estimates of M along may benefit from a tighter likelihood bound, the final result can be deteriorated due to the worse inference network as shown in Rainforth et al. (2018). To mitigate this problem, one simple solution is to increase S of the same order, but such modification takes more computational costs and slows down the training. We apply the doubly reparameterized gradient estimator (DReG) from Tucker et al. (2018), a recently developed method that gets rid of a similar issue. Specifically, we use the below estimator to update the inference network Empirically, computing IW-ELBO and its gradient estimates can be numerically unstable due to exponential operations involved in and . To solve this problem, we compute , and apply the well-known log-sum-exp trick (Zhang et al., 2021) to in Equation (14) and in Equations (15) and (16) as follows:

2.3. Implementation details

2.3.1. MLP and optimization

We provide a basic overview of multilayer perceptron (MLP) applied in this study, which is used to model the variational distribution q as in Equation (8). For more details about MLP and DNN, we recommend readers to Goodfellow et al. (2016). Multilayer perceptron, also known as feedforward neural networks (FNN), is one of the most popular architectures of neural networks because of its simple form and flexibility. To approximate an unknown function f* such that = f*() where ∈ ℝ, ∈ ℝ, MLP takes the recursive form = f(W + ), l = 1, …, L, and 0 = , = . Here, f1, …, f are scalar functions which are almost everywhere differentiable and are applied elementwisely when inputs are vectors. These functions are typically termed as activation functions. When f1, …, f are set to identity function g(z) = z, MLP will reduce to linear regression; using non-linear activation functions, we get a flexible function u = f(v). MLP has been shown an universal approximator under a variety of activation functions (Cybenko, 1989; Hornik, 1991; Sonoda and Murata, 2017), including sigmoid function g(x) = 1/(e− + 1), rectified linear unit function (ReLU, Nair and Hinton, 2010) g(x) = max(0, x), and hyperbolic tangent function (Tanh) g(x) = (e − e−)/(e + e−); refer to Goodfellow et al. (2016) for other choices of activation functions. In this article, we use Tanh activation for f1…, f. The f at the last layer is chosen depending on the data form. To see this, note that the last layer of MLP = f(W + ) can be seen as a generalized linear model with independent variable . When is continuous, f can be set to the identity function and we get the last layer a linear regression. When is binary (categorical), f can be set to the sigmoid (softmax) function and we get a logistic (multinomial logistic) regression, respectively. In this article, we use the following encoder network Here, denotes the intermediate output of the encoder given input the i-th individual data , and we have . In the decoder, to effectively utilize the gradient based method, following Kucukelbir et al. (2017), we map and from constrained ranges [0, 1] to unconstrained space ℝ through the differentiable logit(x) = log x/(1 − x) transformation and conduct gradient ascent in the unconstrained space. To avoid cluttering, we still use original notations and in the following.

2.3.2. Handling missing data

When given latent factors are conditionally independent, exactly as the MIRT models assume, MLP based VAE and IWAE can handle incomplete data containing entries missing at random (MAR) readily (Nazabal et al., 2020). Here, we provide a brief summary of the input drop-out trick. First, we replace missing entries in with zeros and denote the resultant vector as ; we further use indicator vector 1 to record which entries are observed, specifically, 1 ≜ 1(y is observed). Next, we replace with where is defined as For now, we use to emphasize that the inference network takes as input. Next we show the imputing missing entries with 0 does not influence the training. For M, based on Equation (14), its gradient estimate is determined by and does not depend on imputed entries because of the multiplication of 1, therefore M is also independent of them. Additionally, if neither nor is affected by imputed entries, then such imputation will not influence the model training as (and ) does not rely on these entries. To this end, we rely on the MLP architecture. The output of each neuron in MLP is a non-linear transformation of a linear combination of its inputs. This property ensures that all intermediate states and output of the inference network, which determines and for variational distribution , does not depend on zero entries in its inputs. These observations together guarantee the condition for (and ) being independent of imputed entries, as in both ELBO and IW-ELBO, gradient estimates of all parameters are determined by collections of these terms.

2.3.3. Training strategy and hyperparameter choices

We propose a three-stage training strategy for VAE by enhancing it with IW-ELBO. We first train a standard VAE through maximizing its own objective function ELBO. After reaching a local optimum, we train it to maximize the tighter IW-ELBO until it converges again. Since the computation cost of IW-ELBO is more expensive than ELBO, our strategy is cheaper than training an IWAE from scratch. We refer to our model as importance-weighted sampling enhanced VAE(IWVAE). To be more specific, in the first 1% of total iterations, we apply the KL annealing technique, i.e., at step t, we multiply the KL divergence term DKL[qϕ( ∣ )||p()] by a factor , where Tanl = ⌈0.01Tmax⌉ and Tmax = 2,00,000 is a pre-specified maximum number of iterations to avoid the algorithm running forever due to convergence issues. In this stage, the weight of the KL term increases from 0 to 1 linearly. KL annealing has shown great improvement in deep generative models (Gulrajani et al., 2016; Sønderby et al., 2016). The rationale behind this technique is that the KL divergence term can over-regularize the model by forcing the approximate posterior qϕ() close to the prior p() and leading the model to converge early to unsatisfactory local minimums. To mitigate this issue, at the beginning of training, we simply reduce the effect of the KL term. During the annealing stage, we fix and and only update ϕ, A, . After the annealing stage, we train IWVAE until its estimated ELBO converges such that the averaged ELBO value in every 100 steps stops increasing for L = 50 times. We refer to this stage as ELBO converging. Finally, we use importance-weighted samples to train IWVAE until it converges again in terms of IW-ELBO with this same rule. This stage is referred to IW-ELBO converging. After this stage, we end up training. Algorithm 1 demonstrates our training method in a simplified version where at each step only 1 sample is drawn randomly from the data to estimate gradients. In practice, people can instead collect multiple samples (known as a mini batch) at each step and take the average for better gradients estimators. In practice, we used a mini batch size of 16 for each iteration step throughout all stages, S = 1 Monte Carlo sample in all three stages, and R = 5 importance samples in the last IW-ELBO converging stage following (Urban and Bauer, 2021). In terms of parameter updates, we use stochastic gradient ascent with fixed step size to maximize the ELBO or IW-ELBO. We assign a smaller step size (0.001) for parameters and as their ranges are smaller, and all other parameters are optimized with step size (0.01). No further tweaks such as gradient clippings (Pascanu et al., 2013) are used.
Algorithm 1

Stochastic gradient ascent of IWVAE.

Input: data Y; latent factor's dimension K; Monte Carlo and importance sample sizes S, R; maximum number of iterations Tmax.
Initialize ϕ^,M^p using random samples
(KL annealing stage)
while iteration number t not reaching Tanl = ⌈0.01Tmaxdo
    randomly draw yi from Y;
    draw S samples θis~N(μi,σi2) with Equation (10) where (μi,σi2)=Fϕ(yi);
    compute logp(yiθis;Mp) with Equation (1) or Equation (2), DKL[qϕ(θiyi)||p(θi)] with Equation (11). Take 1 gradient ascent step on
ϕ^,M^p=argmaxϕ,Mp1Ss=1Slogp(yiθis;Mp)-tTanlDKL[qϕ(θiyi)||p(θi)]
end while
(ELBO converging stage)
while iteration number t not reaching Tmax and ELBO not converging do
    randomly draw yi from Y;
    draw S samples θis~N(μi,σi2) with Equation (10) where (μi,σi2)=Fϕ(yi);
    compute logp(yiθis;Mp) with Equation (1) or Equation (2), DKL[qϕ(θiyi)||p(θi)] with Equation (11). Take 1 gradient ascent step on
ϕ^,M^p=argmaxϕ,Mp1Ss=1Slogp(yiθis;Mp)-DKL[qϕ(θiyi)||p(θi)]
end while
(IW-ELBO converging stage)
while iteration number t not reaching Tmax and IW-ELBO not converging do
    randomly draw yi from Y;
    draw SR samples θirs~N(μi,σi2) with Equation (10) where (μi,σi2)=Fϕ(yi);
    compute logp(yiθis;Mp) with Equation (1) or Equation (2), wirs=exp[logp(yiθirs;Mp)-logqϕ(θirsyi)+logp(θirs)]. Take 1 gradient ascent step on
ϕ^,M^p=argmaxϕ,Mp1Ss=1S[log1Rr=1Rwirs]
end while
Output: parameter estimates ϕ^,M^p
Stochastic gradient ascent of IWVAE.

3. Simulation study

3.1. Data generation

To evaluate the performances of applying IWVAE to M3PL and M4PL models, we conducted a thorough simulation study. We considered both within item and between item multidimensionality. In particular, for the within item multidimensionality, each item was loaded on two factors; and for the between item multidimensionality, each item was loaded on one factor. Under both settings, items dependency were distributed to different factors evenly in an indirect way through a sparse J × K loading matrix A. Specifically, we first generated a blocked diagonal submatrix A′. Next, we repeated two steps iteratively: (1) flipped A′ horizontally, and (2) concatenated to previous results, until we have the full s-shaped matrix A. When J is not a multiple of row numbers of A′, we truncated the resultant matrix at the bottom. To make the design more realistic and challenging, we considered a missing data design. For datasets with large J, it is impractical to have all items observed from every single respondent in realistic scenarios. To reflect this concern, we randomly masked a large portion (80% in our experiments) of responses from each respondent, assuming each respondent only answer 20% of the items. Parameters M and latent factors Θ were generated as follows. For latent factor , under the independent factors setting, it was sampled from the standard multivariate Gaussian . Under the correlated factors setting, a covariance matrix Σ was first generated and shared by all . Specifically, the diagonal entries were set to 1 so that each factor has unit variance; and off-diagonal (specifically, upper diagonal) entries were sampled independently from . This Σ was accepted if it was positive semi-definitive, otherwise, another matrix was regenerated. For free parameters in the discrimination matrix , we sampled it from . For J pairs of guessing and upper asymptote parameters (c, d), we sampled them from c ~ Beta(1, 9), d ~ Beta(9, 1) in parallel and kept them if all c < d. Our experiments were conducted as follows. First, we chose latent factors for i = 1, …, N to be uncorrelated and studied two asymptotic regimes. Specifically, in the single asymptotic regime, the dimensions of items J and factors K were fixed to 100 and 5 respectively, and sample size N was increased from 500 to 10,000. In the double asymptotic regime, only K was fixed to 5 and J was increased from 100 to 500 as N grew. Under both settings, we chose N ∈ {500, 1,000, 5,000, 10,000} and in the double asymptotic settings we further chose J ∈ {100, 200, 300, 500}. Under each combination of N, J, K, we evaluated performances of IWVAE, MCEM, and MHRM on the M3(4)PL model by checking item parameters M estimation. Finally, we duplicated this series of experiments to correlated factors settings. We implemented IWVAE in PyTorch (Paszke et al., 2019) and MCEM in the mirt R package (Chalmers, 2012). All experiments were run on the same high performance computing cluster (HPCC) with 4 CPUs and 4 GB memory, and no GPU was used. MHRM was implemented with FlexMIRT (Chung and Houts, 2020) and all experiments were fitted on a laptop with Intel Intel(R) Core(TM) i7-10750H CPU and 16 GB memory. Because of the platform difference, we ran B = 100 independent replications for IWVAE and MCEM on each simulated dataset, and B = 20 replications for MHRM. To evaluate the performances of MCEM, MHRM, and IWVAE, we followed Cho et al. (2021) and Urban and Bauer (2021) and reported rooted mean squared error (RMSE) across B independent experiment replications. Specifically, for each scalar parameter ξ (one of α, b, c, d for j = 1, …, J, k = 1, …, K), RMSE for each parameter was computed by where is the estimated value from the b-th replication. The final reported RMSEs were averages of corresponding entries in matrix A or vectors , , , and standard error were shown after each value in the parenthesis. Note that the matrix A in MIRT (IRT) models can be only identified up to a rotation if no further prior constraint is imposed, and we conducted post-hoc processing on following other literature. Our transformation consisted of three steps. First, we applied the promax (Hendrickson and White, 1964) rotation to the estimated , which allowed different factors to be correlated; we denoted this intermediate result with . Next, for each column in that had a negative sum, we flipped its sign and the corresponding factor (refer to, e.g., Urban and Bauer, 2021), we marked the resultant matrix in this step as . Finally, we searched over the best permutation of columns of such that RMSE was minimized, and the corresponding RMSEs were reported in tables. We also utilized the CF-Quartimax rotation as in Cho et al. (2022) to evaluate the sparsity structure estimation of different methods. However, since sparsity estimation is not the main focus of this article, we defer presenting these results to the appendix. Finally, Considering that M3PL is notoriously hard for MHRM to fit (Cho et al., 2021), and M4PL is expected to be more difficult, we reported the success rate of each method, which refers to the percentage of successful replications. The exact definition of success for different methods differs. For MCEM, it refers to the case where the MCEM algorithm terminates and provides estimates successfully, regardless of convergence. For IWVAE, it also refers to successful termination without reaching the maximum iteration number, which implies proper convergence. The difference in success, as we shall see later, is influential: MHRM usually performed the best if it succeeds. MCEM, on the contrary, had much worse performances while succeeding in all experiments.

3.2. Numeric results

In this section, we show detailed numeric results on M estimations, which are summarized in Tables 1–8. In a nutshell, IWVAE achieved competitive or better performances compared to the two other statistical methods. IWVAE achieved much lower RMSE on nearly all item parameters in almost all experiments than MCEM; and unlike MHRM, IWVAE succeed in all experiments from small- to large-scale datasets. Additionally, IWVAE required much more scalable training times on all experiments, while MCEM and MHRM had time costs growing faster as sample size increased.
Table 1

Mean and SE of RMSE of M estimate on M4PL models under single regime setting, best results are in bold.

N, J Item structure Model rot(A) b c d Success rates
500100BetweenMCEM9.400 ± 0.18111.477 ± 0.4240.175 ± 0.0100.183 ± 0.0101.00
MHRM/////
IWVAE 0.674 ± 0.02 0.384 ± 0.027 0.081 ± 0.008 0.087 ± 0.008 1.00
WithinMCEM10.406 ± 0.24011.500 ± 0.4810.163 ± 0.0100.146 ± 0.0081.00
MHRM/////
IWVAE 0.744 ± 0.022 0.402 ± 0.034 0.073 ± 0.008 0.088 ± 0.008 1.00
1000100BetweenMCEM8.230 ± 0.17011.785 ± 0.4500.189 ± 0.0120.178 ± 0.0111.00
MHRM/////
IWVAE 0.498 ± 0.019 0.341 ± 0.028 0.079 ± 0.008 0.080 ± 0.008 1.00
WithinMCEM7.799 ± 0.2308.999 ± 0.4640.132 ± 0.0100.161 ± 0.0111.00
MHRM/////
IWVAE 0.609 ± 0.022 0.386 ± 0.029 0.069 ± 0.007 0.078 ± 0.008 1.00
5,000100BetweenMCEM3.240 ± 0.1564.351 ± 0.2760.189 ± 0.0110.155 ± 0.0111.00
MHRM/////
IWVAE 0.369 ± 0.027 0.378 ± 0.043 0.091 ± 0.011 0.082 ± 0.009 1.00
WithinMCEM3.235 ± 0.2212.939 ± 0.2540.139 ± 0.0120.133 ± 0.0101.00
MHRM/////
IWVAE 0.535 ± 0.029 0.383 ± 0.035 0.075 ± 0.008 0.086 ± 0.010 1.00
10,000100BetweenMCEM1.988 ± 0.0992.690 ± 0.1840.174 ± 0.0100.186 ± 0.0111.00
MHRM/////
IWVAE 0.379 ± 0.028 0.399 ± 0.042 0.084 ± 0.008 0.079 ± 0.008 1.00
WithinMCEM1.823 ± 0.1451.674 ± 0.1510.136 ± 0.0090.125 ± 0.0071.00
MHRM/////
IWVAE 0.516 ± 0.030 0.343 ± 0.038 0.084 ± 0.010 0.079 ± 0.008 1.00

Factors are diagonal. Between item structure: each item depends on 1 factor. Within item structure: each item depends on 2 factors.

Table 8

Mean and SE of RMSE of M estimate on M3PL models under double regime setting, best results are in bold.

N, J Item structure Model rot(A) b c Success rates
500100BetweenMCEM9.137 ± 0.39612.034 ± 0.7990.192 ± 0.0121.00
MHRM 0.331 ± 0.040 0.441 ± 0.062 0.077 ± 0.006 0.05
IWVAE0.659 ± 0.020 0.411 ± 0.029 0.081 ± 0.0081.00
WithinMCEM7.644 ± 0.4657.291 ± 0.6120.153 ± 0.0111.00
MHRM 0.492 ± 0.048 0.364 ± 0.054 0.064 ± 0.005 0.45
IWVAE0.733 ± 0.0200.440 ± 0.0380.073 ± 0.0081.00
1000200BetweenMCEM4.662 ± 0.2007.364 ± 0.4100.220 ± 0.0101.00
MHRM///0.00
IWVAE 0.512 ± 0.014 0.329 ± 0.022 0.080 ± 0.006 1.00
WithinMCEM2.233 ± 0.1712.294 ± 0.2660.104 ± 0.0061.00
MHRM///0.00
IWVAE 0.670 ± 0.014 0.318 ± 0.020 0.080 ± 0.005 1.00
5,000300BetweenMCEM0.576 ± 0.0170.450 ± 0.0480.101 ± 0.0061.00
MHRM///0.00
IWVAE 0.450 ± 0.012 0.263 ± 0.016 0.084 ± 0.005 1.00
WithinMCEM0.728 ± 0.013 0.215 ± 0.017 0.050 ± 0.003 1.00
MHRM///0.00
IWVAE 0.624 ± 0.013 0.290 ± 0.0180.082 ± 0.0051.00
10,000500BetweenMCEM0.571 ± 0.0120.720 ± 0.0490.089 ± 0.0041.00
MHRM///0.00
IWVAE 0.451 ± 0.011 0.246 ± 0.013 0.086 ± 0.004 1.00
WithinMCEM0.858 ± 0.0181.411 ± 0.087 0.065 ± 0.002 1.00
MHRM///0.00
IWVAE 0.587 ± 0.011 0.267 ± 0.014 0.085 ± 0.0041.00

Factors are correlated. Between item structure: each item depends on 1 factor. Within item structure: each item depends on 2 factors.

Mean and SE of RMSE of M estimate on M4PL models under single regime setting, best results are in bold. Factors are diagonal. Between item structure: each item depends on 1 factor. Within item structure: each item depends on 2 factors. Mean and SE of RMSE of M estimate on M4PL models under double regime setting, best results are in bold. Factors are diagonal. Between item structure: each item depends on 1 factor. Within item structure: each item depends on 2 factors. Mean and SE of RMSE of M estimate on M3PL models under single regime setting, best results are in bold. Factors are diagonal. Between item structure: each item depends on 1 factor. Within item structure: each item depends on 2 factors. Tables 1–4 show RMSE of M estimation in M4PL and M3PL models under single and double asymptotic regimes, where different entries in each latent factor were generated independently. Two item structures were reported together in the same table. First, we observed that MCEM and IWVAE are more robust, as they succeed in all experiments, while MHRM achieved a success rate of 50% on few experiments in the M3PL model. Next, IWVAE reached much lower RMSE than MCEM, especially on small to medium sized data. In addition, IWVAE showed similar tendencies as MCEM and MHRM did: as N grew, its RMSE showed remarkable decreases, and on more challenging within-item structure scenarios, IWVAE also had slightly higher RMSEs.
Table 4

Mean and SE of RMSE of M estimate on M3PL models under double regime setting, best results are in bold.

N, J Item structure Model rot(A) b c Success rates
500100BetweenMCEM10.020 ± 0.31813.638 ± 0.7520.213 ± 0.0131.00
MHRM 0.217 ± 0.026 0.567 ± 0.0800.099 ± 0.0070.35
IWVAE0.641 ± 0.022 0.391 ± 0.031 0.081 ± 0.008 1.00
WithinMCEM8.133 ± 0.4598.687 ± 0.7270.194 ± 0.0141.00
MHRM 0.417 ± 0.034 0.345 ± 0.049 0.078 ± 0.0050.40
IWVAE0.708 ± 0.0210.461 ± 0.039 0.073 ± 0.008 1.00
1000200BetweenMCEM5.224 ± 0.1928.544 ± 0.4070.243 ± 0.0111.00
MHRM///0.00
IWVAE 0.506 ± 0.015 0.348 ± 0.024 0.080 ± 0.006 1.00
WithinMCEM2.976 ± 0.1933.441 ± 0.3040.157 ± 0.0081.00
MHRM///0.00
IWVAE 0.638 ± 0.015 0.345 ± 0.020 0.080 ± 0.005 1.00
5,000300BetweenMCEM0.612 ± 0.0190.508 ± 0.0480.114 ± 0.0071.00
MHRM///0.00
IWVAE 0.459 ± 0.012 0.261 ± 0.016 0.084 ± 0.005 1.00
WithinMCEM0.693 ± 0.0150.306 ± 0.020 0.075 ± 0.004 1.00
MHRM///0.00
IWVAE 0.595 ± 0.013 0.271 ± 0.017 0.082 ± 0.0051.00
10,000500BetweenMCEM0.561 ± 0.0100.572 ± 0.0320.097 ± 0.0041.00
MHRM///0.00
IWVAE 0.465 ± 0.011 0.258 ± 0.013 0.086 ± 0.004 1.00
WithinMCEM0.678 ± 0.0100.581 ± 0.047 0.058 ± 0.003 1.00
MHRM///0.00
IWVAE 0.592 ± 0.011 0.271 ± 0.014 0.085 ± 0.0041.00

Factors are diagonal. Between item structure: each item depends on 1 factor. Within item structure: each item depends on 2 factors.

Mean and SE of RMSE of M estimate on M3PL models under double regime setting, best results are in bold. Factors are diagonal. Between item structure: each item depends on 1 factor. Within item structure: each item depends on 2 factors. For experiments where each latent factor has correlated components, we organized results in the same way as before. Tables 5–8 show RMSE of M estimation in M4PL and M3PL models under single and double asymptotic regimes. Again, we observed similar results from IWVAE to MCEM and MHRM in terms of success times and RMSE, indicating the advantage of the proposed IWVAE method.
Table 5

Mean and SE of RMSE of M estimate on M4PL models under single regime setting, best results are in bold.

N, J Item structure Model rot(A) b c d Success rates
500100BetweenMCEM11.248 ± 0.21713.315 ± 0.4910.172 ± 0.0110.178 ± 0.0101.00
MHRM/////
IWVAE 0.654 ± 0.023 0.363 ± 0.024 0.081 ± 0.008 0.087 ± 0.008 1.00
WithinMCEM12.038 ± 0.28612.611 ± 0.6650.148 ± 0.0100.138 ± 0.0081.00
MHRM/////
IWVAE 0.736 ± 0.022 0.416 ± 0.032 0.073 ± 0.008 0.088 ± 0.008 1.00
1000100BetweenMCEM8.231 ± 0.15911.774 ± 0.4170.180 ± 0.0110.189 ± 0.0121.00
MHRM/////
IWVAE 0.486 ± 0.019 0.334 ± 0.028 0.079 ± 0.008 0.080 ± 0.008 1.00
WithinMCEM7.181 ± 0.3027.160 ± 0.5270.108 ± 0.0090.134 ± 0.0101.00
MHRM/////
IWVAE 0.623 ± 0.020 0.408 ± 0.033 0.069 ± 0.007 0.078 ± 0.008 1.00
5,000100BetweenMCEM3.315 ± 0.1674.790 ± 0.3540.182 ± 0.0140.159 ± 0.0121.00
MHRM/////
IWVAE 0.379 ± 0.026 0.363 ± 0.043 0.091 ± 0.011 0.082 ± 0.009 1.00
WithinMCEM1.886 ± 0.1521.468 ± 0.1860.094 ± 0.0090.087 ± 0.0071.00
MHRM/////
IWVAE 0.529 ± 0.031 0.397 ± 0.037 0.075 ± 0.008 0.086 ± 0.010 1.00
10,000100BetweenMCEM1.918 ± 0.1142.555 ± 0.2130.157 ± 0.0100.176 ± 0.0111.00
MHRM/////
IWVAE 0.380 ± 0.028 0.402 ± 0.040 0.084 ± 0.008 0.079 ± 0.008 1.00
WithinMCEM1.127 ± 0.0750.943 ± 0.0790.095 ± 0.0070.087 ± 0.0061.00
MHRM/////
IWVAE 0.520 ± 0.033 0.360 ± 0.039 0.085 ± 0.010 0.079 ± 0.008 1.00

Factors are correlated. Between item structure: each item depends on 1 factor. Within item structure: each item depends on 2 factors.

Mean and SE of RMSE of M estimate on M4PL models under single regime setting, best results are in bold. Factors are correlated. Between item structure: each item depends on 1 factor. Within item structure: each item depends on 2 factors. Mean and SE of RMSE of M estimate on M4PL models under double regime setting, best results are in bold. Factors are correlated. Between item structure: each item depends on 1 factor. Within item structure: each item depends on 2 factors. Mean and SE of RMSE of M estimate on M3PL models under single regime setting, best results are in bold. Factors are correlated. Between item structure: each item depends on 1 factor. Within item structure: each item depends on 2 factors. Mean and SE of RMSE of M estimate on M3PL models under double regime setting, best results are in bold. Factors are correlated. Between item structure: each item depends on 1 factor. Within item structure: each item depends on 2 factors. We finally analyzed the fitting time of IWVAE and MCEM and reported averaged time with stand error (in shadow) in Figure 1 (M3PL) and Figure 2 (M4PL). We combined different factor settings (independent and correlated), and item structures (between and within) for every pair of sample size N and item size J. Each point contains 80 trials. As MHRM could not fit M4PL and its convergence results under M3PL were not stable, here, we do not report their results. From Figures 1, 2, compared to MCEM, IWVAE required significantly lower fitting time. Unlike MCEM, IWVAE had a much more stable fitting time across different data sizes, which was also observed in Urban and Bauer (2021) for estimating M2PL. As in Urban and Bauer (2021), we also note that the computational time of IWVAE appeared not to increase with N and J, which may be due to that VAE-based models are more difficult to train on small data sets. Similarly, in some cases, the computational time of MCEM also dropped when N increased to 10, 000, which may also be because of the easier convergence of the algorithm for the larger datasets. Moreover, we observed that fitting time of MCEM under M4PL depended more on the choices of initialization, revealed by the width of empirical intervals in Figure 2.
Figure 1

Fitting times for IWVAE and MCEM with M3PL model under the single (different sample sizes N and fixed item dimension J) and double [different (N, J)] asymptotic regimes. Vertical bar areas mark empirical 95% intervals.

Figure 2

Fitting times for IWVAE and MCEM with M4PL model under the single (different sample sizes N and fixed item dimension J) and double [different (N, J)] asymptotic regimes. Vertical bar areas mark empirical 95% intervals.

Fitting times for IWVAE and MCEM with M3PL model under the single (different sample sizes N and fixed item dimension J) and double [different (N, J)] asymptotic regimes. Vertical bar areas mark empirical 95% intervals. Fitting times for IWVAE and MCEM with M4PL model under the single (different sample sizes N and fixed item dimension J) and double [different (N, J)] asymptotic regimes. Vertical bar areas mark empirical 95% intervals.

4. Real data analysis

In this section, we evaluated the performance of IWVAE, MCEM, and MHRM on the multistage testing (MST) dataset from the National Assessment of Education Progress (NAEP). The data is from the 2011 grade 8 math assessment study. The NAEP MST design takes a two-stage form: in the routing stage, a block of items with medium difficulty is administered. Then in the second stage, there are three targeted blocks with varying difficulty—blocks of easy, medium, and hard items. Based on a person's performance in the routing block, one of the three targeted blocks is assigned in the second-stage accordingly. Because the assignment in stage II depends on the observed student performance in stage I, the MST design essentially generates a unique missing-at-random pattern. Due to the prevalence of MST design in large scale assessments, it would interesting to evaluate how the different estimation methods fare with such a design. The data set contains N = 3,344 respondents and 74 items in total. The routing block contains two parallel forms with 17 items in each form. The three blocks in stage II contain 14, 13, and 13 items, respectively. Each person responded to 31 or 30 items out of 74. The items cover 5 different content domains, i.e., number properties and operations, measurement, geometry, data analysis statistics and probability, and algebra. The break down of items from each content domain in each form is presented in Table 8 by Wang et al. (2020). The content coverage is pretty balanced, which suggests a five-dimensional model to be appropriate. Hence, five-dimensional exploratory M2PL, M3PL, and M4PL models were fitted to the MST data using IWVAE, MHRM, and MCEM. We used the same algorithm, architecture, hyper-parameters, and stop criteria on IWVAE as in the simulation study except that we use a larger learning rate of 0.1 for ϕ, A, and 0.01 for , . First, we studied the estimates of the covariance matrix Σ and the comparison between the three methods. Due to the identifiability issue, in all models we assumed that covariance matrix of latent factors is Σ = I and conducted the promax rotation to estimate the correlation matrix . After rotation, we adjusted the sign of the correlation depending on the sign of post-hoc transformed as in the previous section. In particular, we flipped the sign of each column in if its sum was negative, and did the same to the corresponding columns and rows in the . Table 9 shows estimated matrices under M4PL, M3PL, and M2PL, respectively. Under all settings, the correlation matrix recovered by IWVAE was similar to those from MCEM and MHRM, and a bit even closer to MHRM than MCEM did on M3PL and M2PL models.
Table 9

Comparison of estimated R from different models on MST dataset.

Model IWVAE MCEM MHRM
M4PL [1.0.6241.0.5710.511.0.6950.6250.5511.0.530.5210.4570.5461.] [1.0.6281.0.6160.6041.0.5950.5990.5431.0.6550.6550.6210.6261.] /
M3PL [1.0.5851.0.5570.6991.0.5310.6710.6531.0.470.520.5210.4961.] [1.0.4651.0.6690.3631.0.7120.4490.6621.0.6680.430.6030.6651.] [1.0.5811.0.520.5891.0.5650.6890.5821.0.480.5860.4940.611.]
M2PL [1.0.741.0.6220.6281.0.580.5290.5151.0.5940.5930.5070.4511.] [1.0.2641.0.4220.4691.0.4790.5640.7091.0.4350.4790.6250.691.] [1.0.5181.0.5930.5841.0.5480.5480.6181.0.5510.6010.6240.6381.]
Comparison of estimated R from different models on MST dataset. Next, given that the true parameters are unknown, we evaluated the predictive performances of the three methods using a held-out validation. That is, we randomly marked 20% of items as missing, which played the role of held-out data, and used the remaining data to estimate our person and item parameters. That is, we used the estimated parameters to produce model-based predicted responses, compared them with the observed responses, and computed their consistency as a measure of accuracy. We computed such accuracy on both training data and held-out data. Higher accuracy indicates more alignment between model prediction and observed data. Meanwhile, we also used the estimated model parameters to compute log-likelihood, with a higher likelihood implying the estimated parameters may be better reflective of unknown truth. We reported accuracy and log-likelihood predicted by different methods on the training and held-out data in Table 10. To eliminate potential randomness in generating observed responses, 5 replications were done for each model, and we generated a different train and held-out data in each replication.
Table 10

Mean and SE of train and held-out accuracy/log-likelihood on MST dataset (over 5 replications).

Method Model Train accuracy Held-out accuracy Train log-likelihood Held-out log-likelihood
IWVAEM4PL0.707 ± 0.0010.704 ± 0.002−0.531 ± 0.001−0.539 ± 0.001
M3PL0.707 ± 0.0000.706 ± 0.002−0.530 ± 0.000−0.537 ± 0.000
M2PL0.706 ± 0.0010.703 ± 0.001−0.531 ± 0.001−0.539 ± 0.001
MCEMM4PL0.764 ± 0.0010.693 ± 0.002−0.481 ± 0.001−0.603 ± 0.001
M3PL0.761 ± 0.0000.697 ± 0.001−0.482 ± 0.000−0.599 ± 0.000
M2PL0.759 ± 0.0010.697 ± 0.001−0.485 ± 0.001−0.589 ± 0.001
MHRMM4PL////
M3PL0.612 ± 0.0030.613 ± 0.002−0.682 ± 0.003−0.683 ± 0.003
M2PL0.622 ± 0.0030.623 ± 0.002−0.662 ± 0.003−0.664 ± 0.003
Mean and SE of train and held-out accuracy/log-likelihood on MST dataset (over 5 replications). Table 10 summarizes the averaged accuracy and log-likelihood (of each item) on the train and held-out sets, where values in parentheses are stand errors across 5 replications. In this experiment, IWVAE achieved the highest held-out accuracy and log-likelihood. Figures 3, 4 further showed the corresponding log-likelihood values of each item. First, we observed that IWVAE had much fewer outliers than MCEM; after removing outliers, IWVAE achieved the highest log-likelihood on three MIRT models. Moreover, among the three models, the held-out accuracy, training data log-likelihood, and held-out log-likelihood from IWVAE were the best for M3PL. This is expected in that the operational model for NAEP analysis is indeed 3PL.
Figure 3

Predicted log-likelihood on held-out items using different methods (IWVAE, MHRM, MCEM) to fit different MIRT models on MST data from a randomly selected trial. Outlier predictions are removed.

Figure 4

Predicted log-likelihood on held-out items using different methods (IWVAE, MHRM, MCEM) to fit different MIRT models on MST data from a randomly selected trial. Outlier predictions are kept.

Predicted log-likelihood on held-out items using different methods (IWVAE, MHRM, MCEM) to fit different MIRT models on MST data from a randomly selected trial. Outlier predictions are removed. Predicted log-likelihood on held-out items using different methods (IWVAE, MHRM, MCEM) to fit different MIRT models on MST data from a randomly selected trial. Outlier predictions are kept.

5. Discussions

In this article, we extend a variational autoencoder estimation method (Urban and Bauer, 2021) for the parameter estimation of the M3PL and M4PL models. By approximating the intractable log-likelihood with variational techniques, it provides a computationally efficient and scalable method for the estimation of large-scale assessment data. Simulation studies demonstrate that the proposed method outperforms the widely used MHRM and MCEM methods in terms of parameter recovery and computation time in both M3PL and M4PL. The proposed method is also more robust with many fewer issues of convergence. That said, we do want to caution readers that a robust algorithm cannot compensate for a lack of data. For M3PL and M4PL to be estimated well, there needs to be enough data at the two extreme ends of the latent trait scales to help estimate the lower and upper asymptote adequately. Although this study focuses on the exploratory item factor analysis, the proposed algorithm can be easily applied to the confirmatory item factor analysis, where certain entries of the loading matrix are set to be 0 by users. Such structural restrictions can be naturally incorporated into the estimation. In addition, it would be also of interest to further estimate the sparsity loading structure from the responses. This can be achieved by adding a lasso-type regularization term into the loss function (the marginal log-likelihood function), which would induce sparse estimation results from the regularized algorithms. Finally, a few interesting problems are left for future investigations. Very recent works suggested that some aspects of our training strategy can be improved; for instance, Collier et al. (2021) revealed that the missing data can be handled better than zero-imputation; and Wang et al. (2021) indicated a possible direction of understanding and solving the posterior collapse, which was solved by a KL annealing stage in our proposed method. Moreover, this work does not directly study the estimation uncertainty of the VAE estimation procedure. It is interesting to further develop valid statistical procedures to make inferences for the corresponding estimation results. Such an important problem, however, still remains unaddressed for VAE and related deep learning methods in the machine learning and statistics literature.

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Author contributions

TL contributed to implementing the studies and writing the initial draft. GX contributed to mentoring, conceptualizing ideas, acquiring funding, and manuscript revision. CW contributed to conceptualizing ideas, acquiring funding, and manuscript revision. All authors contributed to the article and approved the submitted version.

Funding

This study was partially supported by IES grant R305D200015, NSF grants SES-1846747 and NSF SES-2150601.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Table 2

Mean and SE of RMSE of M estimate on M4PL models under double regime setting, best results are in bold.

N, J Item structure Model rot(A) b c d Success rates
500100BetweenMCEM9.400 ± 0.18111.477 ± 0.4240.175 ± 0.0100.183 ± 0.0101.00
MHRM/////
IWVAE 0.674 ± 0.021 0.384 ± 0.027 0.081 ± 0.008 0.087 ± 0.008 1.00
WithinMCEM10.406 ± 0.24011.500 ± 0.4810.163 ± 0.0100.146 ± 0.0081.00
MHRM/////
IWVAE 0.744 ± 0.022 0.402 ± 0.034 0.073 ± 0.008 0.088 ± 0.008 1.00
1000200BetweenMCEM5.397 ± 0.0698.312 ± 0.1880.180 ± 0.0070.178 ± 0.0071.00
MHRM/////
IWVAE 0.500 ± 0.016 0.377 ± 0.025 0.080 ± 0.006 0.088 ± 0.007 1.00
WithinMCEM8.242 ± 0.1749.254 ± 0.3540.151 ± 0.0070.158 ± 0.0071.00
MHRM/////
IWVAE 0.600 ± 0.017 0.413 ± 0.024 0.080 ± 0.005 0.088 ± 0.007 1.00
5,000300BetweenMCEM1.624 ± 0.0581.519 ± 0.0680.163 ± 0.0060.156 ± 0.0051.00
MHRM/////
IWVAE 0.429 ± 0.014 0.338 ± 0.019 0.084 ± 0.005 0.081 ± 0.005 1.00
WithinMCEM1.388 ± 0.0790.876 ± 0.0760.092 ± 0.0050.086 ± 0.0041.00
MHRM/////
IWVAE 0.564 ± 0.015 0.319 ± 0.019 0.083 ± 0.005 0.080 ± 0.005 1.00
10,000500BetweenMCEM1.022 ± 0.0231.152 ± 0.0360.150 ± 0.0040.153 ± 0.0041.00
MHRM/////
IWVAE 0.432 ± 0.012 0.338 ± 0.015 0.086 ± 0.004 0.086 ± 0.004 1.00
WithinMCEM0.930 ± 0.0180.993 ± 0.0310.119 ± 0.0040.119 ± 0.0031.00
MHRM/////
IWVAE 0.564 ± 0.013 0.346 ± 0.015 0.085 ± 0.004 0.087 ± 0.004 1.00

Factors are diagonal. Between item structure: each item depends on 1 factor. Within item structure: each item depends on 2 factors.

Table 3

Mean and SE of RMSE of M estimate on M3PL models under single regime setting, best results are in bold.

N, J Item structure Model rot(A) b c Success rates
500100BetweenMCEM10.020 ± 0.31813.638 ± 0.7520.213 ± 0.0131.00
MHRM 0.217 ± 0.026 0.567 ± 0.0800.099 ± 0.0070.35
IWVAE0.641 ± 0.022 0.391 ± 0.031 0.081 ± 0.008 1.00
WithinMCEM8.133 ± 0.4598.687 ± 0.7270.194 ± 0.0141.00
MHRM 0.417 ± 0.034 0.345 ± 0.049 0.078 ± 0.0050.40
IWVAE0.708 ± 0.0210.461 ± 0.039 0.073 ± 0.008 1.00
1000100BetweenMCEM5.338 ± 0.2877.799 ± 0.5440.237 ± 0.0161.00
MHRM 0.159 ± 0.017 0.280 ± 0.025 0.089 ± 0.0070.30
IWVAE0.492 ± 0.0190.320 ± 0.026 0.079 ± 0.008 1.00
WithinMCEM2.564 ± 0.2353.023 ± 0.4230.120 ± 0.0111.00
MHRM 0.438 ± 0.038 0.313 ± 0.040 0.075 ± 0.0050.65
IWVAE0.590 ± 0.0200.325 ± 0.024 0.069 ± 0.007 1.00
5,000100BetweenMCEM1.031 ± 0.1101.190 ± 0.2040.144 ± 0.0131.00
MHRM 0.151 ± 0.024 0.264 ± 0.028 0.090 ± 0.008 0.30
IWVAE0.403 ± 0.024 0.259 ± 0.028 0.091 ± 0.0111.00
WithinMCEM0.881 ± 0.0630.575 ± 0.0770.097 ± 0.0091.00
MHRM 0.292 ± 0.035 0.123 ± 0.009 0.033 ± 0.004 0.90
IWVAE0.562 ± 0.0260.279 ± 0.0320.075 ± 0.0081.00
10,000100BetweenMCEM0.810 ± 0.0781.008 ± 0.1690.112 ± 0.0111.00
MHRM 0.106 ± 0.019 0.381 ± 0.131 0.055 ± 0.006 0.70
IWVAE0.393 ± 0.027 0.318 ± 0.036 0.084 ± 0.0081.00
WithinMCEM0.754 ± 0.0450.662 ± 0.1290.076 ± 0.0071.00
MHRM 0.343 ± 0.035 0.154 ± 0.012 0.040 ± 0.004 0.75
IWVAE0.535 ± 0.0270.283 ± 0.0390.084 ± 0.0101.00

Factors are diagonal. Between item structure: each item depends on 1 factor. Within item structure: each item depends on 2 factors.

Table 6

Mean and SE of RMSE of M estimate on M4PL models under double regime setting, best results are in bold.

N, J Item structure Model rot(A) b c d Success rates
500100BetweenMCEM11.248 ± 0.21713.315 ± 0.4910.172 ± 0.0110.178 ± 0.0101.00
MHRM/////
IWVAE 0.654 ± 0.023 0.363 ± 0.024 0.081 ± 0.008 0.087 ± 0.008 1.00
WithinMCEM12.038 ± 0.28612.611 ± 0.6650.148 ± 0.0100.138 ± 0.0081.00
MHRM/////
IWVAE 0.736 ± 0.022 0.416 ± 0.032 0.073 ± 0.008 0.088 ± 0.008 1.00
1000200BetweenMCEM8.314 ± 0.09711.742 ± 0.2980.178 ± 0.0080.197 ± 0.0081.00
MHRM/////
IWVAE 0.503 ± 0.015 0.362 ± 0.024 0.080 ± 0.006 0.088 ± 0.007 1.00
WithinMCEM7.323 ± 0.2137.628 ± 0.3620.125 ± 0.0060.131 ± 0.0071.00
MHRM/////
IWVAE 0.609 ± 0.015 0.407 ± 0.024 0.080 ± 0.005 0.088 ± 0.007 1.00
5,000300BetweenMCEM2.041 ± 0.0902.292 ± 0.1540.143 ± 0.0060.139 ± 0.0061.00
MHRM/////
IWVAE 0.426 ± 0.013 0.340 ± 0.020 0.084 ± 0.005 0.081 ± 0.005 1.00
WithinMCEM1.049 ± 0.0530.525 ± 0.065 0.066 ± 0.005 0.059 ± 0.004 1.00
MHRM/////
IWVAE 0.582 ± 0.014 0.339 ± 0.019 0.083 ± 0.0050.080 ± 0.0051.00
10,000500BetweenMCEM1.062 ± 0.0361.163 ± 0.0600.129 ± 0.0040.131 ± 0.0041.00
MHRM/////
IWVAE 0.426 ± 0.011 0.332 ± 0.015 0.086 ± 0.004 0.086 ± 0.004 1.00
WithinMCEM0.884 ± 0.0150.944 ± 0.0310.098 ± 0.0030.099 ± 0.0031.00
MHRM/////
IWVAE 0.562 ± 0.012 0.367 ± 0.016 0.085 ± 0.004 0.087 ± 0.004 1.00

Factors are correlated. Between item structure: each item depends on 1 factor. Within item structure: each item depends on 2 factors.

Table 7

Mean and SE of RMSE of M estimate on M3PL models under single regime setting, best results are in bold.

N, J Item structure Model rot(A) b c Success rates
500100BetweenMCEM9.137 ± 0.39612.034 ± 0.7990.192 ± 0.0121.00
MHRM 0.331 ± 0.040 0.441 ± 0.062 0.077 ± 0.006 0.05
IWVAE0.659 ± 0.020 0.411 ± 0.029 0.081 ± 0.0081.00
WithinMCEM7.644 ± 0.4657.291 ± 0.6120.153 ± 0.0111.00
MHRM 0.492 ± 0.048 0.364 ± 0.054 0.064 ± 0.005 0.45
IWVAE0.733 ± 0.0200.440 ± 0.0380.073 ± 0.0081.00
1000100BetweenMCEM5.231 ± 0.2798.080 ± 0.5730.219 ± 0.0141.00
MHRM 0.284 ± 0.027 0.350 ± 0.0410.090 ± 0.0070.25
IWVAE0.483 ± 0.019 0.312 ± 0.024 0.079 ± 0.008 1.00
WithinMCEM2.855 ± 0.2992.912 ± 0.4570.118 ± 0.0101.00
MHRM 0.428 ± 0.034 0.336 ± 0.020 0.045 ± 0.004 0.80
IWVAE0.601 ± 0.024 0.299 ± 0.026 0.069 ± 0.0071.00
5,000100BetweenMCEM0.762 ± 0.0630.638 ± 0.1170.130 ± 0.0151.00
MHRM 0.242 ± 0.019 0.120 ± 0.011 0.051 ± 0.005 0.75
IWVAE0.415 ± 0.0230.256 ± 0.0310.091 ± 0.0111.00
WithinMCEM0.986 ± 0.0750.421 ± 0.0540.066 ± 0.0061.00
MHRM 0.357 ± 0.034 0.460 ± 0.013 0.030 ± 0.002 0.80
IWVAE0.578 ± 0.029 0.308 ± 0.034 0.075 ± 0.0081.00
10,000100BetweenMCEM0.917 ± 0.1071.073 ± 0.1700.110 ± 0.0121.00
MHRM 0.155 ± 0.013 0.114 ± 0.014 0.037 ± 0.006 0.65
IWVAE0.397 ± 0.0270.297 ± 0.0350.084 ± 0.0081.00
WithinMCEM0.898 ± 0.0560.751 ± 0.1650.057 ± 0.0061.00
MHRM 0.378 ± 0.036 0.461 ± 0.027 0.017 ± 0.002 0.40
IWVAE0.547 ± 0.032 0.302 ± 0.040 0.084 ± 0.0101.00

Factors are correlated. Between item structure: each item depends on 1 factor. Within item structure: each item depends on 2 factors.

  11 in total

1.  Estimation of a four-parameter item response theory model.

Authors:  Eric Loken; Kelly L Rulison
Journal:  Br J Math Stat Psychol       Date:  2009-12-23       Impact factor: 3.380

2.  An improved stochastic EM algorithm for large-scale full-information item factor analysis.

Authors:  Siliang Zhang; Yunxiao Chen; Yang Liu
Journal:  Br J Math Stat Psychol       Date:  2018-12-03       Impact factor: 3.380

3.  A Deep Learning Algorithm for High-Dimensional Exploratory Item Factor Analysis.

Authors:  Christopher J Urban; Daniel J Bauer
Journal:  Psychometrika       Date:  2021-02-02       Impact factor: 2.500

4.  A Variational Maximization-Maximization Algorithm for Generalized Linear Mixed Models with Crossed Random Effects.

Authors:  Minjeong Jeon; Frank Rijmen; Sophia Rabe-Hesketh
Journal:  Psychometrika       Date:  2017-02-28       Impact factor: 2.500

5.  Bayesian Modal Estimation of the Four-Parameter Item Response Model in Real, Realistic, and Idealized Data Sets.

Authors:  Niels G Waller; Leah Feuerstahler
Journal:  Multivariate Behav Res       Date:  2017-03-17       Impact factor: 5.923

6.  Gaussian variational estimation for multidimensional item response theory.

Authors:  April E Cho; Chun Wang; Xue Zhang; Gongjun Xu
Journal:  Br J Math Stat Psychol       Date:  2020-10-16       Impact factor: 3.380

7.  Marginalized maximum a posteriori estimation for the four-parameter logistic model under a mixture modelling framework.

Authors:  Xiangbin Meng; Gongjun Xu; Jiwei Zhang; Jian Tao
Journal:  Br J Math Stat Psychol       Date:  2019-09-25       Impact factor: 3.380

8.  Regularized Variational Estimation for Exploratory Item Factor Analysis.

Authors:  April E Cho; Jiaying Xiao; Chun Wang; Gongjun Xu
Journal:  Psychometrika       Date:  2022-07-13       Impact factor: 2.290

9.  Joint Maximum Likelihood Estimation for High-Dimensional Exploratory Item Factor Analysis.

Authors:  Yunxiao Chen; Xiaoou Li; Siliang Zhang
Journal:  Psychometrika       Date:  2018-11-19       Impact factor: 2.500

10.  Bayesian Prior Choice in IRT Estimation Using MCMC and Variational Bayes.

Authors:  Prathiba Natesan; Ratna Nandakumar; Tom Minka; Jonathan D Rubright
Journal:  Front Psychol       Date:  2016-09-27
View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.