| Literature DB >> 30511445 |
Siliang Zhang1, Yunxiao Chen2, Yang Liu3.
Abstract
In this paper, we explore the use of the stochastic EM algorithm (Celeux & Diebolt (1985) Computational Statistics Quarterly, 2, 73) for large-scale full-information item factor analysis. Innovations have been made on its implementation, including an adaptive-rejection-based Gibbs sampler for the stochastic E step, a proximal gradient descent algorithm for the optimization in the M step, and diagnostic procedures for determining the burn-in size and the stopping of the algorithm. These developments are based on the theoretical results of Nielsen (2000, Bernoulli, 6, 457), as well as advanced sampling and optimization techniques. The proposed algorithm is computationally efficient and virtually tuning-free, making it scalable to large-scale data with many latent traits (e.g. more than five latent traits) and easy to use for practitioners. Standard errors of parameter estimation are also obtained based on the missing-information identity (Louis, 1982, Journal of the Royal Statistical Society, Series B, 44, 226). The performance of the algorithm is evaluated through simulation studies and an application to the analysis of the IPIP-NEO personality inventory. Extensions of the proposed algorithm to other latent variable models are discussed.Keywords: Gibbs sampler; full-information item factor analysis; multidimensional item response theory; proximal gradient descent; rejection sampling; stochastic EM algorithm
Year: 2018 PMID: 30511445 DOI: 10.1111/bmsp.12153
Source DB: PubMed Journal: Br J Math Stat Psychol ISSN: 0007-1102 Impact factor: 3.380