Hostname: page-component-89b8bd64d-5bvrz Total loading time: 0 Render date: 2026-05-07T18:45:40.228Z Has data issue: false hasContentIssue false

Deep Computerized Adaptive Testing

Published online by Cambridge University Press:  27 March 2026

Jiguang Li*
Affiliation:
Econometrics and Statistics, The University of Chicago Booth School of Business , USA
Robert Gibbons
Affiliation:
Department of Statistics, The University of Chicago , USA
Veronika Ročková
Affiliation:
Econometrics and Statistics, The University of Chicago Booth School of Business , USA
*
Corresponding author: Jiguang Li; Email: jiguang@chicagobooth.edu
Rights & Permissions [Opens in a new window]

Abstract

Computerized adaptive tests (CATs) play a crucial role in educational assessment and diagnostic screening in behavioral health. Unlike traditional linear tests that administer a fixed set of pre-assembled items, CATs adaptively tailor the test to an examinee’s latent trait level based on their previous responses. We introduce a novel CAT system that builds on recent advances in Bayesian multivariate IRT. Our approach leverages direct sampling from the latent factor posterior distributions, significantly accelerating existing information-theoretic item-selection methods by eliminating the need for computationally intensive Markov chain Monte Carlo simulations. To address the potential suboptimality of one-step-ahead item-selection rules, we also develop a double deep Q-learning algorithm that efficiently learns an optimal item-selection policy offline using a calibrated item bank. Through simulation and real-data studies, we demonstrate that our approach not only accelerates existing item-selection methods but also highlights the potential of reinforcement learning (RL) in CATs. Notably, our Q-learning-based strategy consistently achieves the fastest posterior variance reduction, leading to earlier test termination. These results demonstrate the promise of combining exact posterior sampling with RL to deliver scalable, high-precision CATs.

Information

Type
Theory and Methods
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (https://creativecommons.org/licenses/by/4.0), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2026. Published by Cambridge University Press on behalf of Psychometric Society
Figure 0

Figure 1 Estimated bifactor factor loading matrix for pCAT-COG.

Figure 1

Figure 2 High-level architecture of the Q-network. The shared encoder $\phi _1$ maps each tuple of posterior parameters to $\mathbb {R}^{L_1}$ and the sum yields the permutation invariant representation $g_1(\tilde {\boldsymbol {\xi }}_t)$. The matrix $\boldsymbol {\Psi }_t$ is encoded by $\phi _2$. The concatenated vector in $\mathbb {R}^{L}$ is passed to the classifier $\rho $ to select the jth item (largest value in the J logits). This network is trained offline using Algorithm 1; during live CAT, its weights are fixed and only the posterior state is updated sequentially.

Figure 2

Figure 3 Number of items versus cumulative percentage of completed tests.

Figure 3

Table 1 Comparison of win shares (W.S), termination, and computation

Figure 4

Table 2 MSEs between posterior mean and ground truth for the first three latent factors as a function of test length

Figure 5

Figure 4 Distributions of item exposure rates.

Figure 6

Figure 5 pCAT-COG: Primary factor posterior variance reduction (left) and estimation accuracy (right).

Figure 7

Table 3 Comparison of termination efficiency, primary-factor accuracy, and computation for pCAT-COG

Figure 8

Table 4 Mean-squared errors (MSEs) as a function of test length

Supplementary material: File

Li et al. supplementary material

Li et al. supplementary material
Download Li et al. supplementary material(File)
File 562.3 KB