Hostname: page-component-77f85d65b8-45ctf Total loading time: 0 Render date: 2026-04-22T12:52:16.392Z Has data issue: false hasContentIssue false

Generative Adversarial Networks for High-Dimensional Item Factor Analysis: A Deep Adversarial Learning Algorithm

Published online by Cambridge University Press:  11 November 2025

Nanyu Luo
Affiliation:
Department of Applied Psychology and Human Development, University of Toronto , Canada
Feng Ji*
Affiliation:
Department of Applied Psychology and Human Development, University of Toronto , Canada
*
Corresponding author: Feng Ji; Email: f.ji@utoronto.ca
Rights & Permissions [Opens in a new window]

Abstract

Advances in deep learning and representation learning have transformed item factor analysis (IFA) in the item response theory (IRT) literature by enabling more efficient and accurate parameter estimation. Variational autoencoders (VAEs) are widely used to model high-dimensional latent variables in this context, but the limited expressiveness of their inference networks can still hinder performance. We introduce adversarial variational Bayes (AVB) and an importance-weighted extension (IWAVB) as more flexible inference algorithms for IFA. By combining VAEs with generative adversarial networks (GANs), AVB uses an auxiliary discriminator network to frame estimation as a two-player game and removes the restrictive standard normal assumption on the latent variables. Theoretically, AVB and IWAVB can achieve likelihoods that match or exceed those of VAEs and importance-weighted autoencoders (IWAEs). In exploratory analyses of empirical data, IWAVB attained higher likelihoods than IWAE, indicating greater expressiveness. In confirmatory simulations, IWAVB achieved comparable mean-square error in parameter recovery while consistently yielding higher likelihoods, and it clearly outperformed IWAE when the latent distribution was multimodal. These findings suggest that IWAVB can scale IFA to complex, large-scale, and potentially multimodal settings, supporting closer integration of psychometrics with modern multimodal data analysis.

Information

Type
Theory and Methods
Creative Commons
Creative Common License - CCCreative Common License - BYCreative Common License - NC
This is an Open Access article, distributed under the terms of the Creative Commons Attribution-NonCommercial licence (https://creativecommons.org/licenses/by-nc/4.0), which permits non-commercial re-use, distribution, and reproduction in any medium, provided the original article is properly cited. The written permission of Cambridge University Press or the rights holder(s) must be obtained prior to any commercial use.
Copyright
© The Author(s), 2025. Published by Cambridge University Press on behalf of Psychometric Society
Figure 0

Figure 1 Ten handwritten images sampled from model (a) GAN, (b) WGAN, (c) VAE, and (d) VAE-GAN. Adapted from Mi et al. (2018).

Figure 1

Figure 2 Distribution of latent variables for VAE and AVB trained on a simple synthetic dataset containing samples from four different labels.

Figure 2

Figure 3 Schematic illustration of a standard generative adversarial network.Note: In some GAN variants, real data serve only as true samples and are not fed into the generator. However, in the AVB application to IFA, the generator and discriminator take item response data as input, and the discriminator distinguishes between samples in the latent space.

Figure 3

Figure 4 Schematic comparison of the encoder and decoder designs for the AVB method and a standard VAE.

Figure 4

Figure 5 Parameter MSE and bias comparison for IWAVB and IWAE methods based on 100 replications of simulation with five-dimensional latent variables following a normal distribution.

Figure 5

Figure 6 Computation time and approximate log-likelihood comparison for IWAVB and IWAE methods across 100 replications of simulation with latent variables following a normal distribution.

Figure 6

Figure 7 Parameter MSE comparison for IWAVB, IWAE, and MHRM methods based on 100 replications of simulation with seven-dimensional and ten-dimensional latent variables following a normal distribution.

Figure 7

Table 1 Computation time comparison for IWAE, IWAVB, and MH-RM methods across 100 replications of simulation with seven-dimensional and ten-dimensional latent variables following a normal distribution

Figure 8

Figure 8 Multimodal distribution of latent variables.

Figure 9

Figure 9 Parameter MSE comparison for IWAVB and IWAE methods based on 100 replications of simulation with latent variables following a multimodal distribution.

Figure 10

Table 2 Computation time and approximate log-likelihood comparison for IWAE and IWAVB methods across 100 replications of simulation with latent variables following a multimodal distribution

Figure 11

Figure 10 Scree plot of predicted approximate log-likelihood with different numbers of latent factors.

Figure 12

Figure 11 Heat map of factor loadings estimated by IWAE and IWAVB for IPIP-FFM items.

Figure 13

Table 3 Performance metrics for the IPIP-FFM data set using IWAVB and IWAE methods