Hostname: page-component-77c78cf97d-kmjgn Total loading time: 0.001 Render date: 2026-04-24T14:42:52.154Z Has data issue: false hasContentIssue false

Psychometric Model Framework for Multiple Response Items

Published online by Cambridge University Press:  19 December 2025

Wenjie Zhou
Affiliation:
Faculty of Psychology, Southwest University , Chongqing, China Berkeley School of Education, University of California, Berkeley, USA
Lei Guo*
Affiliation:
Faculty of Psychology, Southwest University , Chongqing, China Southwest University Branch, Collaborative Innovation Center of Assessment toward Basic Education Quality, Chongqing, China
*
Corresponding author: Lei Guo; Email: happygl1229@swu.edu.cn
Rights & Permissions [Opens in a new window]

Abstract

Multiple response (MR) items—such as multiple true-false, multiple-select, and select-N items—are increasingly used in assessments to identify partial knowledge and differentiate latent abilities more accurately. Allowing multiple selections, MR items provide richer information and reduce guessing effects compared to single-answer multiple-choice items. However, traditional scoring methods (e.g., Dichotomous, Ripkey, Partial scoring) compress response combination (RC) data, losing valuable information and ignoring issues like local dependence and incompatibility across item types. To address these challenges, we introduce a novel psychometric model framework: the Multiple Response Model with Inter-option Local Dependencies (MRM-LD), and its simplified version, the Multiple Response Model (MRM). These models preserve RC data across MR item types, offering a more comprehensive understanding for MR assessment. Parameters for MRM-LD and MRM were estimated using Markov chain Monte Carlo algorithms in Stan and R. Empirical data from an eighth-grade physics test showed that MRM-LD and MRM outperform Graded Response Model and Nominal Response Model combined with three scoring methods, by retaining more test information, improving reliability and validity, and providing more detailed analysis of item characteristics. Simulation studies confirmed the proposed models perform robustly under various conditions, including small samples and few items, demonstrating their applicability across diverse testing scenarios.

Information

Type
Theory and Methods
Creative Commons
Creative Common License - CCCreative Common License - BYCreative Common License - NC
This is an Open Access article, distributed under the terms of the Creative Commons Attribution-NonCommercial licence (https://creativecommons.org/licenses/by-nc/4.0), which permits non-commercial re-use, distribution, and reproduction in any medium, provided the original article is properly cited. The written permission of Cambridge University Press or the rights holder(s) must be obtained prior to any commercial use.
Copyright
© The Author(s), 2025. Published by Cambridge University Press on behalf of Psychometric Society
Figure 0

Figure 1 An MTF example item from the PISA 2012 mathematics assessment.

Figure 1

Figure 2 A CMS example item from the NCLEX-RN practice test.

Figure 2

Figure 3 A Select-N example item from the GSAT 2024.

Figure 3

Table 1 Response combinations and corresponding scores for MR items (four-option item).

Figure 4

Table 2 ${\boldsymbol{Z}}_j$ Matrix for four-option MTF items, ${\boldsymbol{Z}}_j$ matrix for four-option CMS items, and ${\boldsymbol{Z}}_j$ Matrix for four-option MR items across different types.

Figure 5

Figure 4 The MRM-LD model structure diagram.

Figure 6

Table 3 Empirical study test booklet design.

Figure 7

Figure 5 Posterior predictive model checks for each model across different MR item types.Note: Red dashed horizontal lines mark the threshold boundaries at PPP = 0.05 and PPP = 0.95. Red points indicate PPP values outside the 0.05–0.95 acceptable range.

Figure 8

Table 4 Results of relative fit indices.

Figure 9

Figure 6 Item parameter distributions (${a}_{jo}$ and ${d}_{jo}$) of the MRM-LD for CMS items.

Figure 10

Figure 7 Item characteristic curves of the MRM-LD and GRMPS for MTF, CMS and select-N item (item 5).Note: Scores calculated using Partial Scoring method (1 point per correctly judged option). For each score level, 1st RC and 2nd RC indicate the most and second-most probable response combinations. Cumulative Probability shows the total probability of all RCs with that score. For MRM-LD, RC probabilities are calculated conditional on ${\gamma}_{ij} = 0$.

Figure 11

Figure 8 Item characteristic curves of the MRM-LD and GRMPS for CMS item type (item 13).Note: Scores calculated using Partial Scoring method (1 point per correctly judged option). For each score level, 1st RC and 2nd RC indicate the most and second-most probable response combinations. Cumulative Probability shows the total probability of all RCs with that score. For MRM-LD, RC probabilities are calculated conditional on ${\gamma}_{ij} = 0$.

Figure 12

Figure 9 The distribution of the estimate local dependence parameters ${a}_j^{\ast }$ with 95% CIs for the MRM-LD.

Figure 13

Table 5 IRT reliability of different models across item types.

Figure 14

Table 6 Criterion-related validity across nine booklets.

Figure 15

Figure 10 Test information curves for each model across different item types.Note: Information curves for MRM-LD are conditional on ${\gamma}_{ij} = 0$.

Figure 16

Table 7 Parameter estimation accuracy of each model in the simulation study 1.

Figure 17

Figure 11 Estimation accuracy of person parameter ${\theta}_i$ by various models in MTF tests.

Figure 18

Figure 12 Estimation accuracy of person parameter ${\theta}_i$ by various models in CMS tests.

Figure 19

Figure 13 Estimation accuracy of person parameter ${\theta}_i$ by various models in select-N tests.

Figure 20

Figure 14 Scatter plots of the true and estimated values of ${\theta}_i$ for the condition: MTF, J = 10, O = 5, and N = 500.

Figure 21

Table 8 Estimation accuracy of the local dependence parameter ${a}_j^{\ast }$ for the MRM-LD in simulation study 3.

Figure 22

Figure 15 Estimation accuracy of person parameter ${\theta}_i$ by various models in simulation study 3.

Figure 23

Figure 16 Estimation accuracy of person parameter ${\theta}_i$ by various models in simulation study 4.Note: True models are highlighted in red.

Figure 24

Table A1 Abbreviations list.

Figure 25

Figure A1 MRM-LD item parameter distributions (${a}_{jo}$ and ${d}_{jo}$) for MTF items in empirical study.

Figure 26

Figure A2 MRM-LD item parameter distributions (${a}_{jo}$ and ${d}_{jo}$) for select-N items in empirical study.

Supplementary material: File

Zhou and Guo supplementary material

Zhou and Guo supplementary material
Download Zhou and Guo supplementary material(File)
File 115.8 KB