Hostname: page-component-89b8bd64d-z2ts4 Total loading time: 0 Render date: 2026-05-06T13:37:52.772Z Has data issue: false hasContentIssue false

Pair-wise comparisons of multiple models

Published online by Cambridge University Press:  01 January 2023

Stephen B. Broomell*
Affiliation:
Carnegie Mellon University, Department of Social and Decision Sciences, 219C Porter Hall, Pittsburgh, PA, 15213
David V. Budescu
Affiliation:
Department of Psychology, Fordham University, Bronx, NY, 10458
Han-Hui Por
Affiliation:
Department of Psychology, Fordham University, Bronx, NY, 10458
*
Rights & Permissions [Opens in a new window]

Abstract

Often research in judgment and decision making requires comparison of multiple competing models. Researchers invoke global measures such as the rate of correct predictions or the sum of squared (or absolute) deviations of the various models as part of this evaluation process. Reliance on such measures hides the (often very high) level of agreement between the predictions of the various models and does not highlight properly the relative performance of the competing models in those critical cases where they make distinct predictions. To address this important problem we propose the use of pair-wise comparisons of models to produce more informative and targeted comparisons of their performance, and we illustrate this procedure with data from two recently published papers. We use Multidimensional Scaling of these comparisons to map the competing models. We also demonstrate how intransitive cycles of pair-wise model performance can signal that certain models perform better for a given subset of decision problems.

Information

Type
Research Article
Creative Commons
Creative Common License - CCCreative Common License - BY
The authors license this article under the terms of the Creative Commons Attribution 3.0 License.
Copyright
Copyright © The Authors [2011] This is an Open Access article, distributed under the terms of the Creative Commons Attribution license (http://creativecommons.org/licenses/by/3.0/), which permits unrestricted re-use, distribution, and reproduction in any medium, provided the original work is properly cited.
Figure 0

Table 1: Percent of correct predictions by six models for the Hertwig etal (2004) data.

Figure 1

Figure 1: MDS mapping of the six models based on their rate of identical predictions for the choices in the Hertwig et al. (2004) data.

Figure 2

Table 2: Proportion of identical predictions (above the diagonal) and proportion of identical correct predictions (below the diagonal) for each pair of models applied to the Herwtig et al. (2004) data.

Figure 3

Table 3: Ratio of the number of choices favoring the row model and the number of choices favoring the column model in the Herwtig et al. (2004) data. Entries above the diagonal are reciprocals of the entries below the diagonal such that rij = 1/rji.

Figure 4

Table 4: The mean squared deviation between the predicted and observed choice proportions in the competition for predicting decisions from description data (Erev et al. 2010).

Figure 5

Table 5: Ratio of row/column counts of model closer to observed choice rate (Erev et al. 2010, decisions from description). Entries above the diagonal are reciprocals of the entries below the diagonal such that rij = 1/rji.

Figure 6

Table A.1. Model Descriptions for Hertwig et al. (2004) data.

Figure 7

Table B.1. Model Descriptions of top three models for Erev et al. (2010) data.

Figure 8

Table C.1. Comparison of squared deviations from the observed choice proportion (SDrow-SDcol) for each pair of models (Erev et al. 2010, decisions from description). Lower numbers show better performance.

Figure 9

Table C.2. Comparison of absolute deviations from the observed choice proportion (ADrow–ADcol) for each pair of models (Erev et al. 2010, decisions from description). Lower numbers show better performance.

Figure 10

Table C.3. Pearson correlation between the scores of the models (above diagonal) and Kendall rank order correlation between their ordering (below the diagonal) for the Erev et al. (2010) data (decisions from description).

Figure 11

Table D.1. All cases of incorrect and correct prediction for each of M=3 models.

Figure 12

Table D.2. Distribution of Kendall τ rank order correlation for simulation.

Figure 13

Table D.3. Mean (standard deviation) of the global measures of fit from the 100 bootstrap re-samples.

Figure 14

Figure D.1. Distribution of the Kendall τ rank order correlation between the model rankings derived from the global fit and pair-wise comparison approaches based on B = 100 bootstrap re-samples from each data set.