Hostname: page-component-54dcc4c588-sq2k7 Total loading time: 0 Render date: 2025-10-02T18:05:09.450Z Has data issue: false hasContentIssue false

A Beta Mixture Model for Careless Respondent Detection in Visual Analogue Scale Data

Published online by Cambridge University Press:  23 September 2025

Lijin Zhang*
Affiliation:
Graduate School of Education, Stanford University https://ror.org/00f54p054 , Stanford, CA, USA
Benjamin W. Domingue
Affiliation:
Graduate School of Education, Stanford University https://ror.org/00f54p054 , Stanford, CA, USA
Leonie V. D. E. Vogelsmeier
Affiliation:
Department of Methodology, Tilburg University https://ror.org/04b8v1s79 , Tilburg, The Netherlands
Esther Ulitzsch
Affiliation:
Centre for Educational Measurement (CEMO), University of Oslo https://ror.org/01xtthb56 , Oslo, Norway Centre for Research on Equality in Education (CREATE), University of Oslo https://ror.org/01xtthb56 , Oslo, Norway
*
Corresponding author: Lijin Zhang; Email: lijinzhang@stanford.edu
Rights & Permissions [Opens in a new window]

Abstract

Visual Analogue scales (VASs) are increasingly popular in psychological, social, and medical research. However, VASs can also be more demanding for respondents, potentially leading to quicker disengagement and a higher risk of careless responding. Existing mixture modeling approaches for careless response detection have so far only been available for Likert-type and unbounded continuous data but have not been tailored to VAS data. This study introduces and evaluates a model-based approach specifically designed to detect and account for careless respondents in VAS data. We integrate existing measurement models for VASs with mixture item response theory models for identifying and modeling careless responding. Simulation results show that the proposed model effectively detects careless responding and recovers key parameters. We illustrate the model’s potential for identifying and accounting for careless responding using real data from both VASs and Likert scales. First, we show how the model can be used to compare careless responding across different scale types, revealing a higher proportion of careless respondents in VAS compared to Likert scale data. Second, we demonstrate that item parameters from the proposed model exhibit improved psychometric properties compared to those from a model that ignores careless responding. These findings underscore the model’s potential to enhance data quality by identifying and addressing careless responding.

Information

Type
Application and Case Studies - Original
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (https://creativecommons.org/licenses/by/4.0), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2025. Published by Cambridge University Press on behalf of Psychometric Society

1 Introduction

Self-report scales are widely used in social science to measure latent constructs. To effectively capture the underlying constructs, selecting an appropriate scale format is crucial. Throughout the last decades, Likert-type scales have been established as the “go-to” standard for research in psychology, social, and educational sciences (Jebb et al., Reference Jebb, Ng and Tay2021). Likert scales involve selecting discrete ordinal categories (e.g., from strong agreement to strong disagreement), which can simplify the response process (Likert, Reference Likert1932). However, this format may also restrict the range of responses, limiting the ability to capture emotions, attitudes, and beliefs in a more nuanced way. Visual Analogue scales (VASs) are an increasingly popular alternative to Likert-type scales. A VAS (Hayes & Patterson, Reference Hayes and Patterson1921) allows respondents to indicate their feelings on a continuous line (Figure 1), which may capture nuance that Likert scales might overlook, providing a more detailed measurement of attitudes or emotions. For instance, research has demonstrated that VASs are more effective in detecting subtle changes in pain perception and are less prone to confounding factors and ceiling effects than Likert scales, particularly in the context of measuring patient satisfaction (Price et al., Reference Price, Bush, Long and Harkins1994; Voutilainen et al., Reference Voutilainen, Pitkäaho, Kvist and Vehviläinen-Julkunen2016). It is, therefore, not surprising that the utilization of VASs in applied psychological, social science, and clinical research has expanded extensively in recent years (Åström et al., Reference Åström, Thet Lwin, Teni, Burström and Berg2023; Haslbeck et al., Reference Haslbeck, Jover Martínez, Roefs, Fried, Lemmens, Groot and Edelsbrunner2024; Kuhlmann et al., Reference Kuhlmann, Dantlgraber and Reips2017; Sung & Wu, Reference Sung and Wu2018).

Figure 1 Likert scale and VAS.

Psychometric research on VASs, however, has not kept pace with this increased usage in applied settings, and standard psychometric approaches to evaluate important aspects of VAS data quality are lacking. The present study aims to begin filling this gap. We do so by focusing on the detection of responding due to careless or insufficient effort—i.e., providing responses without investing effort into carefully evaluating the administered items and leaving data contaminated with responses that do not reflect what researchers want to measure (Huang et al., Reference Huang, Liu and Bowling2015)—as a crucial aspect of ensuring data quality. To provide an approach that is tailored to detecting careless responding in VAS data, we integrate existing measurement models for VASs (Noel & Dauvier, Reference Noel and Dauvier2007) with mixture item response theory (IRT) or factor models for identifying and modeling careless responding. Such developments have previously been limited to Likert-type scales (Uglanova et al., Reference Uglanova, Nagy and Ulitzsch2025; Ulitzsch, Pohl, et al., Reference Ulitzsch, Pohl, Khorramdel, Kroehne and von Davier2022; Ulitzsch, Yildirim-Erbasli, et al., Reference Ulitzsch, Yildirim-Erbasli, Gorgun and Bulut2022; Ulitzsch, Pohl, et al., Reference Ulitzsch, Pohl, Khorramdel, Kroehne and von Davier2024; van Laar & Braeken, Reference van Laar and Braeken2022) and unbounded continuous response data (e.g., normally distributed with infinite range) (Arias et al., Reference Arias, Garrido, Jenaro, Martinez-Molina and Arias2020; Kam & Cheung, Reference Kam and Cheung2023; Zhang et al., Reference Zhang, Ulitzsch and Domingue2025). Our proposed mixture modeling approach not only facilitates gauging VAS data quality and accounting for careless responding when drawing conclusions on constructs measured with VASs but also opens the path to comparative analyses of the occurrence of careless responding across different scale formats.

The remainder of the article is structured as follows: First, we briefly discuss VASs and their usage in psychological, social, and clinical science research. Second, we review existing approaches to the detection of careless responding, which, so far, have predominantly been targeted to Likert-type scales or continuous, unbounded data. To expand these previous approaches to identifying careless responding in VAS settings, we combine them with existing measurement models for VASs, which are reviewed next. After presenting the proposed model, we evaluate its parameter recovery under realistic research settings in a simulation study. We also demonstrate through a simulation study why mixture modeling designed for unbounded continuous data can be unsuitable for VAS data. In an empirical application to data administering the same scales with VAS and Likert-type scale formats, we further illustrate the real-world application of the proposed model and provide initial insights into differences in careless responding across these different scale types.

1.1 VASs

VASs provide a straightforward method for respondents to indicate their position along a continuum. Typically, a VAS includes a line with two anchors at either end. These anchors usually feature verbal descriptors representing opposite ends of a semantic dimension (e.g., “strongly agree” and “strongly disagree,” Figure 1). However, anchors can also be visual or auditory, including pictures or sound clips. For example, visual cues, such as smiley faces, can be used to help participants, such as young children, who may not fully understand verbal descriptions, indicate their feelings (Reips & Funke, Reference Reips and Funke2008).

Several advantages of VASs are discussed in the literature. One key benefit is their high sensitivity to variations in the latent construct, allowing for the detection of subtle shifts in perceptions or emotional states, which may be missed by Likert-type scales (Price et al., Reference Price, Bush, Long and Harkins1994). Another advantage is that, unlike Likert-type scales—whose psychometric properties can be heavily influenced by the number of response categories and for which selecting an appropriate number is not straightforward—VASs entirely avoid such issues (see, e.g., Kutscher & Eid, Reference Kutscher and Eid2024, for a discussion). The unique characteristics of VASs make it a valuable tool in both academic and clinical settings, where precise measurement of continuous variables is essential. In clinical research, for instance, VAS has proven effective for assessing patient pain levels (McCormack et al., Reference McCormack, Horne and Sheather1988; Myles et al., Reference Myles, Myles, Galagher, Boyd, Chew, MacDonald and Dennis2017; Price et al., Reference Price, Bush, Long and Harkins1994). Similarly, in psychological research, VASs have been widely applied to measure constructs of interest, such as emotional states (Askim & Knardahl, Reference Askim and Knardahl2021; Zhou & Chen, Reference Zhou and Chen2009), self-esteem (Brumfitt & Sheeran, Reference Brumfitt and Sheeran1999), and quality of life (Weigl & Forstner, Reference Weigl and Forstner2021).

Nevertheless, VASs may also be more demanding to use and could lead to quicker disengagement (Haslbeck et al., Reference Haslbeck, Jover Martínez, Roefs, Fried, Lemmens, Groot and Edelsbrunner2024), potentially affecting data quality as a result. Comparing Likert-type scales and VASs in a children’s sample, Van Laerhoven et al. (Reference Van Laerhoven, van der Zaag-Loonen and Derkx2004), for instance, found higher rates of missing responses in VAS data as well as self-reported preferences and higher ease of use for the Likert-type scale. Thorough comparisons of whether or not VASs and Likert-type scales elicit respondent disengagement to a different extent and, as such, provide data of different quality, however, are lacking. In this study, we provide a tool that facilitates filling this gap.

1.2 Careless responding and its detection in Likert-type scales

In collecting and investigating survey data, the detection of careless respondents is crucial to ensure the trustworthiness of conclusions drawn from self-report data (Meade & Craig, Reference Meade and Craig2012). Careless and insufficient effort responding occurs when participants fail to engage with the content of the survey questions properly, e.g., by not reading them carefully, rushing through responses, or being distracted (Meade & Craig, Reference Meade and Craig2012). This phenomenon is widespread in research relying on self-report data, with studies indicating that careless response rates can range from 1% up to 50% (Douglas et al., Reference Douglas, Ewell and Brauer2023; Ward & Meade, Reference Ward and Meade2023). Such careless responses are problematic because they do not accurately reflect the latent variables intended to be measured. Consequently, careless and insufficient effort can heavily distort results, potentially diminishing estimates of reliability and construct validity, comprising factor analysis outcomes, and leading to erroneous conclusions on relationships among constructs of interest (Ward & Meade, Reference Ward and Meade2023).

Previous developments in careless respondent detection have predominantly focused on Likert scale data, most commonly drawing on response-pattern-based indicators like the long string index and Mahalanobis distance. The long string index flags excessive uniformity by measuring the number of consecutive identical responses, while Mahalanobis distance detects unusual response patterns conceptualized as multivariate outliers. The application of such indicators to VAS data, however, is not always straightforward. For instance, while Mahalanobis distance could be adaptable for use with VAS data, using the long string index poses challenges. Since VAS responses are continuous rather than discrete, defining “identical” responses is difficult, and small fluctuations in consecutive values can obscure the detection of inattentive patterns.

Alongside these indicator-based approaches, confirmatory mixture IRT or factor models have recently been proposed to detect inattentive or hasty responses by analyzing item responses and, if available, collateral information on respondent behavior as contained in response times (Arias et al., Reference Arias, Garrido, Jenaro, Martinez-Molina and Arias2020; Kam & Cheung, Reference Kam and Cheung2023; Ulitzsch, Pohl, et al., Reference Ulitzsch, Pohl, Khorramdel, Kroehne and von Davier2022; Ulitzsch, Yildirim-Erbasli, et al., Reference Ulitzsch, Yildirim-Erbasli, Gorgun and Bulut2022; van Laar & Braeken, Reference van Laar and Braeken2022; Zhang et al., Reference Zhang, Ulitzsch and Domingue2025). The overarching principle of these confirmatory mixture modeling approaches to careless respondent detection is the translation of theoretical considerations on respondent behavior into two mixture component models—one representing an assumed attentive and the other one an assumed careless data-generating process. Attentive item responses are assumed to reflect the to-be-measured traits, i.e., to follow standard IRT or confirmatory factor analysis (CFA) measurement models (Arias et al., Reference Arias, Garrido, Jenaro, Martinez-Molina and Arias2020; Roman et al., Reference Roman, Schmidt, Miller and Brandt2024; Ulitzsch, Pohl, et al., Reference Ulitzsch, Pohl, Khorramdel, Kroehne and von Davier2022; Ulitzsch, Yildirim-Erbasli, et al., Reference Ulitzsch, Yildirim-Erbasli, Gorgun and Bulut2022; Ulitzsch, Pohl, et al., Reference Ulitzsch, Pohl, Khorramdel, Kroehne and von Davier2024). Careless item responses, in contrast, are assumed to be driven by respondents’ category preferences (Arias et al., Reference Arias, Garrido, Jenaro, Martinez-Molina and Arias2020) or to be random (van Laar & Braeken, Reference van Laar and Braeken2022). These models have proven effective in accurately identifying careless respondents and mitigating biases in survey results, thereby enhancing the reliability of the collected data and ensuring the validity of conclusions drawn from them. However, factor mixture models suggested for identifying careless respondents in continuous data (Kam & Cheung, Reference Kam and Cheung2023; Roman et al., Reference Roman, Schmidt, Miller and Brandt2024; Zhang et al., Reference Zhang, Ulitzsch and Domingue2025) are not directly applicable to VAS data due to their unique continuous and bounded nature (Noel & Dauvier, Reference Noel and Dauvier2007). A key challenge is that bounded responses often display a skewed distribution (Noel & Dauvier, Reference Noel and Dauvier2007; Verkuilen & Smithson, Reference Verkuilen and Smithson2012), rendering the normal distribution assumptions in CFA measurement models inappropriate for modeling. This issue may be further exacerbated by careless responding, which may, as illustrated below, result in distributions that heavily deviate from normality. The aim of the present study is, therefore, to provide a confirmatory mixture modeling approach to careless respondent detection that is tailored to VAS data. To this end, we expand the Beta item response model (IRM) proposed by Noel & Dauvier (Reference Noel and Dauvier2007) by a mixture component for flexibly absorbing careless response patterns on VASs.

1.3 The Beta IRM (Noel & Dauvier, Reference Noel and Dauvier2007)

Data from the VASs is typically converted into numerical values ranging from 0 to 1, representing the percentage of the distance from one endpoint to the selected position. While this data is continuous, its bounded nature poses challenges for traditional models such as CFA. To address this, Noel & Dauvier (Reference Noel and Dauvier2007) introduced the Beta IRM. This model draws on the Beta distribution (Ferrari & Cribari-Neto, Reference Ferrari and Cribari-Neto2004) to analyze continuous bounded responses. The Beta distribution is defined on the interval (0, 1), making it suitable for modeling data that is similarly bounded.

The Beta IRM defines the response variable $X_{ij}$ of respondent $i \in \{1,\ldots ,N\}$ provided to item ${j \in \{1,\ldots ,J\}}$ as a Beta-distributed random variable, $X_{ij} \sim \operatorname {Beta}(\alpha _{ij}, \beta _{ij})$ , where the parameters $\alpha _{ij}$ and $\beta _{ij}$ might be interpreted as “acceptance” and “refusal” parameters. The “acceptance” parameter $\alpha _{ij}$ signifies the inclination toward the higher end of the scale (closer to 1), while the $\beta _{ij}$ parameter signifies the inclination toward the lower end of the scale (closer to 0). In the multidimensional extension for measuring dimension $d \in \{1, \ldots , D\}$ , assuming a simple structure, the parameters $\alpha _{i j}$ and $\beta _{i j}$ are modeled as functions of the latent trait $\theta _{id[j]}$ , which represents respondent i ’s location on the dth ( $d \in \{1,\ldots ,D\}$ ) dimension measured by item j, as well as the item wording $w_j$ , item difficulty $\delta _j$ , and item dispersion $\tau _j$ :

(1) $$ \begin{align} \begin{aligned} & \alpha_{i j}=\exp \left(\frac{w_j\theta_{id[j]}-\delta_j+\tau_j}{2}\right) \\ & \beta_{i j}=\exp \left(\frac{-(w_j\theta_{id[j]}-\delta_j)+\tau_j}{2}\right). \end{aligned} \end{align} $$

Note that the function $d[j]$ maps item j to the latent dimension d it is designed to assess. The item wording $w_j$ is pre-specified, where $w_j=1$ indicates that item j is positively worded, and $w_j=-1$ denotes a negatively worded item. The remaining item parameters, $\delta _j$ and $\tau _j$ , are freely estimated. The item dispersion parameter $\tau _j$ governs the response variability, with higher values indicating lower variance, conditional on the latent trait and item difficulty. For model identification, the means of the latent traits are set to zero.

The expected value of the response variable $X_{ij}$ and the probability density function of its realization $x_{ij}$ are then:

(2) $$ \begin{align} \mathbb{E}\left(X_{i j}\right)=\frac{\alpha_{i j}}{\alpha_{i j}+\beta_{i j}}= \frac{\exp \left(\frac{w_j \theta_{id[j]}-\delta_j + \tau_j}{2}\right)}{\exp \left(\frac{w_j \theta_{id[j]}-\delta_j + \tau_j}{2}\right)+\exp \left(\frac{-(w_j \theta_{id[j]}-\delta_j) + \tau_j}{2}\right)}=\frac{1}{1+\exp \left(-\left(w_j \theta_{id[j]}-\delta_j\right) \right)} \end{align} $$
(3) $$ \begin{align} f(x_{ij} | \theta_{id[j]}, w_j, \delta_j, \tau_j) = x_{ij}^{(\alpha_{ij} -1)} (1-x_{ij})^{(\beta_{ij} -1)} \frac{\Gamma\left(\alpha_{i j}+\beta_{i j}\right)}{\Gamma\left(\alpha_{i j}\right) \Gamma\left(\beta_{i j}\right)}. \end{align} $$

Note that $\frac {\Gamma \left (\alpha _{i j}+\beta _{i j}\right )}{\Gamma \left (\alpha _{i j}\right ) \Gamma \left (\beta _{i j}\right )}$ is the normalizing constant of the Beta distribution, ensuring the probability density integrates to 1 over the interval (0, 1).

The logistic form of the expected response function shows the familiar S-shaped curve, which is commonly used in IRT to model the probability of item endorsement as a function of person and item parameters. To illustrate the item parameters, we focus on the condition with a unidimensional latent factor ( $D=1$ ) and positively worded items ( $\lambda _j=1$ ) to demonstrate how variations in item difficulty $\delta _j$ and dispersion $\tau _j$ influence the expected response and item information. Figure 2a demonstrates the role of the item difficulty parameter, which determines the location of the item characteristic curve (ICC) along the scale of the latent trait. The $\tau _j$ parameter represents the dispersion or precision of the item. It does not influence the expected responses (Eq 2); however, it enables different items to exhibit different response variabilities around the expected value. A higher $\tau _j$ value results in a more peaked response density for a given level of $\theta _i - \delta _j$ , as well as a more peaked information function which indicates that the item provides more information (precision) at certain trait levels (Figure 2b and c. Details regarding the Fisher information function of the Beta IRM can be found in Noel & Dauvier (Reference Noel and Dauvier2007)).

Figure 2 Beta IRM.

Note: (a) ICC for different difficulty parameters, with dispersion fixed at 0. (b) ICC with response densities for $\theta _i - \delta _j = 0$ under different dispersion parameters. (c) Item Information Function for different dispersion parameters, with difficulty fixed at 0.

The Beta IRM provides several advantages. First, it accurately captures the bounded nature of VAS data. VAS data are sometimes analyzed with traditional factor analysis with normality assumptions (Wortmann et al., Reference Wortmann, Gisch and Warschburger2021). These, however, may not hold for bounded data, especially when responses are skewed or clustered near the boundaries. The Beta IRM avoids this limitation, providing a more flexible and accurate framework for analyzing such data. Second, the model offers a straightforward interpretation of the parameters, as they directly relate to the person’s trait location and item difficulty, similar to traditional IRT models. Third, the logistic form of the expected response function allows for easy comparison with existing IRT models, facilitating a seamless integration of the model into existing psychometric frameworks. Overall, the Beta IRM represents a significant advancement in the analysis of VAS data, offering a theoretically sound and practically viable solution for handling the unique challenges posed by continuous bounded responses (Noel & Dauvier, Reference Noel and Dauvier2007).

2 The proposed model

Building upon the Beta IRM (Noel & Dauvier, Reference Noel and Dauvier2007), we propose an extended Beta Mixture IRM to identify careless respondents in VAS data. Adapting concepts from previous confirmatory mixture models (Arias et al., Reference Arias, Garrido, Jenaro, Martinez-Molina and Arias2020; Kam & Cheung, Reference Kam and Cheung2023; Ulitzsch, Pohl, et al., Reference Ulitzsch, Pohl, Khorramdel, Kroehne and von Davier2022; Ulitzsch, Yildirim-Erbasli, et al., Reference Ulitzsch, Yildirim-Erbasli, Gorgun and Bulut2022; van Laar & Braeken, Reference van Laar and Braeken2022), we classify respondents into attentive ( $Z_i = 1$ ) and careless ( $Z_i = 0$ ) groups based on their response patterns’ alignment with measurement models formulated to capture different types of respondent behavior. The proposed model is expressed as follows (Figure 3):

(4) $$ \begin{align} \begin{aligned} f(x_{ij}) & = \pi_i \cdot f(x_{ij} \mid z_i = 1) + (1 - \pi_i) \cdot f(x_{ij} \mid z_i = 0) \\ X_{ij} \mid Z_i &= 1 \sim \operatorname{Beta}\left(\exp \left(\frac{w_j\theta_{id[j]} - \delta_j + \tau_j}{2}\right), \exp \left(\frac{-\left(w_j\theta_{id[j]} - \delta_j\right) + \tau_j}{2}\right)\right) \\ X_{ij} \mid Z_i &= 0 \sim \operatorname{Beta}(m, n), \\ \end{aligned} \end{align} $$

where $f(\cdot )$ denotes a general probability density function, and $x_{ij}$ represents the realization of the continuous bounded response from respondent i to item j. The latent variable $Z_i$ indicates class membership, where $Z_i=1$ denotes the attentive class, and $Z_i=0$ denotes the careless class, and $\pi _i=\Pr (Z_i=1)$ represents the probability that individual i belongs to the attentive class. The response distribution is a mixture of two components. The first component, $f\left (x_{i j} \mid z_i=1\right )$ , models responses for attentive individuals as a Beta distribution governed by the person’s latent trait levels $\theta _{id[j]}$ , item wording $w_j$ , item difficulty level $\delta _j$ , and the dispersion parameter $\tau _j$ . To facilitate disentangling careless from attentive responding, we recommend modeling raw responses obtained from both positively and negatively worded items (as in Kam & Cheung, Reference Kam and Cheung2023; Ulitzsch, Yildirim-Erbasli, et al., Reference Ulitzsch, Yildirim-Erbasli, Gorgun and Bulut2022; Vogelsmeier et al., Reference Vogelsmeier, Uglanova, Rein and Ulitzsch2024). Information on item wording is incorporated into the model through the parameter $w_j$ , which takes the value 1 or -1.

Figure 3 Model structure.

Note: For attentive respondents, responses are assumed to follow a Beta IRM, governed by item and person parameters; a two-factor, six-item structure is used as an example. For careless respondents, responses are assumed to stem from a common Beta distribution, with Beta(2, 5) used as an illustrative example in the figure.

The second component, $f\left (x_{i j} \mid z_i=0\right )$ , models responses for careless individuals using a Beta distribution with common parameters $\alpha _{ij}=m$ and $\beta _{ij}=n$ that generalize across all items and persons. This reflects the assumption that careless respondents do not process item content, and that, hence, the selected responses should neither be governed by respondent’s trait levels nor impacted by item characteristics.

The Beta distribution of the careless component is assumed to “absorb” different careless response patterns; hence, its parameters will reflect the marginal distribution over all types of careless response patterns in the population (see Ulitzsch, Pohl, et al., Reference Ulitzsch, Pohl, Khorramdel, Kroehne and von Davier2022), approximated by a Beta distribution. The unstructured Beta distribution assumed for the careless class offers a flexible means of capturing the marginal distribution of diverse careless behaviors in the population. For example, a $\text {Beta}(1,1)$ distribution suggests marginally uniform selection across the scale, $\text {Beta}(5, 5)$ indicates a marginal tendency to select around the midpoint, and $\text {Beta}(0.5, 0.5)$ reflects marginal preferences for the opposite ends of the slider (Figure 4).

Figure 4 Density plot for different Beta distributions.

We acknowledge that careless behaviors may take various forms and may differ across respondents (e.g., some may position their slider randomly, while others may move their slider toward one end of the scale). As such, the careless component model will most likely be misspecified. However, since the particular type(s) of careless responding displayed are typically not of substantive interest, we do not aim to disentangle possibly different subtypes of careless behavior or to interpret the internal structure of the careless class. Instead, the careless component of the mixture model is conceptualized as a residual class with minimal structure, designed to capture any response pattern that systematically deviates from attentive behavior.

For model estimation, we suggest Bayesian estimation due to its flexibility in handling complex models (Gelman et al., Reference Gelman, Carlin, Stern, Dunson, Vehtari and Rubin2013), with priors assigned to model parameters. These prior values can be informed by theoretical considerations and previous research, or, when prior knowledge is unavailable, can be set as diffuse. In this study, we apply diffuse priors. For the item-specific difficulty and dispersion parameters, $\delta _j$ and $\tau _j$ , we use normal priors with $\mathcal {N}(0,10)$ . The latent trait mean for the attentive group is fixed at zero for model identification. The multidimensional latent traits $\boldsymbol {\theta }_i$ are modeled as $\mathcal {N}_D(\mathbf {0}, \boldsymbol {\Sigma }_\theta )$ , where the means are fixed to zero for model identification. The covariance matrix $\boldsymbol {\Sigma }_{\theta }$ is decomposed into a correlation matrix $\boldsymbol {R}$ and a vector of standard deviations $\boldsymbol {\sigma }_{\theta }$ , such that $\boldsymbol {\Sigma }_{\theta } = \text {diag}(\boldsymbol {\sigma }_{\theta })\boldsymbol {R}\text {diag}(\boldsymbol {\sigma }_{\theta })$ . We place a half-Cauchy prior on each dimension of $\boldsymbol {\sigma }_\theta $ , i.e., $\sigma _{\theta ,d} \sim \text {Cauchy}^+(0, 5)$ , and a Lewandowski–Kurowicka–Joe (LKJ, Lewandowski et al., Reference Lewandowski, Kurowicka and Joe2009) prior on the correlation matrix: $\boldsymbol {R} \sim \text {LKJ}(1)$ . When $D = 1$ , the prior for latent variable simplifies to $\theta _i \sim \mathcal {N}(0, \sigma _\theta ^2)$ without correlation structure. For the parameters m and n, which shape the Beta distribution of careless response patterns, we employ a half-Cauchy distribution $\text {Cauchy}^+(0, 5)$ for accommodating a wide range of values. For the individual attentiveness and carelessness probabilities $\pi _i$ and $1-\pi _i$ , we employ a Dirichlet prior parameterized as $(\pi _i,1-\pi _i)\sim \text {Dirichlet}(\nu (\pi _{\mathcal{P}}, 1-\pi _{\mathcal{P}}))$ , where $\pi _{\mathcal{P}}$ represents the population-level proportion of attentive respondents and $\nu $ is a concentration parameter (Kemp et al., Reference Kemp, Perfors and Tenenbaum2007; Salakhutdinov et al., Reference Salakhutdinov, Tenenbaum and Torralba2012). For $\pi _{\mathcal{P}}$ and its counterpart $1-\pi _{\mathcal{P}}$ , we employ a diffuse Dirichlet prior $(\pi _{\mathcal{P} },1-\pi _{\mathcal{P} })\sim \text {Dirichlet}(1, 1)$ , implying a uniform distribution. For the concentration parameter $\nu $ , a half-Cauchy prior with location 0 and scale 5 is used.

3 Simulation study 1: Performance of the Beta mixture model

The aim of the simulation study was to evaluate the performance of the proposed model in two aspects: its accuracy in detecting careless respondents exhibiting different behavioral patterns and its estimation accuracy of model parameters of the attentive response model that are adjusted for the occurrence of careless responding.

3.1 Data generation and model estimation

We simulated data for a fixed sample size of 300 as it reflects the typical requirement in latent variable analysis (MacCallum et al., Reference MacCallum, Widaman, Zhang and Hong1999). Empirically, this corresponds to the 23rd percentile of sample sizes among the 859 datasets in the Item Response Warehouse (Domingue et al., Reference Domingue, Braginsky, Caffrey-Maffei, Gilbert, Kanopka, Kapoor, Lee, Liu, Nadela, Pan, Zhang, Zhang and Frank2025), representing a realistic lower-bound sample size in applied research. The number of items was set at 10, half of which being negatively worded ( $w_j=-1$ ). This approach enables us to rigorously evaluate the model’s performance under a limited sample size, with the anticipation that as the sample size increases in real-world scenarios, the performance of the mixture model will improve (Zhang et al., Reference Zhang, Ulitzsch and Domingue2025). For generating the attentive responses, difficulty parameters and latent factor scores were drawn from a standard normal distribution $\mathcal {N}(0,1)$ , and dispersion parameters were randomly sampled from a uniform distribution within the [0,3] range following Noel & Dauvier (Reference Noel and Dauvier2007). It is important to note that attentive responses were generated using a unidimensional model, posing a particular challenge for detection. Careless responding is generally easier to identify in multidimensional scales (Ulitzsch, Pohl, et al., Reference Ulitzsch, Pohl, Khorramdel, Kroehne and von Davier2022), where attentive response patterns tend to exhibit specific structures—such as weaker correlations between items from different scales—that help distinguish them more clearly from careless patterns.

We considered three patterns of careless response behaviors in data generation:

  1. 1) Random response pattern at extremes: Responses clustered at the two extremes of the scale, with values randomly selected near either end, following a $\text {Beta}(0.5, 0.5)$ distribution (Figure 4).

  2. 2) Overly consistent response pattern: Careless respondents consistently selected values near one end of the slider scale—either the left or the right—regardless of item wording. Participants were randomly assigned to one of these two groups, with responses generated from either $\text {unif}(0, 0.2)$ for those favoring the left end or $\text {unif}(0.8, 1)$ for those favoring the right end.

  3. 3) Random response pattern at midpoint: Responses clustered around the midpoint of 0.5, generated from a truncated normal distribution with a mean of 0.5, a standard deviation of 0.25, and bounds set between 0 and 1.

Note that only the first pattern is generated from the Beta distribution. In the latter two data-generating conditions, the Beta distribution can only approximate but not perfectly represent the careless response distribution, presenting a realistic yet potentially challenging scenario. This is especially true for Scenario 2, which (a) simulates a mixture of distinct response preferences, thereby showcasing the model’s capability to “absorb” a mixture of different careless response patterns, and (b) introduces sharp endpoints in both contributing distributions, which are poorly captured by common Beta distributions. While these sharp endpoints may not be entirely realistic—since respondents are likely not able to move their slider uniformly within a precisely defined sub-segment of the scale—we included this scenario to explore the extent to which the careless component of our model can tolerate deviations from a Beta distribution.

Additionally, we varied the proportion of careless respondents ( $1-\pi _{\mathcal{P}}$ ). We included four low-proportion conditions (0.05, 0.10, 0.15, and 0.25), following prior research on careless response detection, as these levels reflect commonly observed prevalence rates (Kam & Cheung, Reference Kam and Cheung2023). To evaluate the model’s performance under more extreme conditions, we also included a higher proportion condition (0.40), where the number of careless respondents approaches that of attentive ones—resulting in five proportion levels in total. For each condition, we generated 100 datasets for replication.Footnote 1

Data generation was conducted using R version 4.3.1 (R Core Team, 2023). Each data set was analyzed with the proposed model using Stan (Carpenter et al., Reference Carpenter, Gelman, Hoffman, Lee, Goodrich, Betancourt, Brubaker, Guo, Li and Riddell2017). For model estimation, we ran two Markov chain Monte Carlo (MCMC) chains with 40,000 iterations each, with the first half being employed as burn-in. Model convergence was assessed using the estimated potential scale reduction (EPSR) index (Gelman, Reference Gelman1996), with values below 1.1 indicating satisfactory convergence (Gelman et al., Reference Gelman, Carlin, Stern, Dunson, Vehtari and Rubin2013). We evaluated the model performance in terms of estimation and classification accuracy. For model estimation, we saved the posterior means as point estimates. We evaluated the correlation between the true parameter values and the estimated values for parameters generated from random distributions ( $\delta _j, \tau _j$ ). We considered the relative bias (i.e., ratio of the difference between the estimated parameter and its true value to the true value) for fixed parameters in each condition ( $m, n, \pi _{\mathcal{P}}$ ). Further, the root mean square error (RMSE) was calculated for all parameters to evaluate the accuracy of the parameter estimates.

To assess the classification performance of the proposed model, each respondent was classified based on their individual probability of attentiveness ( $\pi _i$ ). We evaluated two methods of classification: 1) a threshold-based classification approach, where $\pi _i < 0.5$ ( $\pi _i < 1-\pi _i$ , suggesting the probability of attentiveness is lower than the probability of carelessness) indicates that person i is careless and 2) a ranking approach using the population-level proportion of attentive respondents ( $\pi _{\mathcal{P}}$ ). Here, we first obtained the $\pi _{\mathcal{P}}$ point estimate, then ranked $\pi _i$ values, and identified the least likely $(1-\pi _{\mathcal{P}}) \times N$ respondents as careless based on their individual probabilities. This approach allows the classification to reflect the estimate of overall attentiveness. We then evaluated the model’s classification performance using several key metrics: 1) Accuracy, defined as the proportion of respondents correctly classified as either attentive or careless; 2) Sensitivity, defined as the ability of the model to correctly identify attentive respondents, calculated as the proportion of actual attentive respondents correctly classified; 3) Precision, defined as the proportion of respondents classified as attentive who were truly attentive; 4) False positive rate (FPR), defined as the proportion of careless respondents incorrectly classified as attentive; as well as 5) False negative rate (FNR), defined as the proportion of attentive respondents mistakenly classified as careless.

3.2 Results

The model exhibited high convergence rates for all conditions ( $\geq 99\%$ ), except when the careless response pattern was overly consistent. In the overly consistent condition, convergence became increasingly difficult as the proportion of careless respondents rose. Convergence rates exceeded 89% when $1 - \pi _{\mathcal{P}} = 0.05$ to $0.15$ , but declined to 86% and 68% at $1 - \pi _{\mathcal{P}} = 0.25$ and $0.40$ , respectively. This might be due to the fact that the simulated overly consistent careless responses could not be well accommodated by a Beta distribution. To address convergence issues, we increased the number of MCMC iterations for the overly consistent condition with 40% careless respondents. The convergence rate substantially improved—from 68% to 97%—when the number of burn-in iterations was allowed to go up to 100,000. These results suggest that large proportions of careless responses following distributions that challenge the model also pose challenges for model convergence; however, this issue can be mitigated by increasing the number of iterations.

Estimation and classification accuracy were evaluated based on the converged replications. Tables 1 and 2 summarize the classification accuracy based on the threshold 0.5 and the population-level proportion of attentive respondents, respectively. The classification results showed excellent overall performance, with both the threshold-based and population-level proportion methods achieving high accuracy in classifying respondents as attentive or careless. However, when using the threshold method, the FPR reached unacceptably high levels when the proportion of careless respondents ( $1-\pi _{\mathcal{P}}$ ) was between 0.05 and 0.15. This suggests that when $1-\pi _{\mathcal{P}}$ is low, for most careless respondents, the estimated $\pi _i$ values exceeded 0.5, leading to their misclassification as attentive. Although overall accuracy remained high with consistently strong sensitivity and precision, this is likely due to class imbalance where the majority of respondents are attentive. Moreover, the FPR varied across data-generating scenarios, being low in overly consistent conditions, but higher in random and normal conditions. Under these conditions, the FPR reached as high as 0.477 and 0.456, respectively, when $1-\pi _{\mathcal{P}} = 0.05$ .

Table 1 Classification results based on the threshold

Note: Beta = Random responses generated from $\text {Beta}(0.5, 0.5)$ ; Preference = Overly consistent pattern; Normal = Random responses generated from a truncated normal distribution; The term $1-\pi _{\mathcal {P}}$ indicates the true population-level proportion of careless respondents; FPR = False positive rate; FNR = False negative rate.

Table 2 Classification results based on $\pi _{\mathcal {P}}$

Note: Beta = Random responses generated from $\text {Beta}(0.5, 0.5)$ ; Preference = Overly consistent pattern; Normal = Random responses generated from a truncated normal distribution; The term $1-\pi _{\mathcal {P}}$ indicates the true population-level proportion of careless respondents; FPR = False positive rate; FNR = False negative rate.

The high FPR observed under the single-threshold (0.5) approach may be attributed to the hierarchical prior structure and the high proportions of attentive respondents. When the two classes are more balanced (e.g., $\pi _{\mathcal{P}} = 0.4$ ), the FPR is generally within an acceptable range. In contrast, under imbalanced conditions where $\pi _{\mathcal{P}}$ is high, the hierarchical prior structure—with $\nu (\pi _{\mathcal{P}}, 1-\pi _{\mathcal{P}})$ serving as a hyperprior for $(\pi _i, 1 - \pi _i)$ —pulls individual-level estimates $\pi _i$ toward higher values (i.e., attentiveness), thereby diminishing the effectiveness of the 0.5 threshold in detecting careless respondents.

The classification method utilizing the population-level proportion ( $\pi _{\mathcal {P}}$ ) exhibited a slightly lower overall accuracy. Nevertheless, it maintained a high level of performance, especially when the proportion of careless respondents was low. Moreover, it provided significantly improved management of the FPR, with FPRs consistently remaining below 0.05 across most conditions, except in several cases where careless responses deviated from the Beta distribution, in which FPRs remained under 0.1. This method also maintained high sensitivity and precision, with sensitivity values ranging from 0.919 to 0.986 and precision consistently above 0.95. Given the ability of the population-level proportion method to better control the FPR while maintaining strong classification performance, we recommend using this approach when the proportion of careless respondents is low.

Table 3 displays results related to the estimation accuracy of the attentive item parameters, the standard deviation of the content trait, and the population-level proportion of attentive respondents. The attentive item parameters (difficulty $\delta _j$ and response variability $\tau _j$ ) were accurately estimated (e.g., high correlations between true and estimated values and small RMSEs). Likewise, the standard deviation of latent trait ( $\sigma _\theta $ ) was accurately estimated across all conditions, exhibiting very low relative bias and RMSE. To more closely evaluate latent trait estimation, we further examined the factor score estimates for attentive respondents under one selected condition ( $1 - \pi _{\mathcal {P}} = 0.15$ and careless responses were generated from $\text {Beta}(0.5, 0.5)$ ). We compared the proposed model with the Beta IRM, which does not account for careless responding. Results showed that both models achieved high correlations between estimated and true factor scores, with average correlations across replications of 0.959 for both the proposed model and the Beta IRM. However, the proposed model yielded lower RMSE values (0.288 versus 0.346), indicating it produces more precise estimates with less variability around the true factor scores.

Table 3 Estimation accuracy of item parameters and $\pi _{\mathcal {P}}$

Note: Beta = Random responses generated from $\text {Beta}(0.5, 0.5)$ ; Preference = Overly consistent pattern; Normal = Random responses generated from a truncated normal distribution; The term $1-\pi _{\mathcal {P}}$ indicates the true population-level proportion of careless respondents; RB = Relative bias; RMSE = Root mean square error.

The accuracy of the estimated population-level proportion of attentive respondents ( $\pi _{\mathcal {P}}$ ) depended on the data-generating distribution of careless responses. While the relative bias was essentially zero when careless responses were generated from a Beta distribution, we observed estimates to be downward biased in the other two conditions. That is, under conditions where careless responses cannot be perfectly accommodated by a Beta distribution, researchers might underestimate the proportion of attentive respondents—and, consequently, overestimate the proportion of careless respondents. For instance, when careless responses were generated with an overly consistent pattern, the average estimated proportion of careless respondents was 0.075 and 0.186 in conditions with data-generating proportions of 0.05 and 0.15, respectively. However, we note that the amount of bias was small, such that we deem the model to be sufficiently reliable to roughly gauge the extent of careless respondents in these scenarios. The estimation results of $\pi _{\mathcal {P}}$ also help explain the variation in classification accuracy across three careless response patterns (Table 2). Classification accuracy based on the population-level cutoff ( $\pi _{\mathcal {P}}$ ) was highest when careless responses were generated from a Beta distribution, likely due to the fact that $\pi _{\mathcal {P}}$ was estimated most accurately under this condition.

Figure 5, comparing data-generating distributions of careless responses and how these were approximated by the proposed model, provides intuition for the sources of variation in estimation accuracy across conditions. The parameter recovery results of these distributional patterns are also presented in Table 4. When careless responses were generated from a Beta distribution (Figure 5a), the distribution of careless responses was well recovered, as was to be expected. The parameter estimates of the Beta distributions closely matched the true values, with low RMSE indicating high estimation accuracy. Note that RMSE values for the shape parameters m and n could only be computed for the Beta condition, as these parameters are not defined under alternative generating distributions. When careless responses were not generated from a Beta distribution, the proposed model could only approximate but not fully replicate the true distribution of careless responses. For overly consistent responses generated from unif(0, 0.2) or unif(0.8, 1), the estimated distributions typically showed a U-shaped pattern, characterized by underestimated densities around the midpoint and overestimated densities near the scale boundaries (Figure 5b). The parameters of the Beta-approximated distribution showed great variability, as indicated by the large standard deviations in the parameter estimates. Moreover, when data were generated using a truncated normal distribution, the estimated distributions exhibited lighter tails compared to the true distribution (Figure 5c). The Beta approximation in this case resembled a $\text {Beta}(2.5, 2.5)$ distribution—symmetric around 0.5 and bell-shaped, bearing similarity to a normal distribution.

Figure 5 Beta approximation for careless responses.

Note: Beta = Random responses generated from $\text {Beta}(0.5, 0.5)$ ; Preference = Overly consistent pattern; Normal = Random responses generated from a truncated normal distribution. Blue lines indicate the Beta approximation for careless responses across all simulated datasets, while red dashed lines represent the distribution used for data generation.

Table 4 Parameter estimation for the careless response distributions

Note: Beta = Random responses generated from $\text {Beta}(0.5, 0.5)$ ; Preference = Overly consistent pattern; Normal = Random responses generated from a truncated normal distribution; The term $1-\pi _{\mathcal {P}}$ indicates the true population-level proportion of careless respondents; Mean = Averaged estimates across all converged replications; SD = Standard deviation of these estimates; RMSE is reported only for the Beta condition, where the true values of the Beta distribution are known.

4 Simulation study 2: Comparison with the mixture CFA model

As mentioned in the introduction, the unique characteristics of VAS data render previous mixture modeling approaches for careless respondent detection unsuitable, as their underlying assumptions (e.g., normality) may not hold for VAS data. To further demonstrate this, we conducted a simulation study to evaluate the limitations of a mixture CFA model for VAS data. Using the simulated datasets from Simulation Study 1, we applied a mixture CFA model (Roman et al., Reference Roman, Schmidt, Miller and Brandt2024; Zhang et al., Reference Zhang, Ulitzsch and Domingue2025) to the data:

(5) $$ \begin{align} \begin{aligned}f(x_{ij}) & = p_i \cdot \phi(x_{ij}|z_i=1) + (1 - p_i) \cdot \phi(x_{ij}|z_i=0) \\ X_{ij}|Z_i&=1 \sim \mathcal{N}(\mu_{j} + \lambda_j \cdot \theta_{i}, \sigma_{\epsilon,j}) \\ X_{ij}|Z_i&=0 \sim \mathcal{N}(\mu_{C}, \sigma_{C}), \end{aligned}\end{align} $$

where $\phi \left (\cdot \right )$ represents the normal density function. The item parameters $\mu _{j}$ , $\lambda _j$ , and $\sigma _{\epsilon ,j}$ represent intercepts, item loading, and residual standard deviations in the attentive group. It is important to note that, because the simulation design is based on a unidimensional model, $\lambda _j$ is a scalar rather than a vector. In the mixture CFA model, $\lambda _j$ is estimated using a half-normal prior with a mean of 0 and a standard deviation of 10. Its direction (positive or negative) is constrained based on the item’s wording. Careless responses are assumed to stem from a normal distribution $\mathcal {N}(\mu _{C}, \sigma _{C})$ with a common mean and standard deviation. The remaining simulation settings, such as the number of iterations, were kept consistent to ensure comparability.

The model convergence rate significantly declined under the overly consistent careless response condition compared to the proposed model (51% versus 93% when $1-\pi _{\mathcal {P}} = 0.15$ ). This decline is likely due to careless responses strongly deviating from normality, which might impair the mixture CFA model’s ability to differentiate the two groups. In contrast, for the other two conditions involving random careless responses, the model showed better convergence rates (over 80%).

Based on the converged replications, Table 5 summarizes the classification results of the mixture CFA model when $1-\pi _{\mathcal {P}} = 0.15$ . Notably, as $1-\pi _{\mathcal {P}}$ decreases, the classification performance worsens, with the FPR reaching 0.7 under the threshold method. For brevity, we focus on the conditions where $1-\pi _{\mathcal {P}} = 0.15$ . Table 5 highlights that the mixture CFA model fails to classify attentive and careless respondents effectively in terms of accuracy, FPR, and FNR. It is worth noting that when careless responding was generated using a truncated normal distribution, the mixture CFA model performed better than under other conditions. However, it still fell short of achieving acceptable levels of classification error rates. Although the generated and estimated distributions of careless responses were closely aligned in this scenario, the model’s limitations likely stem from a violation of its assumptions for the attentive group.

Table 5 Classification results of the mixture CFA model when $1-\pi _{\mathcal {P}} = 0.15$ .

Note: Beta = Random responses generated from $\text {Beta}(0.5, 0.5)$ ; Preference = Overly consistent pattern; Normal = Random responses generated from a truncated normal distribution; The term $1-\pi _{\mathcal {P}}$ indicates the true population-level proportion of careless respondents; FPR = False positive rate; FNR = False negative rate.

5 Empirical illustration

We applied the model to empirical data with three objectives: (a) to illustrate the practical application of the proposed model for pinpointing careless respondents, (b) to assess the impact of adjusting for careless respondents on parameter estimation, and (c) to compare the prevalence of careless behaviors between VAS and Likert data. The data used in this analysis come from a study that collected responses through two distinct scale formats—VAS and Likert—across three questionnaires, labeled A, B, and C, that differed in which scales and formats were administered (Kalistová & Cígler, Reference Kalistová and Cígler2021). These questionnaires measured two constructs: height, which captures individuals’ perceptions and experiences related to their physical stature (Tancoš, Reference Tancoš2018), and autonomy, which reflects the extent to which individuals feel free to make their own decisions and express their preferences (Johnston & Finney, Reference Johnston and Finney2010). We utilized data from questionnaire C to demonstrate the application of the proposed model, as it includes responses for both the height and autonomy scales in both VAS and Likert formats, allowing for a direct comparison between these response formats.

Questionnaire C is organized into four sequential blocks: 1) Height assessed using a VAS, 2) Autonomy assessed using a VAS, 3) Height assessed using a Likert scale, and 4) Autonomy assessed using a Likert scale. The Likert scale used in this study is a 5-point scale. The height scale contains 11 items (labeled h1–h11), of which 6 are negatively-worded, while the autonomy scale includes seven items (labeled a1–a7) with 3 negatively-worded items. Rows containing complete missing data in either the VAS or Likert scale were excluded from the analysis, resulting in a sample size of $N = 851$ .

We applied four models to analyze the data. For the VAS data, we used the Beta IRM (Noel & Dauvier, Reference Noel and Dauvier2007), which assumes all responses are attentive and incorporates two factors—height and autonomy. This was compared to the proposed model, which is designed to identify careless respondents. For the Likert scale data, we applied the generalized partial credit model (gPCM), which does not account for careless responding. We also included an ordinal mixture model (Ulitzsch, Pohl, et al., Reference Ulitzsch, Pohl, Khorramdel, Kroehne and von Davier2024), which was designed to detect careless respondents in ordinal data. The ordinal mixture model follows a similar approach as the proposed model to disentangle attentive from careless patterns; however, it is designed for ordinal data and utilizes the gPCM instead of the Beta IRM for the attentive class. For the careless class, the model estimates marginal category probabilities of inattentively choosing a given category over all types of careless behavior. Specifically, the probability of observing response category $X_{ij}=k \in \{0, 1, \dots , K\}$ is given by:

(6) $$ \begin{align} \begin{aligned} \Pr(X_{ij} & = k) = \pi_{i} \cdot \Pr(X_{ij} = k \mid Z_i = 1) + (1-\pi_{i}) \cdot \Pr(X_{ij} = k \mid Z_i = 0) \\ \Pr(X_{ij} & = k \mid Z_i = 1) = \frac{\exp\left( \sum_{l=0}^{k} \left(w_{j} \cdot a_j \cdot \theta_{id[j]} - b_{jl} \right) \right)}{\sum_{r=0}^{K} \exp\left( \sum_{l=0}^{r} \left(w_{j} \cdot a_j \cdot \theta_{id[j]} - b_{jl} \right) \right)} \\ \Pr(X_{ij} &= k \mid Z_i = 0) = \kappa_k, \quad \text{with} \sum_{k=0}^K \kappa_k=1, \end{aligned} \end{align} $$

where $a_j$ is the discrimination parameter, controlling how strongly the item differentiates between respondents with different trait levels and $w_j$ is a pre-specified wording parameter. The step difficulty parameter $b_{jl}$ governs the relative difficulty of selecting higher response categories. Note that $\sum _{l=0}^{0} \left (w_{j} \cdot a_j \cdot \theta _{id[j]} - b_{jl} \right ) \equiv 0$ . For more detailed information regarding the ordinal mixture model, please refer to Ulitzsch, Pohl, et al. (Reference Ulitzsch, Pohl, Khorramdel, Kroehne and von Davier2024).

Prior settings for the proposed model and Beta IRM were the same as in the simulation study. However, unlike the simulation study, the empirical study modeled two factors: height and autonomy. To estimate the correlation between these factors, an LKJ (Lewandowski et al., Reference Lewandowski, Kurowicka and Joe2009) prior with a shape parameter of 1 was used for the factor correlation matrix. For the ordinal mixture model and gPCM, we assigned weakly informative priors to the discrimination ( $\ln a_j \sim \mathcal {N}(0, 10)$ ) and threshold parameters ( $b_{jl} \sim \mathcal {N}(0, 10)$ ). The inattentive response probabilities $\boldsymbol {\kappa }$ are drawn from a Dirichlet prior with all parameters equal to 1. The priors for estimating the class probabilities in the ordinal mixture model were consistent with those specified for the proposed model in the simulation study. The variances of the latent traits were fixed to 1 for model identification in both the ordinal mixture model and gPCM. In all models used in the empirical study, two MCMC chains were generated with 10,000 iterations, with the first half designated as burn-in iterations. All models reached convergence, as indicated by EPSR values of less than 1.1 (Gelman, Reference Gelman1996).

5.1 Investigating careless responding in VAS data

The population-level proportion of careless respondents ( $1-\pi _{\mathcal {P}}$ ) for the VAS was 0.08, with a 95% credibility interval of [0.069, 0.092]. The parameters for the Beta distribution of careless responses, with $m = 0.388$ (95% CI: [0.357, 0.420]) and $n = 0.728$ (95% CI: [0.640, 0.823]), indicated that, marginally, careless respondents tended to select the left side of the slider scale (Figure 6).

Figure 6 Beta(0.388, 0.728): Model-implied distribution of careless responses in the empirical illustration.

Given results from the simulation study (Figures 1 and 2), we identified the careless respondents using $\pi _{\mathcal {P}}$ . We ranked the $\pi _i$ values and identified the least likely respondents as careless by selecting the bottom $\lceil (1-\pi _{\mathcal {P}}) \times N \rceil = 69$ individuals according to their individual probabilities.

Figure 7 compares the item correlations within and across the two scales for the attentive and careless groups. Responses from the attentive group showed strong correlations within both the height and autonomy scales. The negatively-worded items exhibited negative correlations with the positively worded items within the same scale. This demonstrates that attentive respondents answered according to item wording, aligning with the scale design. The correlations between items across the two scales were very low, highlighting the distinctiveness of the height and autonomy measures. Additionally, the low factor correlation that was not credibly different from zero further suggested the distinct separation between the height and autonomy constructs ( $\rho _{ha}$ in Table 6).

Figure 7 Inter-item correlations for attentive and careless groups for VAS.

Note: The item labels “h” and “a” represent the respective scales for “height” and “autonomy.”

Table 6 Summary of estimates for VAS data

Note: The 2.50% and 97.50% columns indicate the lower and upper bounds of the 95% credible interval. The symbols $\delta $ and $\tau $ represent item difficulty and dispersion, $\sigma $ and $\rho $ denote the standard deviations of the height and autonomy factors and the correlation between them, m and n indicate the Beta distribution shape for careless responses, and $1-\pi _{\mathcal {P}}$ is the estimated population-level proportion of careless respondents.

In contrast, the careless group displayed a more chaotic correlation structure. There were weaker and more inconsistent correlations within the same scale, and the negatively-worded items did not display the expected negative correlations. This suggests that careless respondents tended to ignore item wording and might not have distinguished between scales, leading to a breakdown in the expected correlation structure.

5.2 Impact of disregarding careless respondents on parameter estimates

Results of parameter estimates and credible intervals for the difficulty and dispersion parameters are provided in Table 6. After accounting for the careless respondents, the factor correlation ( $\rho _{ha}$ ) slightly increased from 0.032 (95% CI: [ $-$ 0.039, 0.106]) to 0.054 (95% CI: [ $-$ 0.029, .134]), though both estimates remained not credibly different from zero. Figure 8 displays a comparison of the estimates between the proposed model and the Beta IRM that does not take careless responding into account. Generally, both models yielded similar estimates for item difficulty parameters ( $\delta $ ) across the height and autonomy scales. However, the proposed model consistently yielded higher estimates for item dispersion parameters. As depicted in Figure 2, increased dispersion implies enhanced item informativeness, indicating an improvement in item psychometric properties upon accounting for careless responses. Figure 9 further compares the factor score estimates between the proposed model and the Beta IRM, plotted against person-level posterior attentiveness probability obtained from the proposed model. As can be seen, differences in factor score estimates were especially pronounced for individuals with lower attentiveness probabilities, while the two models yielded similar factor score estimates when the posterior attentiveness probability was high.

Figure 8 Item parameter estimates for VAS obtained from different models.

Note: The dashed line shows where the obtained estimates are equal; the labels “h” and “a” represent the item parameters of the “height” and “autonomy” scales.

Figure 9 Difference in factor score estimates.

Note: $\pi _i$ denotes the probability that person i belongs to the attentive group; the labels “h” and “a” represent the respective factors of “height” and “autonomy.”

5.3 Comparison between VAS and Likert scale

With a population-level estimate of 0.047 (95% CI [0.035, 0.061]), the proportion of careless respondents in the Likert scale data was lower than that observed in the VAS data. To evaluate whether the difference was indeed credibly different from zero, we estimated it as a derived parameter (0.034; [0.015, 0.052]). Partially mirroring results obtained for the VAS, marginally, careless respondents tended to prefer the lower two ( $\kappa _0 =$ 0.268, [0.160, 0.422]; $\kappa _1=$ 0.260, [0.136, 0.389]) as well as the middle upper categories ( $\kappa _3=$ 0.239 [0.178, 0.304]).

There was some overlap in the identification of careless respondents between the VAS and the Likert scale. Table 7 presents the contingency table for classification in the VAS and Likert scale data. While some respondents were identified as careless on only one of the scales, an overall agreement of 90.71% between the two scales reflects a high level of consistency in the classifications. Cohen’s $\kappa $ (0.23, 95% CI: [0.12, 0.34]) indicated a fair level of agreement between the Likert and VAS classifications (Cohen, Reference Cohen1960). A chi-square test suggested a significant association between the respondent classifications of attentive and careless across two scales ( $\chi ^2(1, N =851)= 44.615, p < 0.001$ ).

Figure 10 compares the individual probabilities of belonging to the attentive group across the two scales. The results showed a moderate correlation between the individual probabilities ( $r = 0.41$ ). Notably, if a classification threshold of 0.5 was applied, none of the respondents would be identified as careless for both scale formats. This finding aligns with the simulation results and suggests that the 0.5 threshold might have difficulty identifying careless respondents when the population-level proportion of such respondents is low.

Table 7 Contingency table for classification.

Figure 10 Comparison of individual probabilities ( $\pi _i$ ) between VAS and Likert scale.

Figure 11 shows the estimated threshold and discrimination parameters between the ordinal mixture model and gPCM. Consistent with the comparison between the proposed model and the Beta IRM (Figure 8), most of the threshold estimates remained very similar before and after accounting for careless responders. The discrimination parameters, however, increased after considering careless responses in the mixture model. Notably, negatively worded items exhibited a greater increase in discrimination estimates. Moreover, the factor correlation in the Likert scales also slightly increased from 0.039 (95% CI: [ $-$ 0.037, 0.113]) to 0.055 (95% CI: [ $-$ 0.017, 0.139]); however, both estimates remained not credibly different from zero.

Figure 11 Item parameter estimates for Likert scale obtained from different models.

Note: The dashed line shows where the obtained estimates are equal; the labels “h” and “a” represent the item parameters of the “height” and “autonomy” scales.

6 Discussion

We proposed a Beta Mixture IRM designed to identify careless respondents in VAS data. VASs are frequently used in fields like psychology, education, and health research to capture subjective experiences with greater sensitivity than ordinal scales. However, careless responses can undermine this sensitivity, skewing results, and compromising the validity of findings. The proposed model allows to detect and account for these responses.

In our simulation study, the proposed model effectively distinguished between attentive and careless respondents. It performed best in both classification and parameter estimation when careless responses followed the assumed Beta distribution. When the distribution of careless responses deviated from the Beta distribution used to “absorb” careless responses, although the proportion of careless respondents was slightly overestimated, the amount of bias was, in our view, small enough to still render the proposed model a useful tool to gauge the overall extent of careless responding in the data. Simulation results still demonstrated high classification accuracy across these varied scenarios. These results suggest that the proposed model provides a flexible and practical framework for capturing a broad range of careless response behaviors. Moreover, item parameters were well recovered across all conditions, indicating that the proposed model is an effective tool to adjust for careless responding when estimating substantive parameters of interest. Furthermore, the simulation results underscored the limitations of existing mixture CFA models, which are challenged by violations of normality in both attentive and careless responding.

Our empirical results further demonstrated the practical applicability of the proposed model by highlighting distinct correlation patterns between attentive and careless respondents. Attentive respondents showed strong, consistent correlations within the same scale, with negatively-worded items correlating negatively as expected, whereas careless respondents displayed weak and inconsistent correlations, indicating a disregard for item wording. When compared with the Beta IRM that assumes all responses to be attentive, accounting for careless respondents yielded higher dispersion estimates for VAS items, which suggests that once careless respondents are identified and accounted for, the item information improves.

The unique design of the empirical data, administering the same scales with both VAS and Likert scale formats, further allowed us to conduct initial evaluations on differential scale-specific prevalences of careless responding. In the data at hand, the Likert scale indicated a lower proportion of careless respondents. However, while the probability of attentiveness was moderately correlated between the VAS and Likert scales, some respondents were classified as careless in only one format. A possible explanation for this finding may be that some respondents may engage more effectively with certain response formats. For instance, Dourado et al. (Reference Dourado, Volpato, de Almeida-Pedrin, Oltramari, Fernandes and Conti2021) found that most respondents preferred the Likert scale in assessing facial pleasantness, as they found it easier to express their opinions compared to the VAS. Future research could investigate whether allowing respondents to choose their preferred response format (Kutscher & Eid, Reference Kutscher and Eid2024) might help reduce careless responding. It should be noted, however, that the within-subject design used in this study does not allow to disentangle order effects from scale effects, leaving room for alternative interpretations. For instance, respondents might have been influenced by the order in which the response formats were presented, rather than engaging differently with the scales themselves. Additionally, it is possible that the models did not adequately uncover careless responding, which could account for some of the observed discrepancies. Future research is encouraged to address these design limitations, further evaluate the robustness of the proposed model, and explore attentiveness differences across different scale formats.

6.1 Limitations and future directions

Practically, the proposed model offers a safeguard for researchers using VASs to ensure their data accurately reflects the constructs being measured. There are several promising directions for future research to further enhance careless respondent detection and survey data reliability. One such direction is the integration of collateral information, such as response times, to enhance the detection of inattentive behaviors. Response time has been shown to correlate with engagement, with shorter times often reflecting careless responding. Adding this diagnostic layer could create a more robust approach to identifying carelessness, ultimately improving the reliability and validity of survey data (Ulitzsch, Pohl, et al., Reference Ulitzsch, Yildirim-Erbasli, Gorgun and Bulut2022; Zhang et al., Reference Zhang, Ulitzsch and Domingue2025). We suspect that incorporating response time information into the proposed model may be especially advantageous when the distribution of careless responses deviates from a perfect fit with the Beta distribution.

Note that the proposed model identifies careless responding at the respondent level. However, inattentiveness may vary across specific items or sections of a questionnaire. For instance, respondents may start attentively but be more prone to careless responding later in the questionnaire. This raises the need for future research to identify careless responding at the level of single item responses, enabling researchers to pinpoint areas within a survey that may be more prone to inattention. A straightforward way to do so would be to integrate the proposed models with previous model developments that allow for varying attentiveness across the questionnaire but, so far, have only been available for ordinal and continuous, unbounded item responses (Roman et al., Reference Roman, Schmidt, Miller and Brandt2024; Ulitzsch, Pohl, et al., Reference Ulitzsch, Pohl, Khorramdel, Kroehne and von Davier2022; Ulitzsch, Yildirim-Erbasli, et al., Reference Ulitzsch, Yildirim-Erbasli, Gorgun and Bulut2022).

Furthermore, the model’s reliance on assumptions about attentive behavior poses a limitation; if the attentive response model is mis-specified (e.g., a uni- instead is assumed but a multidimensional model holds), the accuracy of detecting careless responses could be compromised (Vogelsmeier et al., Reference Vogelsmeier, Uglanova, Rein and Ulitzsch2024). Future research is recommended to evaluate the performance of the proposed model under these conditions. Additionally, the data generation setting considered in the simulation study ( $\delta _j \sim \mathcal {N}(0,1)$ ), which results in highly heterogeneous items, is particularly favorable for the model. It is important to note that the model’s performance may decline when item heterogeneity is reduced.

Another important consideration is that we presented the model with a deliberately misspecified careless component model, designed to “absorb” different careless response patterns present in the data, described marginally by a common Beta distribution. If researchers want to disentangle different careless response patterns, they could specify the model with an empirically determined number of careless classes, each with its own Beta distribution. Note, however, that this approach may be challenging and unstable to estimate when some specific careless classes are very small, i.e., when a specific careless response pattern is only exhibited by very few respondents. Future research may compare this approach with the residual class approach taken in the present study and derive guidelines on how to choose between these different approaches.

We have introduced and evaluated the proposed model for cross-sectional data. Nevertheless, we see its highest applicability in the context of longitudinal ecological momentary assessments, where VASs are a common scale of choice to capture real-time changes in participant experiences (Haslbeck et al., Reference Haslbeck, Jover Martínez, Roefs, Fried, Lemmens, Groot and Edelsbrunner2024). As participants are at risk of becoming inattentive over time due to the repetitive nature and increased respondent burden of ecological momentary assessments (Eisele et al., Reference Eisele, Vachon, Lafit, Kuppens, Houben, Myin-Germeys and Viechtbauer2022; Ulitzsch, Nestler, et al., Reference Ulitzsch, Nestler, Lüdtke and Nagy2024; Ulitzsch, Viechtbauer, et al., Reference Ulitzsch, Viechtbauer, Lüdtke, Myin-Germeys, Nagy, Nestler and Eisele2025; Vogelsmeier et al., Reference Vogelsmeier, Uglanova, Rein and Ulitzsch2024), future research could integrate the mixture model with dynamic models to track attentiveness over time, which, so far, have only been developed for Likert-type scales (Vogelsmeier et al., Reference Vogelsmeier, Uglanova, Rein and Ulitzsch2024). This extension would be particularly valuable for health and psychological well-being studies, where real-time, accurate data is critical for understanding changes in subjective experiences.

Data availability statement

A preprint of this manuscript is available at https://doi.org/10.31219/osf.io/tp6df_v1 and the code for the empirical study can be accessed through this OSF Project.

Funding statement

This work was supported by the Research Council of Norway through its Centres of Excellence scheme (project number 33160), the Jacobs Foundation, and the Stanford Data Science Fellowship.

Competing interests

The authors declare no competing interests.

Footnotes

1 We conducted a test run with 200 replications for one simulation condition. The results were highly consistent with those from the 100-replication run. Given this consistency and the high computational cost (around 1-hour per dataset on an M2 Max chip), we retained 100 replications for efficiency.

References

Arias, V. B., Garrido, L., Jenaro, C., Martinez-Molina, A., & Arias, B. (2020). A little garbage in, lots of garbage out: Assessing the impact of careless responding in personality survey data. Behavior Research Methods, 52, 24892505. https://doi.org/10.3758/s13428-020-01401-8 CrossRefGoogle ScholarPubMed
Askim, K., & Knardahl, S. (2021). The influence of affective state on subjective-report measurements: Evidence from experimental manipulations of mood. Frontiers in Psychology, 12, 601083. https://doi.org/10.3389/fpsyg.2021.601083 CrossRefGoogle ScholarPubMed
Åström, M., Thet Lwin, Z. M., Teni, F. S., Burström, K., & Berg, J. (2023). Use of the visual analogue scale for health state valuation: A scoping review. Quality of Life Research, 32(10), 27192729. https://doi.org/10.1007/s11136-023-03411-3 CrossRefGoogle ScholarPubMed
Brumfitt, S. M., & Sheeran, P. (1999). The development and validation of the visual analogue self-esteem scale (VASES). British Journal of Clinical Psychology, 38(4), 387400. https://doi.org/10.1348/014466599162980 CrossRefGoogle ScholarPubMed
Carpenter, B., Gelman, A., Hoffman, M. D., Lee, D., Goodrich, B., Betancourt, M., Brubaker, M. A., Guo, J., Li, P., & Riddell, A. (2017). Stan: A probabilistic programming language. Journal of Statistical Software, 76, 132.10.18637/jss.v076.i01CrossRefGoogle Scholar
Cohen, J. (1960). A coefficient of agreement for nominal scales. Educational and Psychological Measurement, 20(1), 3746. https://doi.org/10.1177/001316446002000104 CrossRefGoogle Scholar
Domingue, B., Braginsky, M., Caffrey-Maffei, L., Gilbert, J., Kanopka, K., Kapoor, R., Lee, H., Liu, Y., Nadela, S., Pan, G., Zhang, L., Zhang, S., & Frank, M. (2025). An introduction to the Item Response Warehouse (IRW): A resource for enhancing data usage in psychometrics. Behavior Research Methods, 57, 276. https://doi.org/10.3758/s13428-025-02796-y CrossRefGoogle Scholar
Douglas, B. D., Ewell, P. J., & Brauer, M. (2023). Data quality in online human-subjects research: Comparisons between MTurk, Prolific, CloudResearch, Qualtrics, and SONA. PLoS One, 18(3), e0279720. https://doi.org/10.1371/journal.pone.0279720 CrossRefGoogle ScholarPubMed
Dourado, G. B., Volpato, G. H., de Almeida-Pedrin, R. R., Oltramari, P. V. P., Fernandes, T. M. F., & Conti, A. C. d. C. F. Placeholder Text(2021). Likert scale vs visual analog scale for assessing facial pleasantness. American Journal of Orthodontics and Dentofacial Orthopedics, 160(6), 844852. https://doi.org/10.1016/j.ajodo.2020.05.024 CrossRefGoogle ScholarPubMed
Eisele, G., Vachon, H., Lafit, G., Kuppens, P., Houben, M., Myin-Germeys, I., & Viechtbauer, W. (2022). The effects of sampling frequency and questionnaire length on perceived burden, compliance, and careless responding in experience sampling data in a student population. Assessment, 29(2), 136151. https://doi.org/10.1177/1073191120957102 CrossRefGoogle Scholar
Ferrari, S., & Cribari-Neto, F. (2004). Beta regression for modelling rates and proportions. Journal of Applied Statistics, 31(7), 799815. https://doi.org/10.1080/0266476042000214501 CrossRefGoogle Scholar
Gelman, A. (1996). Markov chain Monte Carlo in practice. Chapman & Hall.Google Scholar
Gelman, A., Carlin, J. B., Stern, H. S., Dunson, D. B., Vehtari, A., & Rubin, D. B. (2013). Bayesian data analysis. CRC press.10.1201/b16018CrossRefGoogle Scholar
Haslbeck, J. M. B., Jover Martínez, A., Roefs, A., Fried, E. I., Lemmens, L. H., Groot, E., & Edelsbrunner, P. A. (2024). Comparing likert and visual analogue scales in ecological momentary assessment. PsyArXiv. https://doi.org/10.31234/osf.io/yt8xw CrossRefGoogle Scholar
Hayes, M. S., & Patterson, D. (1921). Experimental development of the graphic rating method. Psychological Bulletin, 18, 9899.Google Scholar
Huang, J. L., Liu, M., & Bowling, N. A. (2015). Insufficient effort responding: Examining an insidious confound in survey data. Journal of Applied Psychology, 100(3), 828845. https://doi.org/10.1037/a0038510 CrossRefGoogle ScholarPubMed
Jebb, A. T., Ng, V., & Tay, L. (2021). A review of key likert scale development advances: 1995–2019. Frontiers in Psychology, 12, 637547.10.3389/fpsyg.2021.637547CrossRefGoogle ScholarPubMed
Johnston, M. M., & Finney, S. J. (2010). Measuring basic needs satisfaction: Evaluating previous research and conducting new psychometric evaluations of the basic needs satisfaction in general scale. Contemporary Educational Psychology, 35(4), 280296. https://doi.org/10.1016/j.cedpsych.2010.04.003 CrossRefGoogle Scholar
Kalistová, G., & Cígler, H. (2021). Comparison of psychometric properties of likert scale and visual analog scale. [Last Updated: 2021-11-14]. https://osf.io/xbmpw/ Google Scholar
Kam, C. C. S., & Cheung, S. F. (2023). A constrained factor mixture model for detecting careless responses that is simple to implement. Organizational Research Methods, 27(3), 443476. https://doi.org/10.1177/10944281231195298 CrossRefGoogle Scholar
Kemp, C., Perfors, A., & Tenenbaum, J. B. (2007). Learning overhypotheses with hierarchical Bayesian models. Developmental Science, 10(3), 307321.10.1111/j.1467-7687.2007.00585.xCrossRefGoogle ScholarPubMed
Kuhlmann, T., Dantlgraber, M., & Reips, U.-D. (2017). Investigating measurement equivalence of visual analogue scales and likert-type scales in internet-based personality questionnaires. Behavior Research Methods, 49, 21732181. https://doi.org/10.3758/s13428-016-0850-x CrossRefGoogle ScholarPubMed
Kutscher, T., & Eid, M. (2024). Psychometric benefits of self-chosen rating scales over given rating scales. Behavior Research Methods, 56(7), 74407464. https://doi.org/10.3758/s13428-024-02429-w CrossRefGoogle ScholarPubMed
Lewandowski, D., Kurowicka, D., & Joe, H. (2009). Generating random correlation matrices based on vines and extended onion method. Journal of Multivariate Analysis, 100(9), 19892001.10.1016/j.jmva.2009.04.008CrossRefGoogle Scholar
Likert, R. (1932). A technique for the measurement of attitudes. Archives of Psychology, 140, 555.Google Scholar
MacCallum, R. C., Widaman, K. F., Zhang, S., & Hong, S. (1999). Sample size in factor analysis. Psychological Methods, 4(1), 84.10.1037/1082-989X.4.1.84CrossRefGoogle Scholar
McCormack, H. M., Horne, D. J., & Sheather, S. (1988). Clinical applications of visual analogue scales: A critical review. Psychological Medicine, 18(4), 10071019. https://doi.org/10.1017/s0033291700009934 CrossRefGoogle ScholarPubMed
Meade, A. W., & Craig, S. B. (2012). Identifying careless responses in survey data. Psychological Methods, 17(3), 437. https://doi.org/10.1037/a0028085 CrossRefGoogle ScholarPubMed
Myles, P., Myles, D., Galagher, W., Boyd, D., Chew, C., MacDonald, N., & Dennis, A. (2017). Measuring acute postoperative pain using the visual analog scale: The minimal clinically important difference and patient acceptable symptom state. British Journal of Anaesthesia, 118(3), 424429. https://doi.org/10.1093/bja/aew466 CrossRefGoogle ScholarPubMed
Noel, Y., & Dauvier, B. (2007). A beta item response model for continuous bounded responses. Applied Psychological Measurement, 31(1), 4773. https://doi.org/10.1177/0146621605287691 CrossRefGoogle Scholar
Price, D. D., Bush, F. M., Long, S., & Harkins, S. W. (1994). A comparison of pain measurement characteristics of mechanical visual analogue and simple numerical rating scales. Pain, 56(2), 217226. https://doi.org/10.1016/0304-3959(94)90097-3 CrossRefGoogle ScholarPubMed
R Core Team. (2023). R: A language and environment for statistical computing. R Foundation for Statistical Computing. Placeholder TextPlaceholder Texthttps://www.R-project.org/ Google Scholar
Reips, U.-D., & Funke, F. (2008). Interval-level measurement with visual analogue scales in internet-based research: VAS generator Behavior Research Methods, 40(3), 699704. https://doi.org/10.3758/BRM.40.3.699 CrossRefGoogle ScholarPubMed
Roman, Z. J., Schmidt, P., Miller, J. M., & Brandt, H. (2024). Identifying dynamic shifts to careless and insufficient effort behavior in questionnaire responses: A novel approach and experimental validation. Structural Equation Modeling: A Multidisciplinary Journal, 31(5), 775793. https://doi.org/10.1080/10705511.2024.2304816 CrossRefGoogle Scholar
Salakhutdinov, R., Tenenbaum, J. B., & Torralba, A. (2012). Learning with hierarchical-deep models. IEEE Transactions on Pattern Analysis and Machine Intelligence, 35(8), 19581971.10.1109/TPAMI.2012.269CrossRefGoogle Scholar
Sung, Y.-T., & Wu, J.-S. (2018). The visual analogue scale for rating, ranking and paired-comparison (VAS-RRP): A new technique for psychological measurement. Behavior Research Methods, 50, 16941715. https://doi.org/10.3758/s13428-018-1041-8 CrossRefGoogle ScholarPubMed
Tancoš, M. (2018). The influence of verbal labels of likert-type scales on psychometric properties of questionnaires [Unpublished Bachelor’s Thesis]. Masaryk University. https://is.muni.cz/th/uk8cb/BP_Tancos.pdf Google Scholar
Uglanova, I., Nagy, G., & Ulitzsch, E. (2025). A mixture IRT model for handling different types of careless respondents. PsyArXiv. https://doi.org/10.31219/osf.io/tgys3˙v1 CrossRefGoogle Scholar
Ulitzsch, E., Nestler, S., Lüdtke, O., & Nagy, G. (2024). A screen-time-based mixture model for identifying and monitoring careless and insufficient effort responding in ecological momentary assessment data. Psychological Methods. Advance Online Publication. https://doi.org/10.1037/met0000636 CrossRefGoogle ScholarPubMed
Ulitzsch, E., Pohl, S., Khorramdel, L., Kroehne, U., & von Davier, M. (2022). A response-time-based latent response mixture model for identifying and modeling careless and insufficient effort responding in survey data. Psychometrika, 87(2), 593619. https://doi.org/10.1007/s11336-021-09817-7 CrossRefGoogle ScholarPubMed
Ulitzsch, E., Pohl, S., Khorramdel, L., Kroehne, U., & von Davier, M. (2024). Using response times for joint modeling of careless responding and attentive response styles. Journal of Educational and Behavioral Statistics, 49(2), 173206. https://doi.org/10.3102/10769986231173607 CrossRefGoogle Scholar
Ulitzsch, E., Viechtbauer, W., Lüdtke, O., Myin-Germeys, I., Nagy, G., Nestler, S., & Eisele, G. V. (2025). Investigating the effect of experience sampling study design on careless and insufficient effort responding identified with a screen-time-based mixture model. Psychological Assessment, 37(8), 347359. https://doi.org/10.31234/osf.io/p9cfm CrossRefGoogle Scholar
Ulitzsch, E., Yildirim-Erbasli, S. N., Gorgun, G., & Bulut, O. (2022). An explanatory mixture IRT model for careless and insufficient effort responding in survey data. British Journal of Mathematical and Statistical Psychology, 75, 668698. https://doi.org/10.1111/bmsp.12272 CrossRefGoogle Scholar
Van Laerhoven, H., van der Zaag-Loonen, H. J., & Derkx, B. H. (2004). A comparison of likert scale and visual analogue scales as response options in children’s questionnaires. Acta Paediatrica, 93(6), 830835. https://doi.org/10.1111/j.1651-2227.2004.tb03026.x CrossRefGoogle ScholarPubMed
van Laar, S., & Braeken, J. (2022). Random responders in the TIMSS 2015 student questionnaire: A threat to validity? Journal of Educational Measurement, 59(4), 407501. https://doi.org/10.1111/jedm.12317 CrossRefGoogle Scholar
Verkuilen, J., & Smithson, M. (2012). Mixed and mixture regression models for continuous bounded responses using the beta distribution. Journal of Educational and Behavioral Statistics, 37(1), 82113.10.3102/1076998610396895CrossRefGoogle Scholar
Vogelsmeier, L. V. D. E., Uglanova, I., Rein, M. T., & Ulitzsch, E. (2024). Investigating dynamics in attentive and inattentive responding together with their contextual correlates using a novel mixture IRT model for intensive longitudinal data. British Journal of Mathematical and Statistical Psychology. Advance online publication.10.1111/bmsp.12373CrossRefGoogle ScholarPubMed
Voutilainen, A., Pitkäaho, T., Kvist, T., & Vehviläinen-Julkunen, K. (2016). How to ask about patient satisfaction? the visual analogue scale is less vulnerable to confounding factors and ceiling effect than a symmetric likert scale. Journal of Advanced Nursing, 72(4), 946957. https://doi.org/10.1111/jan.12875 CrossRefGoogle ScholarPubMed
Ward, M., & Meade, A. W. (2023). Dealing with careless responding in survey data: Prevention, identification, and recommended best practices. Annual Review of Psychology, 74, 577596. https://doi.org/10.1146/annurev-psych-040422-045007 CrossRefGoogle ScholarPubMed
Weigl, K., & Forstner, T. (2021). Design of paper-based visual analogue scale items. Educational and Psychological Measurement, 81(3), 595611. https://doi.org/10.1177/0013164420952118 CrossRefGoogle ScholarPubMed
Wortmann, H. R., Gisch, U. A., & Warschburger, P. (2021). Psychometric evaluation of the German version of the variety seeking tendency scale (VARSEEK). Food Quality and Preference, 94, 104319.10.1016/j.foodqual.2021.104319CrossRefGoogle Scholar
Zhang, L., Ulitzsch, E., & Domingue, B. (2025). Bayesian factor mixture modeling with response time for detecting careless respondents. Behavior Research Methods. Advance Online Publication.10.3758/s13428-025-02797-xCrossRefGoogle Scholar
Zhou, W., & Chen, D. (2009). Fear-related chemosignals modulate recognition of fear in ambiguous facial expressions. Psychological Science, 20(2), 177183. https://doi.org/10.1111/j.1467-9280.2009.02263.x CrossRefGoogle ScholarPubMed
Figure 0

Figure 1 Likert scale and VAS.

Figure 1

Figure 2 Beta IRM.Note: (a) ICC for different difficulty parameters, with dispersion fixed at 0. (b) ICC with response densities for $\theta _i - \delta _j = 0$ under different dispersion parameters. (c) Item Information Function for different dispersion parameters, with difficulty fixed at 0.

Figure 2

Figure 3 Model structure.Note: For attentive respondents, responses are assumed to follow a Beta IRM, governed by item and person parameters; a two-factor, six-item structure is used as an example. For careless respondents, responses are assumed to stem from a common Beta distribution, with Beta(2, 5) used as an illustrative example in the figure.

Figure 3

Figure 4 Density plot for different Beta distributions.

Figure 4

Table 1 Classification results based on the threshold

Figure 5

Table 2 Classification results based on $\pi _{\mathcal {P}}$

Figure 6

Table 3 Estimation accuracy of item parameters and $\pi _{\mathcal {P}}$

Figure 7

Figure 5 Beta approximation for careless responses.Note: Beta = Random responses generated from $\text {Beta}(0.5, 0.5)$; Preference = Overly consistent pattern; Normal = Random responses generated from a truncated normal distribution. Blue lines indicate the Beta approximation for careless responses across all simulated datasets, while red dashed lines represent the distribution used for data generation.

Figure 8

Table 4 Parameter estimation for the careless response distributions

Figure 9

Table 5 Classification results of the mixture CFA model when $1-\pi _{\mathcal {P}} = 0.15$.

Figure 10

Figure 6 Beta(0.388, 0.728): Model-implied distribution of careless responses in the empirical illustration.

Figure 11

Figure 7 Inter-item correlations for attentive and careless groups for VAS.Note: The item labels “h” and “a” represent the respective scales for “height” and “autonomy.”

Figure 12

Table 6 Summary of estimates for VAS data

Figure 13

Figure 8 Item parameter estimates for VAS obtained from different models.Note: The dashed line shows where the obtained estimates are equal; the labels “h” and “a” represent the item parameters of the “height” and “autonomy” scales.

Figure 14

Figure 9 Difference in factor score estimates.Note: $\pi _i$ denotes the probability that person i belongs to the attentive group; the labels “h” and “a” represent the respective factors of “height” and “autonomy.”

Figure 15

Table 7 Contingency table for classification.

Figure 16

Figure 10 Comparison of individual probabilities ($\pi _i$) between VAS and Likert scale.

Figure 17

Figure 11 Item parameter estimates for Likert scale obtained from different models.Note: The dashed line shows where the obtained estimates are equal; the labels “h” and “a” represent the item parameters of the “height” and “autonomy” scales.