Hostname: page-component-89b8bd64d-ksp62 Total loading time: 0 Render date: 2026-05-08T14:14:47.535Z Has data issue: false hasContentIssue false

Honesty repeats itself: comparing manual and automated coding on the veracity cues total details and redundancy

Published online by Cambridge University Press:  21 October 2024

Haneen Deeb*
Affiliation:
Department of Psychology, University of Portsmouth, Portsmouth, Hampshire, UK
Aldert Vrij
Affiliation:
Department of Psychology, University of Portsmouth, Portsmouth, Hampshire, UK
Nicola Palena
Affiliation:
Netherlands Institute for the Study of Crime and Law Enforcement, Amsterdam, Netherlands
Petra Hypšová
Affiliation:
Department of Psychology, Palacký University, Olomouc, Czech Republic
Gerges Dib
Affiliation:
Amazon.com, Inc., Seattle, Washington, USA
Sharon Leal
Affiliation:
Department of Psychology, University of Portsmouth, Portsmouth, Hampshire, UK
Samantha Mann
Affiliation:
Department of Psychology, University of Portsmouth, Portsmouth, Hampshire, UK
*
Corresponding author: Haneen Deeb; Email: haneen.deeb@port.ac.uk
Rights & Permissions [Opens in a new window]

Abstract

Lie detection research comparing manual and automated coding of linguistic cues is limited. In Experiment 1, we attempted to extend this line of research by directly comparing the veracity differences in manual coding and two coding software programs (Text Inspector and Linguistic Inquiry and Word Count [LIWC]) on the linguistic cue “total details” across eight published datasets. Mixed model analyses revealed that LIWC showed larger veracity differences in total details than Text Inspector and manual coding. Follow-up classification analyses showed that both automated coding and manual coding could accurately classify honest and false accounts. In Experiment 2, we examined if LIWC’s sensitivity to veracity differences was the result of honest accounts including more redundant (repeated) words than false accounts as LIWC—but not Text Inspector or manual coding—accounts for redundancy. Our prediction was supported, and the most redundant words were function words. The results implicated that automated coding can detect veracity differences in total details and redundancy, but it is not necessarily better than manual coding at accurately classifying honest and false accounts.

Information

Type
Original Article
Creative Commons
Creative Common License - CCCreative Common License - BYCreative Common License - NCCreative Common License - ND
This is an Open Access article, distributed under the terms of the Creative Commons Attribution-NonCommercial-NoDerivatives licence (https://creativecommons.org/licenses/by-nc-nd/4.0/), which permits non-commercial re-use, distribution, and reproduction in any medium, provided that no alterations are made and the original article is properly cited. The written permission of Cambridge University Press must be obtained prior to any commercial use and/or adaptation of the article.
Copyright
© The Author(s), 2024. Published by Cambridge University Press
Figure 0

Table 1. Summary of the datasets and conditions included in the present paper

Figure 1

Table 2. Fixed effects parameters for total details as a function of veracity and coding method

Figure 2

Figure 1. Simple effects for total details as a function of veracity and coding method.

Figure 3

Table 3. Simple effects for total details as a function of veracity and coding method

Figure 4

Table 4. Classification accuracy for each coding method based on total details using linear discriminant analysis, XGBoost classifier, and random forest classifier

Figure 5

Table 5. Feature importance of the model in Experiment 1

Figure 6

Figure 2. Means of the redundancy ratio as a function of veracity.

Figure 7

Table 6. T-test results for redundant words that significantly differentiated truth tellers and lie tellers

Figure 8

Table 7. Classification accuracy based on redundancy ratio using linear discriminant analysis, XGBoost classifier, and random forest classifier