Hostname: page-component-8448b6f56d-m8qmq Total loading time: 0 Render date: 2024-04-24T05:32:56.500Z Has data issue: false hasContentIssue false

The evolution of distorted beliefs vs. mistaken choices under asymmetric error costs

Published online by Cambridge University Press:  20 May 2020

Charles Efferson*
Affiliation:
Faculty of Business and Economics, University of Lausanne, Switzerland
Ryan McKay
Affiliation:
Department of Psychology, Royal Holloway, University of London, UK
Ernst Fehr
Affiliation:
Department of Economics, University of Zurich, Switzerland
*
*Corresponding author. E-mail: charles.efferson@unil.ch

Abstract

Why do people sometimes hold unjustified beliefs and make harmful choices? Three hypotheses include (a) contemporary incentives in which some errors cost more than others, (b) cognitive biases evolved to manage ancestral incentives with variation in error costs and (c) social learning based on choice frequencies. With both modelling and a behavioural experiment, we examined all three mechanisms. The model and experiment support the conclusion that contemporary cost asymmetries affect choices by increasing the rate of cheap errors to reduce the rate of expensive errors. Our model shows that a cognitive bias can distort the evolution of beliefs and in turn behaviour. Unless the bias is strong, however, beliefs often evolve in the correct direction. This suggests limitations on how cognitive biases shape choices, which further indicates that detecting the behavioural consequences of biased cognition may sometimes be challenging. Our experiment used a prime intended to activate a bias called ‘hyperactive agency detection’, and the prime had no detectable effect on choices. Finally, both the model and experiment show that frequency-dependent social learning can generate choice dynamics in which some populations converge on widespread errors, but this outcome hinges on the other two mechanisms being neutral with respect to choice.

Type
Research Article
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted re-use, distribution, and reproduction in any medium, provided the original work is properly cited.
Copyright
Copyright © The Author(s) 2020

Media summary: An evolutionary analysis of distorted beliefs and widespread errors when incentives, cognitive bias, and social learning mix.

Introduction

In early modern Europe, Christian zealots put thousands of innocent people to death because these innocents were seen to be agents of Satan (Boyer, Reference Boyer2001). In parts of Asia and Africa, ‘penis panics’ have occurred repeatedly. These outbreaks of paranoia centred on the belief that one's genitals were receding into one's body or had been stolen. People resorted to self-destructive measures to protect themselves, and they arbitrarily accused others of genital thievery (Yap, Reference Yap1965; Sachdev, Reference Sachdev1985; Ilechukwu, Reference Ilechukwu1992; Cheng, Reference Cheng1996; Buckle et al., Reference Buckle, Chuah, Fones and Wong2007; Bures, Reference Bures2008). In 2003, the United States justified its invasion of Iraq by arguing that the Iraqi government was manufacturing weapons of mass destruction. In the wake of the invasion, the evidence overwhelmingly indicated that Iraq did not have such weapons, but many US citizens maintained the opposite belief (Gaines et al., Reference Gaines, Kuklinski, Quirk, Peyton and Verkuilen2007).

Whether witches, disappearing genitalia or bombs that do not exist, why do large numbers of people sometimes hold seemingly unjustified beliefs or make seemingly bad choices? Put differently, what is the evolutionary explanation for widespread errors in a risky and uncertain world (Haselton & Nettle, Reference Haselton and Nettle2006; Foster & Kokko, Reference Foster and Kokko2009)? We distinguish between beliefs in error and choices in error. A belief in error means a decision maker holds a belief about her situation that available evidence does not fully justify. A choice in error comes in two forms, ex ante and ex post. From an ex ante perspective, the decision maker makes a sub-optimal choice given the information she has about her situation at the time of her decision. From an ex post perspective, the outcome that follows from a choice is not the best possible outcome given the situation the decision maker actually faces. Under risk and uncertainty, the two perspectives are linked, but not perfectly so. Bad decision making may increase the probability of bad outcomes, but good outcomes are still possible. Analogously, good decision making may increase the probability of good outcomes, but bad outcomes can still occur.

Here, we address errors in belief and errors in choice by examining the effects of three different mechanisms. Each mechanism represents a prominent hypothesis about the origin and persistence of errors. No hypothesis is mutually exclusive with respect to any other, and so we examine the three mechanisms in all combinations.

First, the incentive structure of the current decision-making task can support frequent errors in choice if some errors are costlier than others (Pascal, Reference Pascal and Levi1995, originally published in 1670). Under asymmetric error costs, choosing optimally requires the decision maker to minimise expected costs. In turn, people should commit cheap errors with a relatively high probability to reduce the probability of expensive errors. This is simply optimal decision making under risk (McKay & Efferson, Reference McKay and Efferson2010). It does not require unjustified beliefs, and choices are errors only from an ex post perspective that conditions on the decision maker's realised environment or situation. Imagine, for example, that Charles offers Ryan a gamble. With probability 0.01, Charles pays Ryan 10,000 dollars. With probability 0.99, Ryan pays Charles one dollar. Objective probabilities are known and verifiable, and so beliefs are perfectly accurate. Ryan accepts the offer, but ex post he typically walks away with one less dollar, a good bet with a bad outcome. The only mechanism of interest is that losing one dollar is far better than losing the opportunity to gain 10,000 dollars. In evolutionary terms, the hypothesis of interest is that evolution has shaped parts of the mind to develop so that they are tolerably adept at general-purpose optimising (Frankenhuis et al., Reference Frankenhuis, Panchanathan and Clark Barrett2013; Barrett, Reference Barrett2015).

Second, error management theory (Haselton & Nettle, Reference Haselton and Nettle2006; McKay & Efferson, Reference McKay and Efferson2010; Johnson et al., Reference Johnson, Blumstein, Fowler and Haselton2013; Marshall et al., Reference Marshall, Trimmer, Houston and McNamara2013; Haselton et al., Reference Haselton, Nettle and Murray2015) posits that the incentive structures of ancestral decision-making tasks can support contemporary errors in belief and, by extension, errors in choice. The claim is that many decision-making domains in the ancestral past involved consistent variation in error costs. Cognition for a given domain evolved to process information from the environment in a way that helped ancestral humans avoid especially costly errors. Contemporary humans retain the psychologies in question. If environmental stimuli activate an associated cognitive process, the decision maker chooses in accord with ancestral conditions, and she does so even if the contemporary setting is quite different (Cosmides & Tooby, Reference Cosmides and Tooby2013). Because underlying beliefs are potentially distorted by psychologies adapted to ancestral conditions that may no longer hold, beliefs can be errors, as can attendant choices from both ex ante and ex post perspectives.

Finally, we consider the hypothesis that frequency-dependent social learning supports multiple equilibria and path-dependent cultural evolutionary dynamics (Boyd & Richerson, Reference Boyd and Richerson1985; Bikhchandani et al., Reference Bikhchandani, Hirshleifer and Welch1992; Young, Reference Young1996, Reference Young2015; Toyokawa et al., Reference Toyokawa, Whalen and Laland2019). When people learn socially by observing how common or rare different behaviours are, the path a population takes can depend sensitively on small random events that occur early in the cultural evolutionary process. Consequently, a tradition in which nearly everyone chooses in error, given their situation, can be a stable equilibrium that obtains with positive probability (Boyd & Richerson, Reference Boyd and Richerson1985; Bikhchandani et al., Reference Bikhchandani, Hirshleifer and Welch1992). This mechanism has been offered as an explanation for the persistence and even origin of many harmful traditions, including foot binding and female genital cutting (Mackie, Reference Mackie1996; Efferson et al., Reference Efferson, Vogt, Elhadi, Ahmed and Fehr2015, Reference Efferson, Vogt and Fehr2020; Platteau et al., Reference Platteau, Camilotti, Auriol, Anderson, Beaman and Platteau2018), child marriage (Cloward, Reference Cloward2016) and duelling (Young, Reference Young2015).

Importantly, social learning and path-dependent dynamics are perfectly consistent with a rational benchmark in which everyone has Bayesian beliefs and maximises expected payoffs given these beliefs (Bikhchandani et al., Reference Bikhchandani, Hirshleifer and Welch1992). Consequently, the mechanism requires neither unjustified beliefs nor sub-optimal decision making. Nonetheless, social learning could amplify or attenuate the effects of other mechanisms, and so social learning also does not preclude unjustified beliefs and sub-optimal decision making. One of our main tasks here is to examine interactions between social learning and other mechanisms, specifically mechanisms related to contemporary error costs and to psychologies evolved to manage ancestral error costs.

We present a model that combines the three mechanisms above, and we then show results from a behavioural experiment with a design that parallels the model closely. This jointly theoretical and empirical approach has key advantages. It requires us to provide precise operational definitions of the three mechanisms. This is especially critical with respect to contemporary vs. ancestral cost asymmetries as the distinction between these two mechanisms is not always clear (McKay & Efferson, Reference McKay and Efferson2010; Marshall et al., Reference Marshall, Trimmer, Houston and McNamara2013). Moreover, our modelling and empirical efforts jointly link the present study to other fields, some theoretical and some empirical. These include expected utility and cumulative prospect theories (von Neumann and Morgenstern, Reference von Neumann and Morgenstern1944; Kahneman and Tversky, Reference Kahneman and Tversky1979; Tversky & Kahneman, Reference Tversky and Kahneman2000), error management theory (Haselton & Nettle, Reference Haselton and Nettle2006; McKay & Efferson, Reference McKay and Efferson2010; Johnson et al., Reference Johnson, Blumstein, Fowler and Haselton2013; Haselton et al., Reference Haselton, Nettle and Murray2015), risk dominance in game theory (Harsanyi & Selten, Reference Harsanyi and Selten1988; Camerer, Reference Camerer2003), information cascades and herding (Bikhchandani et al., Reference Bikhchandani, Hirshleifer and Welch1992; Anderson & Holt, Reference Anderson and Holt1997; Chamley, Reference Chamley2004; Goeree et al., Reference Goeree, Palfrey, Rogers and McKelvey2007; Goeree & Yariv, Reference Goeree and Yariv2015), and the study of cultural evolution (Cavalli-Sforza & Feldman, Reference Cavalli-Sforza and Feldman1981; Sperber, Reference Sperber1996; Richerson & Boyd, Reference Richerson and Boyd2005; Mesoudi, Reference Mesoudi2011; Henrich, Reference Henrich2015; Morin, Reference Morin2016; Heyes, Reference Heyes2018). Finally, our joint approach ensures that we have a framework for interpreting our experimental results (Muthukrishna and Henrich, Reference Muthukrishna and Henrich2019). For example, as explained below, past implementations of the generic experimental paradigm we used (e.g. Anderson & Holt, Reference Anderson and Holt1997) led us to expect that social learning would routinely generate path-dependent dynamics and associated harmful traditions in our experiment. The model suggested a more subtle result. Our intuition was wrong; the model was not.

A model of beliefs and choices with explicit incentives, biased cognition and social learning

The environment takes one of two states, 0 or 1. The ex ante probability of state 1 is p 1 ∈ (0, 1). Individuals have two actions to choose from, which we also label as 0 and 1. Choosing 0 in state 0 is better than choosing 1, and choosing 1 in state 1 is better than choosing 0. Specifically, u 00 is the payoff from choosing 0 in state 0, u 10 is the payoff from choosing 1 in state 0, and u 00 > u 10. Similarly, u 11 is the payoff from choosing 1 in state 1, u 01 is the payoff from choosing 0 in state 1, and u 11 > u 01. This is the ‘explicit incentive structure’ of the decision-making task, which represents the first mechanism we consider.

Two errors are possible. A decision maker can choose 1 when the state is 0, with a loss of u 00 − u 10, or the decision maker can choose 0 when the state is 1, with a loss of u 11 − u 01. If error costs are the same, the explicit incentive structure is ‘symmetric’. If error costs are different, it is ‘asymmetric’. We arbitrarily designate state 1 as the state with relatively large error costs under asymmetric incentives, which means u 11 − u 01 ≥ u 00 − u 10. To illustrate asymmetric incentives, imagine a woman walking through the forest. She spies something long, dark and skinny. Is it a dangerous snake, equivalent to state 1, or just a stick, equivalent to state 0? Treating a snake as a stick is worse than treating a stick as a snake, and so the woman decides to tread carefully. Treading carefully does not require her to process snake-like stimuli in a special way, although we consider this possibility below. Rather, treading carefully simply requires the woman to recognise explicitly that snakes are more dangerous than sticks. A long dark skinny object in the forest is in this sense equivalent to the gamble above that Charles offers to Ryan. Put differently, the explicit incentive structure of a task at hand is a general mechanism that can shape choices via asymmetric costs in ways unrelated to specific decision-making domains and associated selection in the ancestral past.

Our treatment of cognition, in contrast, centres on hypotheses about past selection. Specifically, decision makers do not simply respond to explicit incentives and an ex ante probability; they also learn. Learning may or may not yield beliefs in error depending on the structure of evolved cognition. This is the second mechanism we examine.

Before choosing, each decision maker observes a private signal about the state. She interprets this signal and learns accordingly. Like the state, a signal can take a value of 0 or 1. After observing her signal, a decision maker updates her beliefs about the environment via Bayes’ Rule. Beliefs take the form of a subjective probability that the state is 1. If the decision maker observes 0 as a private signal, her belief goes down. If she observes 1, her belief goes up. Because the signal is private, associated learning is individual. With Bayesian updating, we do not suggest that real people routinely perform Bayesian calculations. Rather, whatever the actual cognitive processes people use, Bayesian beliefs provide the unbiased benchmark. Using Bayes’ Rule in the model thus allows us to posit a precise operational definition of beliefs formed via biased vs. unbiased cognition.

In particular, decision makers interpret private signals in a way that may or may not be accurate. Decision makers think that private signals match the state with probability $\hat{q}\in \lpar {0.5\comma \;1} \rpar$. The actual probabilities that govern perceived private signals may be different. If so, decision makers systematically misinterpret signals, and we thus refer to a ‘cognitive bias’. We focus on cognitive biases that distort beliefs in favour of state 1. If the state is 0, the actual probability a decision maker perceives a 0 signal is $\hat{q}-\alpha$ for some $\alpha \in \lsqb {0\comma \;\hat{q}} \rsqb$. If the state is 1, the actual probability that a decision maker perceives a 1 signal is $\hat{q} + \beta$ for some $\beta \in \lsqb {0\comma \;1-\hat{q}} \rsqb$. To see why this represents a cognitive bias favouring belief in state 1, consider the extreme case in which $\alpha = \hat{q}$ and $\beta = 1-\hat{q}$. Perceived signals indicate state 1 regardless of the actual state, and thus private signals are completely uninformative. However, because decision makers are unaware of their bias, they interpret perceived signals as evidence for state 1.

An archetypical scenario illustrates. A contemporary human is walking through the forest, and she spies something long, dark, and skinny. Is it a dangerous snake, equivalent to state 1, or just a stick, equivalent to state 0? Ancestral error costs were presumably asymmetric because treating a snake as a stick in the ancestral past was worse than treating a stick as a snake (Haselton & Galperin, Reference Haselton and Galperin2012). Crucially, however, we considered explicit incentives above as the first of our three mechanisms, and these incentives, whether ancestral or contemporary, are not our concern here. Rather, our concern is the hypothesis that ancestral cost asymmetries led to the evolution of a cognition that distorts contemporary belief formation. Such a cognition would imply that contemporary humans are hypersensitive to snake-like stimuli and draw inferences about the threat of dangerous snakes that are not fully justified by the evidence at hand. Formally, α, β > 0, and thus beliefs deviate systematically from the unbiased Bayesian benchmark. Many cognitive biases of this sort have been hypothesised owing to ancestral cost asymmetries (Abbey, Reference Abbey1982; Haselton & Nettle, Reference Haselton and Nettle2006; McKay & Efferson, Reference McKay and Efferson2010; Delton et al., Reference Delton, Krasnow, Cosmides and Tooby2011; Johnson et al., Reference Johnson, Blumstein, Fowler and Haselton2013; Marshall et al., Reference Marshall, Trimmer, Houston and McNamara2013; Haselton et al., Reference Haselton, Nettle and Murray2015; Perilloux & Kurzban, Reference Perilloux and Kurzban2015; Zimmermann & Efferson, Reference Zimmermann and Efferson2017; Murray et al., Reference Murray, Murphy, von Hippel, Trivers and Haselton2017), and our experiment below considers a specific bias known as ‘hyperactive agency detection’ (Guthrie et al., Reference Guthrie, Agassi, Andriolo, Buchdahl, Earhart, Greenberg and Tissot1980; Guthrie, Reference Guthrie1993; Barrett, Reference Barrett2000, Reference Barrett2004, Reference Barrett2012; McKay et al., Reference McKay, Ross, O'Lone and Efferson2018; Maij et al., Reference Maij, van Schie and van Elk2019).

As the final mechanism, decision makers learn socially by observing others. They make choices one at a time in a sequence indexed by t. At any given point in the sequence, the decision maker in question has a prior belief that the state is 1, which we call $\hat{p}_t$. The decision maker observes her private signal and updates her belief. Given the explicit incentive structure and her updated belief, the decision maker has an expected payoff from choosing 0 and an expected payoff from choosing 1. She then makes a choice. Decision makers tend to choose in a way that maximises expected payoffs, but they do not do so with certainty. The parameter λ ∈ [0, ∞) controls how strongly choices respond to expected payoffs. For low values of λ, choices are relatively noisy and only somewhat responsive to expected payoffs. For high values of λ, decision makers are extremely responsive and almost always maximise expected payoffs, which means decision making is relatively systematic.

Social learning occurs because choices are observable. For the first decision maker, at position t = 1, her subjective prior is the ex ante objective prior, $\hat{p}_1 = p_1$. The decision maker observes her private signal, updates her beliefs, and makes a choice observable to everyone else. All downstream decision makers (t > 1) update their beliefs after observing the choice at t = 1. The ex ante objective prior (p 1), the incentive structure (u 00, u 10, u 11 and u 01), $\hat{q}$ and λ are all common knowledge, and so everyone downstream updates beliefs in the same way. The result is $\hat{p}_2$, which we treat as a new prior for the decision maker in position t = 2. The decision maker at t = 2 observes her private signal, updates her beliefs accordingly, and makes an observable choice. Downstream decision makers update their beliefs based on this observable choice, and so the process goes. In sum, the first decision maker only learns individually, while all subsequent decision makers learn both individually and socially. Importantly, if cognition is biased (α, β > 0), no one is aware of the bias. Decision makers do not account for the bias when updating their beliefs by learning individually via private signals, and they do not account for the bias when learning socially by observing the choices of others.

To avoid any confusion, we would like to explain our use of the word ‘bias’ when referring to social learning. We do not mean biased social learning in the sense of Boyd and Richerson (Reference Boyd and Richerson1985), who used the term ‘biased’ to mean any social learning strategy that generates endogenous cultural evolutionary dynamics. When we say ‘biased’, we mean a cognitive system that processes information in some way that deviates systematically from a Bayesian with an accurate understanding of priors, private information and observed choices. The two views of bias are not at odds; they simply emphasise different questions. Interestingly, recent theoretical research has shown that genetic evolution can support social learning strategies that are both consistent with Bayesian updating and generate endogenous cultural evolutionary dynamics (Perreault et al., Reference Perreault, Moya and Boyd2012; Efferson et al., Reference Efferson, Lalive, Cacault and Kistler2016). Such strategies are biased in the sense of Boyd and Richerson (Reference Boyd and Richerson1985) but unbiased by our definition.

Model results, analytical

Cost asymmetries in the explicit incentive structure exert a powerful influence on choice by weakening the belief a decision maker requires before choosing 1. Moreover, this mechanism does not require a large asymmetry. Indeed, the largest effects on behaviour occur when moving from no asymmetry to small asymmetries (Supplementary Information). Moreover, the potency of explicit cost asymmetries has nothing to do with the origin of beliefs. Explicit cost asymmetries exert their considerable influence on behaviour regardless of whether beliefs are prior to learning or posterior, and regardless of whether or not beliefs are distorted by cognitive bias. Cost asymmetries mean that decision makers require a relatively weak belief that the state is 1 before choosing 1, and this claim is independent of how beliefs are formed.

Belief formation is a separate process, and we show that beliefs can only evolve in the wrong direction if cognition is biased in a sufficiently strong way (Supplementary Information). Specifically, if cognition is unbiased (α = β = 0), beliefs evolve in expectation in the right direction. If the state is 0, beliefs that the state is 1 are expected to go down. If the state is 1, beliefs that the state is 1 are expected to go up. If cognition is biased (α, β > 0), beliefs may or may not evolve in expectation in the right direction. Because we focus on cognitive biases that distort beliefs in favour of state 1, belief evolution when the state is 1 is not especially interesting. The associated cognitive bias (β > 0) may speed up the evolution of beliefs in favour of state 1, but it cannot send belief evolution off in the wrong direction. Information from the environment and the cognitive bias point towards the same conclusion.

The interesting scenario centres on the evolution of beliefs in favour of state 1 when the actual state is 0. We show that beliefs evolve in expectation in favour of state 1, and hence in the wrong direction, if and only if $\hat{p}_t \lt \alpha /\lpar {2\hat{q}-1} \rpar$. In effect, beliefs are expected to evolve consistently away from reality if and only if decision makers think that private signals are much noisier than these signals really are. For example, if $\hat{q} = 0.6$, decision makers think that private signals are relatively noisy, α can take any value in [0, 0.6], and $2\hat{q}-1 = 0.2$. By extension, for any feasible α ≥ 0.2, $\alpha /\lpar {2\hat{q}-1} \rpar \ge 1$. Consequently, $\hat{p}_t \lt \alpha /\lpar {2\hat{q}-1} \rpar$ always holds, and thus beliefs favouring state 1 always increase in expectation. In contrast, if $\hat{q} = 0.9$, α can take any value in [0, 0.1], and $2\hat{q}-1 = 0.8$. Even if α = 0.1, the maximum feasible value, beliefs favouring state 1 only increase in expectation if the prior is sufficiently small, namely if $\hat{p}_t \lt \alpha /\lpar {2\hat{q}-1} \rpar = 1/8$. Here, the effects of the cognitive bias are fundamentally limited. The cognitive bias prevents beliefs from converging on the truth, but the bias does not systematically distort belief evolution in the wrong direction. For most priors ($\hat{p}_t \gt 1/8$), beliefs are expected to move in the right direction, namely towards zero, even though cognition is biased in the other direction.

All in all, our model isolates $\alpha /\lpar {2\hat{q}-1} \rpar$ as an important measure of cognitive bias. If $\alpha /\lpar {2\hat{q}-1} \rpar$ is sufficiently large, beliefs are always expected to evolve in the wrong direction when reality is inconsistent with the bias. Alternatively, if $\alpha /\lpar {2\hat{q}-1} \rpar$ is sufficiently small, beliefs are only expected to evolve in the wrong direction when the decision maker's prior is sufficiently close to the truth. Otherwise, beliefs are distorted by the cognitive bias, but they still evolve in expectation in the right direction.

Model results, simulation

To provide a complete depiction of how beliefs and behaviour evolve, we developed an agent-based simulation of the exact model detailed above, and we ran simulations under a diverse array of parameter values (Supplementary Information). In particular, we varied the explicit incentive structure (u 00, u 10, u 11 and u 01) and the properties of cognition (α and β). Like the analysis above, we arbitrarily limit attention to associated biases in favour of 1. This means, if error costs are asymmetric, the asymmetry favours choosing 1 (u 11 − u 01 > u 00 − u 10). We varied the incentive structure to range from no asymmetry to large asymmetries. As explained above, however, the difference between symmetric error costs and small asymmetries is the difference that matters most. In addition, if cognition is biased, the bias distorts beliefs in favour of 1 (α, β > 0). We specifically set $\hat{q} = 0.6$ and varied α and β over the full range of possible values. As explained above, any associated distortions in information processing may or may not be strong enough to lead beliefs to evolve in the wrong direction under state 0.

Finally, we also varied how strongly choices respond to expected payoffs (λ), and we varied the ex ante probability of state 1 (p 1). For every combination of parameter values, we simulated 100 independent sequences of 201 decision makers each. Below we provide a link to the files for the simulation, a script for managing the entire project over a user-defined parameter space, and a script for graphing results for each parameter combination.

Here we focus on the interesting case in which state 0 is the most likely state ex ante and the actual state ex post. Specifically, we consider cases in which p 1 = 1/3, and 0 is the actual state, which is expected to happen for 2/3 of all simulated populations. This scenario is interesting for two key reasons. First, if the actual state is 1, cognitive biases and explicit cost asymmetries can only reinforce the evolution of beliefs and choices in a way that is consistent with the actual environment. The tendency for cognitive biases and explicit cost asymmetries to support errors, in contrast, hinges on the environment being in the opposite state, and this is why we focus on cases in which the actual state is 0. Second, by making state 1 ex ante unlikely (p 1 = 1/3), populations that converge on choosing 1 in state 0 are converging on a behavioural tradition that is not just an error, but an error with a relatively high ex ante probability. As explained later, we chose p 1 = 1/3 in our experiment for exactly this reason.

Figures 1–3 show the evolution of beliefs and choices under these conditions. We use bubble plots because they allow us to depict the complete distribution of outcomes over all relevant simulations, and the graphs thus provide complete information about simulation results. To read the graphs, take Figure 1a as an example. First consider beliefs. The prior belief for a given decision maker can take values in [0, 1]. Accordingly, we partition this interval into 10 bins, {[0, 0.1], (0.1, 0.2], (0.2, 0.3], …, (0.9, 1]}. For a given position (i.e. a decision maker in the sequence), we calculate the distribution over the prior beliefs of the decision makers in that position, with one decision maker per simulation, and show that distribution as a bubble plot in open blue circles. Bubbles are centred for each of the 10 bins, and the sizes of bubbles are proportional to the frequency of observations for the bin in question. For example, all decision makers in position 1 have the ex ante objective prior as their subjective priors ($\hat{p}_1 = p_1 = 1/3$), and so this position has one large open circle at the centre of the interval (0.3, 0.4]. To prevent clutter, we only show the distribution of priors for every 10 positions in the sequence, and distributions are offset slightly to the left relative to the sequence position in question.

Figure 1. The evolution of beliefs and choices when error costs are symmetric (u 11 = u 00 = 1 and u 01 = u 10 = 0). For each panel, we simulated 100 independent sequences with 1/3 as the ex ante probability of state 1. Each panel shows results for the specific sequences in which the actual state is 0. Over these sequences, the graph shows the distribution of prior beliefs that the state is 1 (open circles) for every tenth decision maker in the sequence. It also shows the associated distribution over the cumulative proportions, by sequence, of decision makers incorrectly guessing state 1 (closed circles). Distributions are represented as bubble plots. (a) Cognition is biased (α = 0.5) and decision making relatively noisy (λ = 10). (b) Cognition is biased (α = 0.5) and decision making relatively systematic (λ = 100). (c) Cognition is unbiased (α = 0) and decision making relatively noisy (λ = 10). (d) Cognition is unbiased (α = 0) and decision making relatively systematic (λ = 100). See the main text for a detailed description of how to read the graphs and a summary of key results.

Figure 2. The evolution of beliefs and choices when error costs involve a relatively weak asymmetry (u 11 = 1, u 00 = 0.75, u 01 = 0, and u 10 = 0.25). For each panel, we simulated 100 independent sequences with 1/3 as the ex ante probability of state 1. Each panel shows results for the specific sequences in which the actual state is 0. Over these sequences, the graph shows the distribution of prior beliefs that the state is 1 (open circles) for every tenth decision maker in the sequence. It also shows the associated distribution over the cumulative proportions, by sequence, of decision makers incorrectly guessing state 1 (closed circles). Distributions are represented as bubble plots. (a) Cognition is biased (α = 0.5) and decision making relatively noisy (λ = 10). (b) Cognition is biased (α = 0.5) and decision making relatively systematic (λ = 100). (c) Cognition is unbiased (α = 0) and decision making relatively noisy (λ = 10). (d) Cognition is unbiased (α = 0) and decision making relatively systematic (λ = 100). See the main text for a detailed description of how to read the graphs and a summary of key results.

Figure 3. The evolution of beliefs and choices when error costs involve a relatively strong asymmetry (u 11 = 1, u 00 = 0.6, u 01 = 0, and u 10 = 0.4). For each panel, we simulated 100 independent sequences with 1/3 as the ex ante probability of state 1. Each panel shows results for the specific sequences in which the actual state is 0. Over these sequences, the graph shows the distribution of prior beliefs that the state is 1 (open circles) for every tenth decision maker in the sequence. It also shows the associated distribution over the cumulative proportions, by sequence, of decision makers incorrectly guessing state 1 (closed circles). Distributions are represented as bubble plots. (a) Cognition is biased (α = 0.5) and decision making relatively noisy (λ = 10). (b) Cognition is biased (α = 0.5) and decision making relatively systematic (λ = 100). (c) Cognition is unbiased (α = 0) and decision making relatively noisy (λ = 10). (d) Cognition is unbiased (α = 0) and decision making relatively systematic (λ = 100). See the main text for a detailed description of how to read the graphs and a summary of key results.

Now consider choices. For a given point in the sequence, t, we calculate the cumulative proportion of decision makers choosing 1 for each simulated sequence. Let c n ∈ {0, 1} denote the choice of the decision maker choosing in position n for a specific sequence. The cumulative proportion choosing 1 is simply ${\sum _{n = 1}^{t}} c_n / t$. For a given t, we have multiple cumulative proportions, one for each simulated sequence. A bubble plot in closed blue circles represents the distribution of these values over sequences. As above, we partition the unit interval into 10 bins and show the distribution of cumulative proportion values over these 10 bins. To illustrate with t = 1 in Figure 1a, most simulations under state 0 have cumulative proportions in [0, 0.1], but a few also have cumulative proportions in (0.9, 1]. We show the distribution for every 10 positions in the sequence, and distributions are offset slightly to the right relative to the sequence position in question.

Simulations show four key results.

  1. (1) The dynamics of both beliefs and choices unfold slowly when choices respond strongly to expected payoffs and are thus relatively systematic (Figures 1–3, panels b and d, λ = 100). To see why, consider the extreme case in which all decision makers always maximise expected payoffs (λ → ∞). When this is true, decision makers start off choosing in perfect accord with their private signals. Downstream decision makers can thus infer, with complete accuracy, the private signals of the decision makers in question. Prior beliefs evolve towards one of the two boundaries as the sample of signals and congruent choices grows. Before long, however, prior beliefs become so strong that the weight of history exceeds the informational value of a private signal. At this point, the choice maximising expected payoffs is independent of the signal the decision maker observes (Bikhchandani et al., Reference Bikhchandani, Hirshleifer and Welch1992, Reference Bikhchandani, Hirshleifer and Welch1998). All learning stops because choices no longer reveal private information. For finite but large values of λ (Figures 1–3, panels b and d, λ = 100), learning never stops in this way (Goeree et al., Reference Goeree, Palfrey, Rogers and McKelvey2007), but it is slow. It is slow because, even though choices are not independent of signals, probability distributions over choices are nonetheless highly skewed. Observing a choice thus conveys some information but only a little (Cover & Thomas, Reference Cover and Thomas2006). With more noise (Figures 1–3, panels a and c, λ = 10), probability distributions are less skewed, and observed choices convey more information. This speeds up learning and associated cultural evolutionary dynamics.

  2. (2) When explicit error costs are symmetric, beliefs and choices are congruent. If the belief in state 1 is high, choosing 1 is common (Figure 1a). If the belief in state 1 is low, choosing 1 is uncommon (Figure 1b–d). When error costs are asymmetric, beliefs and choices can be incongruent. Choosing 1 can be common even when the belief in state 1 is low (Figures 2b, d and 3bd).

  3. (3) Beliefs only evolve consistently in the wrong direction, namely away from 0, when cognition is biased (Figures 1a, 2a, b and 3a). Here we implement a cognitive bias that we know, given our analytical results above, is a strong bias because $\alpha /\lpar {2\hat{q}-1} \rpar = 0.5/\lpar {2\lpar {0.6} \rpar -1} \rpar = 2.5$.

  4. (4) Social learning can generate path-dependent dynamics, but it has no general tendency to do so. Figure 2c, d shows specific situations that support path dependence. Some sequences converge on nearly everyone correctly choosing 0, while other sequences converge on nearly everyone incorrectly choosing 1. For these figures, the cost of an error if the state is 1 is twice the cost of an error if the state is 0. When coupled with the fact that state 1 is half as likely as state 0 (1/3 vs. 2/3), the explicit incentive structure creates no initial bias towards choosing 0 or 1. In addition, cognition is unbiased, and so decision makers do not process information in a systematically distorted way. As a result, both the explicit incentive structure and cognition are neutral with respect to choice. Because incentives and cognition are neutral in this way, the tendency for social learning to generate path-dependent dynamics in behaviour can rise to the surface. In contrast, if explicit incentives are not neutral (Figure 1a–d), or if cognition is biased (Figures 13, a and b), sequences do not exhibit this path dependence, and all sequences tend to evolve towards most decision makers choosing either 0 or 1.

Experimental methods

We conducted behavioural experiments in which subjects made decisions under risk in a shared environment (Supplementary Information). This shared environment took one of two possible states, labelled simply as ‘red’ and ‘blue’. The actual state was not known with certainty, and the basic experimental task was to guess the state correctly. Errors amounted to either guessing blue when the state was red or guessing red when the state was blue. Both errors cost subjects real money. To draw a link with the model above, blue is state 0, and red is state 1.

For a given experimental session, we typically had 34 decision-making subjects and one additional, randomly selected subject who served as the monitor. With probability 1/3 subjects faced the red state, while with probability 2/3 the state was blue. Red was thus ex ante unlikely. Accordingly, a group in which red choices spread under the blue state was a group generating a shared tradition expected to be an error with a relatively high ex ante probability.

In each experimental session we repeated the experiment five times. Treatment conditions were held constant for all repetitions within a session, and each repetition proceeded as follows. The monitor rolled a die to determine the state. A roll of 1 or 2 resulted in the red state, while 3, 4, 5 or 6 resulted in the blue state. The monitor rolled the die while out of sight but within earshot of the other subjects, and all subjects knew the process for determining the state. After rolling the die, the monitor returned to her computer, entered the realised state for the remaining 34 subjects, and the experiment proper began. The experimenter observed all of the monitor's activities to ensure accuracy, and the monitor's payoff did not depend in any way on the realised state or the choices of the other participants.

After the monitor had entered the state, the remaining subjects made their guesses one at a time in a randomly determined order that was independent for each repetition of the experiment. Before making a guess, each subject received a private signal only she could observe. This private signal matched the realised state with probability 0.6, and thus private signals were informative but noisy.

After observing her private signal, each decision-making subject made a guess about the state. Subjects received an increase in payoffs for correct guesses and a decrease for incorrect guesses. As explained below, in some treatments guesses were publicly displayed immediately. As a result, downstream decision makers, in addition to relying on their private signals, could learn socially. In other treatments, guesses were not publicly observable, subjects had only their private signals, and thus learning was strictly individual. Our overall experimental design included eight treatments based on variation in explicit incentives, the presence of a payoff-irrelevant prime intended to induce biased cognition and whether social learning was possible.

Explicit incentives

In the symmetric case, each subject received an endowment of 8 CHF (Swiss Francs). An incorrect guess about the state resulted in a loss of 3 CHF, while a correct guess resulted in a gain of 3 CHF. These gains and losses held regardless of whether the realised state was red or blue. In the asymmetric case, endowments were the same, and the losses and gains in the blue state were the same. If the realised state was red, however, a correct guess resulted in a gain of 6 CHF, while an incorrect guess led to a loss of 6 CHF. This means that the cost of an error in the red state was 6 − ( − 6) = 12 CHF, which was twice as much as the error cost of 3 − ( − 3) = 6 CHF when the state was blue.

Our asymmetric treatments involved a weak asymmetry in which the cost of choosing blue when the state was red was only twice the cost of choosing red when the state was blue. Moreover, we exactly offset this asymmetry by making blue twice as likely as red ex ante. The net result was that, in treatments with an asymmetric payoff structure, the a priori expected payoff from choosing red ((1/3)(6) + (2/3)( − 3) = 0) exactly equalled the a priori expected payoff from choosing blue ((1/3)( − 6) + (2/3)(3) = 0). Put differently, before the first subject in a sequence had received her private signal, the cost asymmetry did not produce an initial bias favouring blue or red for risk-neutral subjects. Once subjects started to receive information and learn, this equivalence no longer held.

Agency prime

Treatments varied in terms of how we described payoffs. In no agency prime treatments, the payoff consequences associated with a correct guess were explained in the instructions and on the decision-making screen during the experiment by saying, for example, ‘Your income rises by 3 CHF’. With an incorrect guess, we analogously said, ‘Your income falls by 3 CHF’. This frame was used for both the red and blue states. In agency prime treatments, the frame was the same for describing the correct and incorrect guesses under the blue state. However, when describing the payoff consequences under the red state, with the symmetric case as an example, we said, ‘We will reward you with an increase of 3 CHF’ or ‘We will punish you with a reduction of 3 CHF’. For screen shots in the original German, see the Supplementary Information (Figures S3 and S4).

Our aim here was to use a linguistic manipulation to activate the concept of an intentional agent associated with a specific environmental state. The hypothesised cognitive bias of interest is hyperactive agency detection (Guthrie et al., Reference Guthrie, Agassi, Andriolo, Buchdahl, Earhart, Greenberg and Tissot1980; Guthrie, Reference Guthrie1993; Barrett, Reference Barrett2000, Reference Barrett2004, Reference Barrett2012; McKay et al., Reference McKay, Ross, O'Lone and Efferson2018; Maij et al., Reference Maij, van Schie and van Elk2019). The hypothesis posits that one of the biggest threats ancestral humans faced was other people with furtive, malevolent intentions. Ancestral humans faced two associated errors. They could have assumed an unseen agent was trying to harm them when no such agent was present, or they could have ignored the possibility of an unseen hostile agent when in fact one did exist. The latter error was typically more costly, and this would have increased the tendency to guard preemptively against the hazards of unseen agents. In the end, humans evolved a cognitive bias that discounts the role of chance and overestimates the probability that unseen agents are responsible for many events in life, from the mundane to the extraordinary. Metaphorically, people do not simply see clouds; they see faces in the clouds (Guthrie, Reference Guthrie1993).

For our purposes, guessing the state is theoretically equivalent to guessing if an unseen agent is nearby. Under one state, in our case blue, no unseen agent is nearby and costs and benefits simply occur. Under the other state, in our case red, an agent is nearby to discharge costs and benefits as punishments and rewards. This is why we attach the agency prime to the red state. Previous experiments have successfully used both word primes (Shariff & Norenzayan, Reference Shariff and Norenzayan2007; Gervais & Norenzayan, Reference Gervais and Norenzayan2012) and face primes (Haley & Fessler, Reference Haley and Fessler2005; Nettle et al., Reference Nettle, Harper, Kidson, Stone, Penton-Voak and Bateson2013; Sparks & Barclay, Reference Sparks and Barclay2013) to activate agency concepts, and linguistic priming effects have been widely documented in behavioural experiments (Tversky & Kahneman, Reference Tversky and Kahneman1981). The question is whether this manipulation shapes choices in a way that is distinct from the effects of explicit material incentives. As explained in our theory section above, such an effect would reflect our operational definition of a cognitive bias.

Individual and social learning

In social treatments, guesses were posted in order, as they occurred, across the top of every subject's screen using either an ‘R’ or a ‘B’. In asocial treatments, the character ‘X’ appeared instead, regardless of whether the relevant decision was red or blue.

Overall, our empirical strategy was to design an experiment that captures the potential effects of three mechanisms in all eight combinations. The three mechanisms are asymmetric error costs associated with contemporary explicit incentives, evolved cognitive biases owing to ancestral cost asymmetries and path-dependent cultural evolutionary dynamics. Each mechanism represents a distinct hypothesis about the origin and persistence of errors, and our model suggests that all three should affect if and how decision makers exhibit errors given environmental states.

First, for subjects maximising expected payoffs, treatments including an explicit cost asymmetry should reduce the belief in the red state a subject requires before actually guessing red, which should increase red choices all else equal. Second, treatments using agency language to describe outcomes under the red state should evoke a psychology to mollify the unseen agents who distribute punishments and rewards if the state is red, which should also increase red choices all else equal. Finally, under publicly observable choices, social learning should lead to path-dependent dynamics when other mechanisms are neutral and by extension a subset of populations that converge on a harmful tradition. The neutrality of other mechanisms holds in asymmetric treatments without an agency prime.

We conducted experiments under anonymous laboratory conditions on a local computer network running z-Tree (Fischbacher, Reference Fischbacher2007) in the Department of Economics at the University of Zurich. We ran sequences of length 34 because this was the maximum length we could implement given the size of the laboratory. Altogether, we ran 20 sessions with 670 subjects, recruited via the laboratory's standard subject pool, for a total of 3365 observations over 100 separate sequences (Supplementary Information). The final sample was 47.5% female with an average age of 22.1 (SD 4.45). Excluding the monitors, subjects made an average of 43.89 CHF in the experiment, and they additionally received 10 CHF each as a show-up fee. Monitors received fixed total payments of 50 CHF. The study was approved by the Human Subjects Committee of the Faculty of Economics, Business Administration, and Information Technology at the University of Zurich. We did not pre-register the study because we collected the data before recent replication studies (Open-Science-Collaboration, 2015; Camerer et al., Reference Camerer, Dreber, Forsell, Ho, Huber, Johannesson and Wu2016) and the trend towards pre-registration that followed.

Experimental results

For all analyses, whether modelling binary or continuous response variables, we rely on ordinary least squares with robust clustered standard errors as a robust approach to estimating average treatment effects under minimal assumptions (Angrist & Pischke, Reference Angrist and Pischke2009). We conservatively cluster at the session level, which yields 20 clusters. Accordingly, we calculate both the variance-covariance matrix allowing for heteroskedastic errors correlated within clusters (Wooldridge, Reference Wooldridge2002; Arai, Reference Arai2011), and we use clustered bootstrapping to obtain 95 and 99% confidence intervals (Cameron & Trivedi, Reference Cameron and Trivedi2005; Angrist & Pischke, Reference Angrist and Pischke2009). We emphasise results robust to multiple approaches to statistical inference. As additional robustness checks, we also model treatment effects both with and without controls. We provide a link to the data below.

Red choices

Figure 4 shows the proportion of red guesses conditional on the treatment and the realised state. The figure reveals that explicitly asymmetric error costs had an overwhelmingly dominant effect. Modelling red choices as a function of the treatments confirms this conclusion by showing that asymmetric error costs produced a large and highly significant increase in the rate at which participants guessed red (Table 1, Asym). This means that asymmetric error costs reduced the error rate when the state was red, which was relatively rare, and increased the error rate when the state was blue, which was relatively common. The other treatment dimensions did not robustly affect the rate of red guesses. The results provide some suggestive evidence that the availability of social information slightly increased the probability of participants choosing red, but the effect is not robust to multiple forms of statistical inference (Table 1, Social). In addition, observing a red private signal had a strong and robust positive effect on the probability of making a red choice (Table 1, Signal red).

Figure 4. Red choices by treatment and realised state. The proportion of red choices for (a) asocial treatments and (b) for social treatments. The colour of the bars signifies the realised state. Consequently, the red bars represent correct choices in the relatively rare case of a red state, while the blue bars represent errors in the relatively common case of a blue state. These results show that explicit cost asymmetries had an overwhelmingly dominant effect on average choices (see Table 1).

Table 1. Red choices in all treatments. Linear probability models with red choices as the response variable and robust clustered standard errors calculated by clustering on session. In addition, the table shows 95% and 99% confidence intervals calculated with a non-parametric bootstrap clustered at the session level. Independent variables include a dummy for the realised environment for the sequence (Env red), the realised private signal (Signal red), the subject's gender, order in the sequence and treatment dummies (Asym, Agency prime, Social). Columns 2–4 are for models that only include treatments as independent variables. Columns 5–7 add controls.

*(0.05); **(0.01); ***(0.001).

Although the availability of social information did not have a robust effect on the average tendency to choose red, this does not mean that social learning was unimportant. Rather, the result simply means that the possibility to learn socially did not affect average behaviour. For any given social learner, however, the social information available might still have affected her decision making, and before running the experiments we expected social information to influence choices. Indeed, a long tradition of research has shown that people exhibit some tendency to conform when presented with social information about how common or rare different behaviours are (Sherif, Reference Sherif1936; Asch, Reference Asch1955; Anderson & Holt, Reference Anderson and Holt1997; Berns et al., Reference Berns, Chappelow, Zink, Pagnoni, Martin-Shurski and Richards2005; Morgan et al., Reference Morgan, Rendell, Ehn, Hoppitt and Laland2012; Goeree & Yariv, Reference Goeree and Yariv2015; Efferson et al., Reference Efferson, Lalive, Cacault and Kistler2016; Muthukrishna et al., Reference Muthukrishna, Morgan and Henrich2016; Efferson & Vogt, Reference Efferson and Vogt2018). We examined such effects by analysing choices in social treatments as a function of frequency-dependent social information. In particular, because our paradigm relies on sequential choices with one choice per subject in a given environment, this analysis avoids the interpretive problems that plague many attempts to identify the causal impact of information about how others behave (Manski, Reference Manski2000; Angrist, Reference Angrist2014).

Accordingly, Table 2 shows an analysis of red choices in social treatments, where we have added lagged frequency-dependent social information to the independent variables. To specify this variable, let c n = 1 denote a red choice in position n of a sequence and c n = 0 a blue choice. For any position t > 1, lagged social information is the centred proportion of upstream subjects choosing red, $\mathop \sum \nolimits_{n = 1}^{t-1} c_n/\lpar {t-1} \rpar -0.5$. Like the analysis of all treatments (Table 1), the analysis of social treatments shows large and robust effects associated with asymmetric costs and the participant's private signal (Table 2, Asym and Signal red). Asymmetric costs and observing a red private signal both resulted in large and robust increases in the probability a participant chose red. The proportion of observed upstream participants choosing red also had a robust positive effect on the probability of a red choice (Table 2, Lagged social info).

Table 2. Red choices in social treatments with frequency-dependent social information. Linear probability models with red choices as the response variable and robust clustered standard errors calculated by clustering on session. In addition, the table shows 95% and 99% confidence intervals calculated with a non-parametric bootstrap clustered at the session level. Independent variables include a dummy for the realised environment for the sequence (Env red), the realised private signal (Signal red), the subject's gender, order in the sequence, the centred cumulative proportion choosing red through the previous period (Lagged social info), and relevant treatment dummies (Asym, Agency prime).

*(0.05); **(0.01); ***(0.001).

Altogether, our results indicate that frequency-dependent social information affected individual decision making, but it did so without affecting average choices. This result suggests that social learning might have instead affected the variance in choices. Put differently, in comparison to asocial treatments, social treatments might have shifted some of the overall variation in choices from within sequences to between sequences. Such a result would reduce the variation in choices within sequences, and it would be consistent with the hypothesis that social learning supports path-dependent dynamics (Young, Reference Young1996, Reference Young2015; Bowles, Reference Bowles2004). To test the idea, we now turn to an analysis of the variance in choices by sequence.

Variance in choices within sequences

Frequency-dependent social learning can generate path-dependent cultural evolutionary dynamics. Path-dependent dynamics, in turn, can generate an important aggregate pattern in which groups differ from one another, but choices within groups are relatively homogeneous (Young, Reference Young2015). Whether this aggregate-level pattern occurs, however, can be extremely sensitive to the details of how heterogeneous decision makers respond to social information (Granovetter, Reference Granovetter1978; Young, Reference Young2009; Efferson et al., Reference Efferson, Vogt and Fehr2020). Specifically, conformist social learning at the individual level may or may not translate into path-dependent dynamics at the aggregate level. The most direct route to examining this question is to analyse outcomes directly at the aggregate level (Efferson & Vogt, Reference Efferson and Vogt2018).

To do so, we treated each sequence of choices as a sample from a Bernoulli distribution and modelled sample variance by sequence as a function of the treatments. Given that our generic experimental paradigm has proven quite conducive to path dependence in the past (Anderson & Holt, Reference Anderson and Holt1997), we initially imagined that social sequences would be uniformly more homogeneous than asocial sequences. Our experimental results turned out to be more subtle than this in a way that was consistent with our model. We found that choices were extremely homogeneous, but without path dependence, in asocial sequences with symmetric error costs. For these treatments, blue choices predominated, and adding social information to the mix did not increase homogeneity because choices were already extremely homogeneous. In contrast, social learning led to path-dependent cultural evolution and an associated increase in homogeneity when explicit error costs were asymmetric. As explained below, the incentive structure in asymmetric treatments was relatively neutral with respect to choice, and choices within asocial sequences were correspondingly heterogeneous. Adding social information could then generate path-dependent dynamics and an associated increase in homogeneity.

Figure 5 shows the choice dynamics for all treatments, and Table 3 shows an analysis of the variance in choices within sequences. In asocial treatments with symmetric costs, the symmetry of error costs did not offset the fact that blue was twice as likely as red ex ante. All sequences converged on blue (Figure 5a), which resulted in homogeneity without path dependence. This left little scope for social information to homogenise choices further (Figure 5c vs. Figure 5a), and regression results show no effect (Table 3, Social and Social × AgencySym).

Figure 5. Choice dynamics for all treatments. Let cn = 0 denote a blue choice in sequence position n and cn = 1 a red choice. Given position t, the graphs show the cumulative proportion choosing red by sequence, ${\sum _{n = 1}^{t}} c_n / t$, as a function of sequence position, t, for all sequences in the experiment. The colour of the line shows the realised state for the sequence in question. Solid lines are for sequences in no agency prime treatments, while dashed lines are for sequences in agency prime treatments. Panels (a) and (b) show asocial treatments, while (c) and (d) show social treatments. Panels (a) and (c) show treatments with symmetric error costs, while (b) and (d) show asymmetric error costs. Social learning led to path-dependent dynamics and an associated increase in homogeneity within sequences when error costs were asymmetric.

Table 3. Ordinary least squares models with the variance in choices by sequence as the response variable and robust clustered standard errors calculated by clustering on session. The table also shows 95% and 99% confidence intervals calculated with a non-parametric bootstrap clustered at the session level. Independent variables include a dummy for the realised environment by sequence (Env red), a dummy for treatments that allowed social learning (Social), and dummies for the four treatment combinations involving the agency prime and the explicit payoff structure. We used combined dummies for these four combinations in order to avoid three-way interactions. The dummies are defined according to the presence (Agency) or absence (NoAgency) of the agency prime and either symmetric (Sym) or asymmetric (Asym) error costs. Columns 2–4 are for models that only include treatments as independent variables. Columns 5–7 add the sequence-level control.

*(0.05); **(0.01); ***(0.001).

Adding asymmetric error costs, however, neutralised the tendency for explicit incentives to favour blue choices. Although the blue state was twice as likely as red ex ante, choosing blue in the red state cost twice as much as choosing red in the blue state. Neutralising incentives in this way allowed natural variation in private signals to create considerable variation in choices within asocial sequences (Figure 5b). The result was a highly significant increase in the variance relative to asocial treatments with symmetric error costs (Table 3, NoAgencyAsym and AgencyAsym). Moreover, by pushing choices away from blue, asymmetric error costs increased the scope for social information to homogenise choices within sequences via path-dependent dynamics. This reduction in variance within sequences is exactly what happened (Figures 5d vs. 5b), and interacting the availability of social information with asymmetric costs (Table 3, Social × NoAgencyAsym and Social × AgencyAsym) produced highly significant negative interactions. In sum, asymmetric costs decreased homogeneity within groups, and social information, given asymmetric costs, increased homogeneity within groups.

Discussion

Both our model and experiment indicate that asymmetries in the explicit incentive structure can exert a powerful effect on choices. Indeed, even without a cognitive bias, rational optimisers with Bayesian beliefs can choose in a way that is optimal under risk, and yet the resulting choices are likely to be costly errors given the state that actually obtains. In effect, smart bets do not always produce good outcomes. More interestingly, explicit contemporary cost asymmetries readily yield scenarios in which smart bets are extremely likely to lead to bad outcomes. We can hijack a staple example from the error management literature to illustrate the point (Haselton & Buss, Reference Haselton and Buss2000; Haselton, Reference Haselton2003; Haselton & Nettle, Reference Haselton and Nettle2006; Johnson et al., Reference Johnson, Blumstein, Fowler and Haselton2013; Haselton et al., Reference Haselton, Nettle and Murray2015; Perilloux & Kurzban, Reference Perilloux and Kurzban2015; Murray et al., Reference Murray, Murphy, von Hippel, Trivers and Haselton2017). Imagine a man in a bar approaching every woman there, only to be repeatedly rejected. His persistence could follow from the fact that he is overestimating his chances, but this is not necessary. Perhaps he simply views a missed opportunity as far more costly than a rejection. Such an asymmetry could easily generate the behaviour in question even if the man has an accurate and extremely precise understanding of just how bad his objective chances are (McKay & Efferson, Reference McKay and Efferson2010). He persists as a general-purpose optimiser, and he usually goes home alone.

Importantly, cost asymmetries do not need to be especially large, just something other than trivial. Consistent with this idea, we used a relatively small cost asymmetry in our experiment, and it produced a large behavioural effect. The cost of choosing blue when the state was red was only twice the cost of choosing red when blue. Introducing this moderate asymmetry, however, increased the rate of red choices by a factor of nearly 3.5, from 16.9% to 58.8% (Figure 4).

Frequency-dependent social learning can also have dramatic effects via path-dependent dynamics, but the details are decisive. Path dependence implies at least two dynamically stable equilibria. One of the equilibria has most people choosing correctly given the state, while another equilibrium has most people choosing in error given the state. Because widespread errors can be a stable equilibrium, social learning provides a cogent hypothesis about the origin and persistence of seemingly harmful traditions (Mackie, Reference Mackie1996). The existence of path dependence owing to social learning, however, can be hypersensitive to the effects of other mechanisms. Ordinary individual heterogeneity (Granovetter, Reference Granovetter1978; Young, Reference Young2009; Muthukrishna et al., Reference Muthukrishna, Morgan and Henrich2016; Efferson et al., Reference Efferson, Vogt and Fehr2020), an evolved combination of individual and social learning (Perreault et al., Reference Perreault, Moya and Boyd2012; Efferson et al., Reference Efferson, Lalive, Cacault and Kistler2016), systematic errors (Goeree et al., Reference Goeree, Palfrey, Rogers and McKelvey2007) and identity concerns (Efferson et al., Reference Efferson, Vogt and Fehr2020) can all destabilise or eliminate equilibria in a system that would otherwise exhibit multiple equilibria, path-dependent dynamics and the potential for stable harmful traditions.

Our results support this overall picture. We found that frequency-dependent social learning strongly influenced cultural evolutionary dynamics, but the details were critical. When error costs were symmetric, the prior distribution over environmental states favoured blue 2/3 to 1/3. Most participants simply chose blue, which left little scope for social learning to amplify random variation in choices early in sequences. In these cases, path dependence played little or no role. Asocial and social learning were similar (Table 3), with perhaps only a small and uncertain tendency for social learning to increase red choices (Table 1 and Figures 5a and c).

When error costs were asymmetric, in contrast, the cost asymmetry favoured red, while the prior distribution over states favoured blue. These two countervailing forces neutralised each other, allowing the aggregate-level effects of social learning to appear. We found that social learning had little effect on average behaviour (Table 1), but it had a large effect on how choices were distributed within vs. between groups. Social learning and associated path dependence shifted much of the behavioural variation from within groups to between groups (Table 3 and Figures 5b, d). Social learning thus transformed errors from a question about individual decision making into a question about the cultural evolution of group traditions. Importantly, we have focussed on frequency-dependent social learning and associated cultural evolutionary dynamics, but this is only one of many possible cultural evolutionary processes (Cavalli-Sforza & Feldman, Reference Cavalli-Sforza and Feldman1981; Sperber, Reference Sperber1996; Richerson & Boyd, Reference Richerson and Boyd2005; Mesoudi, Reference Mesoudi2011; Henrich, Reference Henrich2015; Morin, Reference Morin2016; Heyes, Reference Heyes2018). Our study does not allow conclusions about the effects of social learning more broadly.

Finally, we found no evidence that our linguistic prime activated a cognitive bias. Error management theory (Haselton & Nettle, Reference Haselton and Nettle2006; Johnson et al., Reference Johnson, Blumstein, Fowler and Haselton2013; Haselton et al., Reference Haselton, Nettle and Murray2015) has provided a rich framework for generating hypotheses and empirical studies about evolved cognitive biases. As discussed above, the key idea is that persistent cost asymmetries in the past shaped the evolution of human cognition to avoid errors that would have been especially costly in ancestral environments. If stimuli activate this ancestral psychology in a contemporary setting, the person in question will choose in a way that avoids the ancestral error. Moreover, she will do so in a way that is somehow distinct from the effects of explicit contemporary information and explicit contemporary incentives.

We have written before about the challenges associated with defining and identifying cognitive biases of this sort (McKay & Efferson, Reference McKay and Efferson2010; Haselton et al., Reference Haselton, Nettle and Murray2015; McKay et al., Reference McKay, Ross, O'Lone and Efferson2018). Notwithstanding the difficulties, at least two empirical strategies exist. First, the researcher can compare two situations that are equivalent in terms of the value of the information available to subjects. One situation is consistent with the hypothesised error management bias, while the other is not. If decision making varies between the two situations in the predicted direction, the result provides support for the cognitive bias. Predicting outcomes in negatively vs. positively autocorrelated sequences (Scheibehenne et al., Reference Scheibehenne, Wilke and Todd2011) provides an example of this approach. Second, the researcher can manipulate payoff-irrelevant stimuli in a way that runs orthogonal to explicit incentives. This is what we have done here. Economic games with eye spots (Haley & Fessler, Reference Haley and Fessler2005; Nettle et al., Reference Nettle, Harper, Kidson, Stone, Penton-Voak and Bateson2013; Sparks & Barclay, Reference Sparks and Barclay2013; Vogt et al., Reference Vogt, Efferson, Berger and Fehr2015; Northover et al., Reference Northover, Pedersen, Cohen and Andrews2017) and alternative framings of the Wason task (Cosmides & Tooby, Reference Cosmides and Tooby1992) provide other examples.

In choosing this latter strategy, we varied the frame used to describe outcomes by switching between a neutral frame and a frame that associated unseen agents with a specific state. By varying the frame independently of explicit incentives, our design effectively decomposed asymmetric error costs into a component related to the explicit incentives of the contemporary decision making task and a component related to how cognition evolved owing to ancestral incentive structures. With respect to contemporary incentives, we implemented incentive structures that either did or did not involve explicit error cost asymmetries. With respect to activating biased cognition, our agency prime posited a state (blue) in which costs and benefits simply occur and a state (red) in which unseen agents distribute these costs and benefits as punishments and rewards. We compared this with a control in which costs and benefits for both states were framed in neutral terms. Interestingly, this decomposition mirrors the two dimensions at work when we ask, ‘Does God exist?’ First, the question involves incentives (Pascal, Reference Pascal and Levi1995, originally published in 1670). Going to hell instead of heaven is quite costly, while going to church when you could stay home is less so. Second, the question involves beliefs about the origin of outcomes (Barrett, Reference Barrett2012). Under one state, outcomes originate from the dispassionate workings of nature. Under the other state, God is responsible.

Although our prime did not affect choices, this does not mean that people do not have a cognitive bias that overinfers the existence of unseen agents. People may have such a bias, but our agency frame failed to activate it. We cannot exclude this interpretation, and this represents a key limitation of our study. Another recent study, however, also failed to find positive evidence for hyperactive agency detection (Maij et al., Reference Maij, van Schie and van Elk2019).

Regardless, our model suggests that a cognitive bias, even if present and active, may have limited effects in a setting with individual and social learning. To consistently distort the evolution of beliefs away from the truth, the bias must be relatively strong. Otherwise, even if cognition is biased, beliefs can still evolve towards the truth under a wide range of conditions. This implies yet another possibility. Namely, our agency prime may have activated a cognitive bias, but only weakly so, in which case the behavioural effects were too small to detect. Consequently, future research could profit from a more forceful approach to attaching unseen agents to a particular environmental state. For example, some studies have used a paradigm in which participants play a game against partners who are either real people or computers programmed to play like real people (e.g. Falk et al., Reference Falk, Fehr and Fischbacher2008). This isolates the effects of playing against an active decision maker vs. playing against nature when nature's choices mimic active decision makers. To modify this approach for present purposes, one would require instead that active decision makers, who fill the role of unseen agents, mimic nature.

Finally, even if a cognitive bias does not always lead beliefs to evolve in the wrong direction when the bias conflicts with reality, our model supports the general claim that bias always prevents beliefs from getting arbitrarily close to the truth. This suggests that empiricists can maximise the chances of detecting biases if they work in settings where prior beliefs are close to accurate. We hope this unexpected theoretical finding highlights the value of modelling cognitive biases relative to an explicitly unbiased benchmark (McKay & Efferson, Reference McKay and Efferson2010; Zimmermann & Efferson, Reference Zimmermann and Efferson2017).

Supplementary material

The supplementary material for this article can be found at https://doi.org/10.1017/ehs.2020.25.

Acknowledgements

We would like to thank Holger Herz, two anonymous referees, and seminar participants at Queen Mary, the University of St Andrews, University College London, the Santa Fe Institute, ETH Zurich, the University of Geneva, the University of Zurich, and the University of Lausanne.

Author contributions

CE, RM, and EF designed the experiment. CE and RM implemented the experiment. CE developed and analysed the model. CE analysed the experimental data. CE wrote the paper with input from RM and EF.

Financial support

For supporting this research, CE would like to thank the Swiss National Science Foundation (grant no 100018_185417/1), and CE and RM would like to thank the Cogito Foundation (grant no. R10917).

Conflict of interest

CE is a member of the editorial board at Evolutionary Human Sciences.

Research transparency and reproducibility

Code for simulations, raw data, and code for data analysis are available at https://github.com/cmefferson/evoHumSciEffersonetal2020_codeAndData.

References

Abbey, A. (1982). Sex differences in attributions for friendly behavior: Do males misperceive females’ friendliness? Journal of Personality and Social Psychology, 42(5), 830.CrossRefGoogle Scholar
Anderson, L. R., & Holt, C. A. (1997). Information cascades in the laboratory. The American Economic Review, 87(5), 847862.Google Scholar
Angrist, J. D. (2014). The perils of peer effects. Labour Economics, 30, 98108.CrossRefGoogle Scholar
Angrist, J. D., & Pischke, J. S. (2009). Mostly harmless econometrics: An empiricist's companion. Princeton University Press.CrossRefGoogle Scholar
Arai, M. (2011). Cluster-robust standard errors using R. http://people.su.se/ma/clustering.pdfGoogle Scholar
Asch, S. E. (1955). Opinions and social pressure. Scientific American, 193(5), 17.CrossRefGoogle Scholar
Barrett, H. C. (2015). The shape of thought: How mental adaptations evolve. Oxford University Press.CrossRefGoogle Scholar
Barrett, J. L. (2000). Exploring the natural foundations of religion. Trends in Cognitive Sciences, 4(1), 2934.CrossRefGoogle Scholar
Barrett, J. L. (2004). Why would anyone believe in God? AltaMira Press.Google Scholar
Barrett, J. L. (2012). Born believers: The science of children's religious belief. Simon and Schuster.Google Scholar
Berns, G. S., Chappelow, J., Zink, C. F., Pagnoni, G., Martin-Shurski, M. E., & Richards, J. (2005). Neurobiological correlates of social conformity and independence during mental rotation. Biological Psychiatry, 58(3), 245253.CrossRefGoogle ScholarPubMed
Bikhchandani, S., Hirshleifer, D., & Welch, I. (1992). A theory of fads, fashion, custom, and cultural change as information cascades. The Journal of Political Economy, 100(5), 9921026.CrossRefGoogle Scholar
Bikhchandani, S., Hirshleifer, D., & Welch, I. (1998). Learning from the behavior of others: Conformity, fads, and informational cascades. The Journal of Economic Perspectives, 12(3), 151170.CrossRefGoogle Scholar
Bowles, S. (2004). Microeconomics: Behavior, institutions, and evolution. Russell Sage.Google Scholar
Boyd, R., & Richerson, P. J. (1985). Culture and the evolutionary process. University of Chicago Press.Google Scholar
Boyer, P. (2001). Religion explained: The evolutionary origins of religious thought. Basic Books.Google Scholar
Buckle, C., Chuah, Y. M. L., Fones, C. S. L., & Wong, A. H. C. (2007). A conceptual history of Koro. Transcultural Psychiatry, 44(1), 2743.CrossRefGoogle ScholarPubMed
Bures, F. (2008). A mind dismembered: In search of magical penis thieves. Harper's Magazine, June, 60–65.Google Scholar
Camerer, C. F. (2003). Behavioral game theory: Experiments in strategic interaction. Princeton University Press.Google Scholar
Camerer, C. F., Dreber, A., Forsell, E., Ho, T.-H., Huber, J., Johannesson, M., … Wu, H. (2016). Evaluating replicability of laboratory experiments in economics. Science.CrossRefGoogle ScholarPubMed
Cameron, A. C., & Trivedi, P. K. (2005). Microeconometrics: Methods and applications. Cambridge University Press.CrossRefGoogle Scholar
Cavalli-Sforza, L. L., & Feldman, M. W. (1981). Cultural transmission and evolution: A quantitative approach. Princeton University Press.Google ScholarPubMed
Chamley, C. P. (2004). Rational herds: Economic models of social learning. Cambridge University Press.Google Scholar
Cheng, S.-T. (1996). A critical review of Chinese Koro. Culture, Medicine, and Psychiatry, 20, 6782.CrossRefGoogle ScholarPubMed
Cloward, K. (2016). When norms collide: Local responses to activism against female genital mutilation and early marriage. Oxford University Press.CrossRefGoogle Scholar
Cosmides, L., & Tooby, J. (1992). Cognitive adaptations for social exchange. The Adapted Mind: Evolutionary Psychology and the Generation of Culture, 163, 163228.Google Scholar
Cosmides, L., & Tooby, J. (2013). Evolutionary psychology: New perspectives on cognition and motivation. Annual Review of Psychology, 64, 201229.CrossRefGoogle ScholarPubMed
Cover, T. M., & Thomas, J. A. (2006). Elements of Information Theory, 2nd ed.John Wiley & Sons.Google Scholar
Delton, A. W., Krasnow, M. M., Cosmides, L., & Tooby, J. (2011). Evolution of direct reciprocity under uncertainty can explain human generosity in one-shot encounters. Proceedings of the National Academy of Sciences, 108(32), 1333513340.CrossRefGoogle ScholarPubMed
Efferson, C., Lalive, R., Cacault, M. P., & Kistler, D. (2016). The evolution of facultative conformity based on similarity. PLoS One, 11(12), e0168551.CrossRefGoogle ScholarPubMed
Efferson, C., & Vogt, S. (2018). Behavioural homogenization with spillovers in a normative domain. Proceedings of the Royal Society B, 285(1879), 20180492.CrossRefGoogle Scholar
Efferson, C., Vogt, S., Elhadi, A., Ahmed, H. E. F., & Fehr, E. (2015). Female genital cutting is not a social coordination norm. Science, 349(6255), 14461447.CrossRefGoogle Scholar
Efferson, C., Vogt, S., & Fehr, E. (2020). The promise and the peril of using social influence to reverse harmful traditions. Nature Human Behaviour, 4, 5568.CrossRefGoogle ScholarPubMed
Falk, A., Fehr, E., & Fischbacher, U. (2008). Testing theories of fairness – Intentions matter. Games and Economic Behavior, 62(1), 287303.CrossRefGoogle Scholar
Fischbacher, U. (2007). z-Tree: Zurich toolbox for ready-made economic experiments. Experimental Economics, 10(2), 171178.CrossRefGoogle Scholar
Foster, K. R., & Kokko, H. (2009). The evolution of superstitious and superstition-like behaviour. Proceedings of the Royal Society of London B: Biological Sciences, 276(1654), 3137.CrossRefGoogle ScholarPubMed
Frankenhuis, W. E., Panchanathan, K., & Clark Barrett, H. (2013). Bridging developmental systems theory and evolutionary psychology using dynamic optimization. Developmental Science, 16(4), 584598.CrossRefGoogle ScholarPubMed
Gaines, B. J., Kuklinski, J. H., Quirk, P. J., Peyton, B., & Verkuilen, J. (2007). Same facts, different interpretations: Partisan motivation and opinion on Iraq. The Journal of Politics, 69(04), 957974.CrossRefGoogle Scholar
Gervais, W. M., & Norenzayan, A. (2012). Like a camera in the sky? Thinking about God increases public self-awareness and socially desirable responding. Journal of Experimental Social Psychology, 48(1), 298302.CrossRefGoogle Scholar
Goeree, J. K., Palfrey, T. R., Rogers, B. W., & McKelvey, R. D. (2007). Self-correcting information cascades. Review of Economic Studies, 74, 733762.CrossRefGoogle Scholar
Goeree, J. K., & Yariv, L. (2015). Conformity in the lab. Journal of the Economic Science Association, 1(1), 1528.CrossRefGoogle Scholar
Granovetter, M. (1978). Threshold models of collective behavior. American Journal of Sociology, 83(6), 14201443.CrossRefGoogle Scholar
Guthrie, S. (1993). Faces in the clouds: A new theory of religion. Oxford University Press.Google Scholar
Guthrie, S., Agassi, J., Andriolo, K. R., Buchdahl, D., Earhart, H. B., Greenberg, M., … Tissot, G. (1980). A cognitive theory of religion [and comments and reply]. Current Anthropology, 21(2), 181203.CrossRefGoogle Scholar
Haley, K. J., & Fessler, D. M. T. (2005). Nobody's watching?: Subtle cues affect generosity in an anonymous economic game. Evolution and Human Behavior, 26(3), 245256.CrossRefGoogle Scholar
Harsanyi, J. C., & Selten, R. (1988). A general theory of equilibrium selection in games. The MIT Press.Google Scholar
Haselton, M. G. (2003). The sexual overperception bias: evidence of a systematic bias in men from a survey of naturally occurring events. Journal of Research in Personality, 37(1), 3447.CrossRefGoogle Scholar
Haselton, M. G., & Buss, D. M. (2000). Error management theory: A new perspective on biases in cross-sex mind reading. Journal of Personality and Social Psychology, 78(1), 8191.CrossRefGoogle ScholarPubMed
Haselton, M., & Galperin, A. (2012). Error management and the evolution of cognitive bias. Social Thinking and Interpersonal Behaviour, 45, 63.Google Scholar
Haselton, M. G., & Nettle, D. (2006). The paranoid optimist: An integrative evolutionary model of cognitive biases. Personality and Social Psychology Review, 10(1), 4766.CrossRefGoogle ScholarPubMed
Haselton, M. G., Nettle, D., & Murray, D. R. (2015). The evolution of cognitive bias. In The Handbook of Evolutionary Psychology (pp. 120). https://doi.org/10.1002/9781119125563.evpsych241Google Scholar
Henrich, J. (2015). The secret of our success: How culture is driving human evolution, domesticating our species, and making us smarter. Princeton University Press.CrossRefGoogle Scholar
Heyes, C. (2018). Cognitive gadgets: The cultural evolution of thinking. Harvard University Press.CrossRefGoogle Scholar
Ilechukwu, S. T. (1992). Magical penis loss in Nigeria: Report of a recent epidemic of a Koro-like syndrome. Transcultural Psychiatric Research Review, 29(2), 91108.CrossRefGoogle Scholar
Johnson, D. D., Blumstein, D. T., Fowler, J. H., & Haselton, M. G. (2013). The evolution of error: Error management, cognitive constraints, and adaptive decision-making biases. Trends in Ecology & Evolution, 28(8), 474481.CrossRefGoogle ScholarPubMed
Kahneman, D., & Tversky, A. (1979). Prospect theory: An analysis of decision under risk. Econometrica, 47(2), 263292.CrossRefGoogle Scholar
Mackie, G. (1996). Ending footbinding and infibulation: A convention account. American Sociological Review, 61, 9991017.CrossRefGoogle Scholar
Maij, D. L., van Schie, H. T., & van Elk, M. (2019). The boundary conditions of the hypersensitive agency detection device: An empirical investigation of agency detection in threatening situations. Religion, Brain & Behavior, 9(1), 2351.CrossRefGoogle Scholar
Manski, C. F. (2000). Economic analysis of social interactions. The Journal of Economic Perspectives, 14(3), 115136.CrossRefGoogle Scholar
Marshall, J. A., Trimmer, P. C., Houston, A. I., & McNamara, J. M. (2013). On evolutionary explanations of cognitive biases. Trends in Ecology & Evolution, 28(8), 469473.CrossRefGoogle ScholarPubMed
McKay, R., & Efferson, C. (2010). The subtleties of error management. Evolution and Human Behavior, 31(5), 309319.CrossRefGoogle Scholar
McKay, R., Ross, R. M., O'Lone, K., & Efferson, C. (2018). The error of God, revisited. Religion, Brain & Behavior, 8(3), 306310.CrossRefGoogle Scholar
Mesoudi, A. (2011). Cultural evolution: How Darwinian theory can explain human culture & synthesize the social sciences. University of Chicago Press.CrossRefGoogle Scholar
Morgan, T. J. H., Rendell, L. E., Ehn, M., Hoppitt, W., & Laland, K. N. (2012). The evolutionary basis of human social learning. Proceedings of the Royal Society B, 279(1729), 653662.CrossRefGoogle ScholarPubMed
Morin, O. (2016). How traditions live and die. Oxford University Press.Google Scholar
Murray, D. R., Murphy, S. C., von Hippel, W., Trivers, R., & Haselton, M. G. (2017). A preregistered study of competing predictions suggests that men do overestimate women's sexual intent. Psychological Science, 28(2), 253255.CrossRefGoogle ScholarPubMed
Muthukrishna, M., & Henrich, J. (2019). A problem in theory. Nature Human Behaviour, 3(3), 221229.CrossRefGoogle ScholarPubMed
Muthukrishna, M., Morgan, T. J., & Henrich, J. (2016). The when and who of social learning and conformist transmission. Evolution and Human Behavior, 37(1), 1020.CrossRefGoogle Scholar
Nettle, D., Harper, Z., Kidson, A., Stone, R., Penton-Voak, I. S., & Bateson, M. (2013). The watching eyes effect in the Dictator Game: It's not how much you give, it's being seen to give something. Evolution and Human Behavior, 34(1), 3540.CrossRefGoogle Scholar
Northover, S. B., Pedersen, W. C., Cohen, A. B., & Andrews, P. W. (2017). Artificial surveillance cues do not increase generosity: Two meta-analyses. Evolution and Human Behavior, 38(1), 144153.CrossRefGoogle Scholar
Open-Science-Collaboration (2015). Estimating the reproducibility of psychological science. Science, 349(6251).Google Scholar
Pascal, B. (1995). Pensées and other writings (English translation by Levi, H.). Oxford University Press. (Original work published in 1670.)Google Scholar
Perilloux, C., & Kurzban, R. (2015). Do men overperceive women's sexual interest? Psychological Science, 26(1), 7077.CrossRefGoogle ScholarPubMed
Perreault, C., Moya, C., & Boyd, R. (2012). A Bayesian approach to the evolution of social learning. Evolution and Human Behavior, 33(5), 449459.CrossRefGoogle Scholar
Platteau, J.-P., Camilotti, G., & Auriol, E. (2018). Eradicating women-hurting customs. In Anderson, S., Beaman, L. & Platteau, J. (Eds.), Towards gender equity in development (pp. 319356). Oxford University Press.Google Scholar
Richerson, P. J., & Boyd, R. (2005). Not by genes alone: How culture transformed the evolutionary process. University of Chicago Press.Google Scholar
Sachdev, P. (1985). Koro epidemic in north-east India. Australian and New Zealand Journal of Psychiatry, 19(4), 433438.CrossRefGoogle ScholarPubMed
Scheibehenne, B., Wilke, A., & Todd, P. M. (2011). Expectations of clumpy resources influence predictions of sequential events. Evolution and Human Behavior, 32(5), 326333.CrossRefGoogle Scholar
Shariff, A. F., & Norenzayan, A. (2007). God is watching you: Priming god concepts increases prosocial behavior in an anonymous economic game. Psychological Science, 18(9), 803809.CrossRefGoogle Scholar
Sherif, M. (1936). The Psychology of Social Norms. Harper and Brothers.Google Scholar
Sparks, A., & Barclay, P. (2013). Eye images increase generosity, but not for long: The limited effect of a false cue. Evolution and Human Behavior, 34(5), 317322.CrossRefGoogle Scholar
Sperber, D. (1996). Explaining culture: A naturalistic approach. Cambridge.Google Scholar
Toyokawa, W., Whalen, A., & Laland, K. N. (2019). Social learning strategies regulate the wisdom and madness of interactive crowds. Nature Human Behaviour, 3(2), 183193.CrossRefGoogle ScholarPubMed
Tversky, A., & Kahneman, D. (1981). The framing of decisions and the psychology of choice. Science, 211(4481), 453458.CrossRefGoogle ScholarPubMed
Tversky, A., & Kahneman, D. (2000). Choices, values, and frames. Cambridge University Press.Google Scholar
Vogt, S., Efferson, C., Berger, J., & Fehr, E. (2015). Eye spots do not increase altruism in children. Evolution and Human Behavior, 36(3), 224231.CrossRefGoogle Scholar
von Neumann, J., & Morgenstern, O. (1944). Theory of games and economic behavior. Princeton University Press, Princeton.Google Scholar
Wooldridge, J. M. (2002). Econometric analysis of cross section and panel data. The MIT Press.Google Scholar
Yap, P. M. (1965). Koro – A culture-bound depersonalization syndrome. British Journal of Psychiatry, 111, 4350.CrossRefGoogle Scholar
Young, H. P. (1996). The economics of convention. The Journal of Economic Perspectives, 10(2), 105122.CrossRefGoogle Scholar
Young, H. P. (2009). Innovation diffusion in heterogeneous populations: Contagion, social influence, and social learning. American Economic Review, 99(5), 18991924.CrossRefGoogle Scholar
Young, H. P. (2015). The evolution of social norms. Annual Review of Economics, 7(1), 359387.CrossRefGoogle Scholar
Zimmermann, J., & Efferson, C. (2017). One-shot reciprocity under error management is unbiased and fragile. Evolution and Human Behavior, 38(1), 3947.CrossRefGoogle Scholar
Figure 0

Figure 1. The evolution of beliefs and choices when error costs are symmetric (u11 = u00 = 1 and u01 = u10 = 0). For each panel, we simulated 100 independent sequences with 1/3 as the ex ante probability of state 1. Each panel shows results for the specific sequences in which the actual state is 0. Over these sequences, the graph shows the distribution of prior beliefs that the state is 1 (open circles) for every tenth decision maker in the sequence. It also shows the associated distribution over the cumulative proportions, by sequence, of decision makers incorrectly guessing state 1 (closed circles). Distributions are represented as bubble plots. (a) Cognition is biased (α = 0.5) and decision making relatively noisy (λ = 10). (b) Cognition is biased (α = 0.5) and decision making relatively systematic (λ = 100). (c) Cognition is unbiased (α = 0) and decision making relatively noisy (λ = 10). (d) Cognition is unbiased (α = 0) and decision making relatively systematic (λ = 100). See the main text for a detailed description of how to read the graphs and a summary of key results.

Figure 1

Figure 2. The evolution of beliefs and choices when error costs involve a relatively weak asymmetry (u11 = 1, u00 = 0.75, u01 = 0, and u10 = 0.25). For each panel, we simulated 100 independent sequences with 1/3 as the ex ante probability of state 1. Each panel shows results for the specific sequences in which the actual state is 0. Over these sequences, the graph shows the distribution of prior beliefs that the state is 1 (open circles) for every tenth decision maker in the sequence. It also shows the associated distribution over the cumulative proportions, by sequence, of decision makers incorrectly guessing state 1 (closed circles). Distributions are represented as bubble plots. (a) Cognition is biased (α = 0.5) and decision making relatively noisy (λ = 10). (b) Cognition is biased (α = 0.5) and decision making relatively systematic (λ = 100). (c) Cognition is unbiased (α = 0) and decision making relatively noisy (λ = 10). (d) Cognition is unbiased (α = 0) and decision making relatively systematic (λ = 100). See the main text for a detailed description of how to read the graphs and a summary of key results.

Figure 2

Figure 3. The evolution of beliefs and choices when error costs involve a relatively strong asymmetry (u11 = 1, u00 = 0.6, u01 = 0, and u10 = 0.4). For each panel, we simulated 100 independent sequences with 1/3 as the ex ante probability of state 1. Each panel shows results for the specific sequences in which the actual state is 0. Over these sequences, the graph shows the distribution of prior beliefs that the state is 1 (open circles) for every tenth decision maker in the sequence. It also shows the associated distribution over the cumulative proportions, by sequence, of decision makers incorrectly guessing state 1 (closed circles). Distributions are represented as bubble plots. (a) Cognition is biased (α = 0.5) and decision making relatively noisy (λ = 10). (b) Cognition is biased (α = 0.5) and decision making relatively systematic (λ = 100). (c) Cognition is unbiased (α = 0) and decision making relatively noisy (λ = 10). (d) Cognition is unbiased (α = 0) and decision making relatively systematic (λ = 100). See the main text for a detailed description of how to read the graphs and a summary of key results.

Figure 3

Figure 4. Red choices by treatment and realised state. The proportion of red choices for (a) asocial treatments and (b) for social treatments. The colour of the bars signifies the realised state. Consequently, the red bars represent correct choices in the relatively rare case of a red state, while the blue bars represent errors in the relatively common case of a blue state. These results show that explicit cost asymmetries had an overwhelmingly dominant effect on average choices (see Table 1).

Figure 4

Table 1. Red choices in all treatments. Linear probability models with red choices as the response variable and robust clustered standard errors calculated by clustering on session. In addition, the table shows 95% and 99% confidence intervals calculated with a non-parametric bootstrap clustered at the session level. Independent variables include a dummy for the realised environment for the sequence (Env red), the realised private signal (Signal red), the subject's gender, order in the sequence and treatment dummies (Asym, Agency prime, Social). Columns 2–4 are for models that only include treatments as independent variables. Columns 5–7 add controls.

Figure 5

Table 2. Red choices in social treatments with frequency-dependent social information. Linear probability models with red choices as the response variable and robust clustered standard errors calculated by clustering on session. In addition, the table shows 95% and 99% confidence intervals calculated with a non-parametric bootstrap clustered at the session level. Independent variables include a dummy for the realised environment for the sequence (Env red), the realised private signal (Signal red), the subject's gender, order in the sequence, the centred cumulative proportion choosing red through the previous period (Lagged social info), and relevant treatment dummies (Asym, Agency prime).

Figure 6

Figure 5. Choice dynamics for all treatments. Let cn = 0 denote a blue choice in sequence position n and cn = 1 a red choice. Given position t, the graphs show the cumulative proportion choosing red by sequence, ${\sum _{n = 1}^{t}} c_n / t$, as a function of sequence position, t, for all sequences in the experiment. The colour of the line shows the realised state for the sequence in question. Solid lines are for sequences in no agency prime treatments, while dashed lines are for sequences in agency prime treatments. Panels (a) and (b) show asocial treatments, while (c) and (d) show social treatments. Panels (a) and (c) show treatments with symmetric error costs, while (b) and (d) show asymmetric error costs. Social learning led to path-dependent dynamics and an associated increase in homogeneity within sequences when error costs were asymmetric.

Figure 7

Table 3. Ordinary least squares models with the variance in choices by sequence as the response variable and robust clustered standard errors calculated by clustering on session. The table also shows 95% and 99% confidence intervals calculated with a non-parametric bootstrap clustered at the session level. Independent variables include a dummy for the realised environment by sequence (Env red), a dummy for treatments that allowed social learning (Social), and dummies for the four treatment combinations involving the agency prime and the explicit payoff structure. We used combined dummies for these four combinations in order to avoid three-way interactions. The dummies are defined according to the presence (Agency) or absence (NoAgency) of the agency prime and either symmetric (Sym) or asymmetric (Asym) error costs. Columns 2–4 are for models that only include treatments as independent variables. Columns 5–7 add the sequence-level control.

Supplementary material: PDF

Efferson et al. supplementary material

Efferson et al. supplementary material

Download Efferson et al. supplementary material(PDF)
PDF 299.8 KB