In retrospect, it was a seminal moment in the psychology world: Stanley Milgram surprised even himself when he discovered just how obedient people can be. It was 1963 and Milgram was looking to answer questions about the cruel and immoral side of human nature: Could Nazi Germany repeat itself? Are people capable of following orders that profoundly conflict with their moral principles? Could normal, law-abiding citizens of New Haven, Connecticut be pushed to do gravely immoral acts? Milgram required his subjects to apply electric shocks to a second subject (a confederate) each time a wrong answer was given during a memory test. The result: 67 percent of them were willing to apply a level of shock sufficient to silence a screaming confederate.
Following on the heels of World War II, Milgram’s experiments, which highlighted just how antisocial people can be, were particularly germane; however, issues relating to morality, prosociality, honesty, trust, and cooperation, have been, and will continue to be, relevant throughout humanity’s tenure. This is largely because the issues central to being prosocial revolve around our capacity for successful social living. Prosocial actions enable an individual to put the needs of another before their own and act in ways that increase the well-being of the other, typically at the expense of oneself (Batson & Powell, Reference Batson, Powell, Millon and Lerner2003; Eisenberg, Reference Eisenberg2014; Penner et al., Reference Penner, Dovidio, Piliavin and Schroeder2005). From mundane acts, such as giving your taxi to a stranger because she is in a hurry, to more consequential behaviors, such as heroically donating organs to those in need (Marsh et al., Reference Marsh, Stoycos, Brethel-Haurwitz, Robinson, VanMeter and Cardinale2014), our prosocial calculus enables society to function in a relatively seamless manner.
Successful social living, however, is only possible when there is a quorum of individuals behaving in altruistic, cooperative, and trustworthy ways. This notion was not readily apparent in social psychology’s early research tradition, which, in its nascent years, primarily focused on documenting antisocial behavior. There was good reason for this focus. As social psychology was finding its footing as a field, the world had just been rocked by two world wars, in which more than 100 million people died. The realization that humanity’s darker side can be so easily uncovered was fresh, and researchers leveraged this insight to explore the depth of this antisocial capacity. Milgram’s obedience study (Milgram, Reference Milgram1963) and Zimbardo’s Stanford prison study (Haney et al., Reference Haney, Banks and Zimbardo1973) are classically evoked examples, and Sherif’s summer camp study in the late 1950s illustrates another storied case for how competition and group membership rapidly unfold in the wild (Sherif et al., Reference Sherif, Harvey, White, Hood and Sherif1961). In the last two decades, however, there has been a seismic shift toward understanding the prosocial mind. If we can take Google’s software that counts frequency of words in written text as a barometer of topic interest, we would see that the word “prosocial” has exponentially grown since the beginning of the 2000s. Prosociality is trending.
In this chapter, we leverage this newfound interest to provide a brief account of the historical relevance of prosocial behavior and the current research objectives within the fields of social psychology and economics. In the last few decades, the topic of prosociality has been covered by a number of different perspectives and research fields. Even within the domain of psychology, the approach has differed substantially depending on tradition, from investigating which personality traits correlate with prosociality (Graziano & Eisenberg, Reference Graziano, Eisenberg, Hogan, Johnson and Briggs1997) to describing the external environmental factors that influence prosocial actions (Latané & Darley, Reference Latané and Darley1970). In the past, major reviews have focused on the many levels of analysis one could take, from micro to macro, which has helped guide at what level of abstraction a research question is pitched (Penner et al., Reference Penner, Dovidio, Piliavin and Schroeder2005). Here we evoke an experimentalist framework, covering ground from a number of different subfields. We begin by discussing how early paradigms originating from behavioral economics were co-opted by social psychologists to investigate prosocial behavior in the laboratory. We then explain how more recent efforts have been leveraged to create paradigms that more adequately capture the tensions experienced in the real world. We then address the potential origins of human prosociality by briefly reviewing both primate and developmental work. We also highlight the data to date that have tried to take a more mechanistic approach to the question of what motivates people to be prosocial (e.g., empathy, risk attitudes, temporary affective states). No chapter on prosociality would be complete without a brief nod to the neuroscientific advances used to map the prosocial brain. Finally, we conclude with often overlooked topics and offer a few suggestions for future research avenues.
12.1 Studying Prosocial Behavior in the Lab
Although entirely unintended, economists have a historical lead on other fields when it comes to examining prosocial behaviors in the laboratory. Economic games – which were largely created to explore negotiation tactics, the boundary conditions of “rationality,” and the strategies that destabilize the economic system – swiftly revealed that humans have a deeply ingrained prosocial tendency. Economic games typically involve the division of some good (in most cases, money) between different parties. Perhaps the most famous paradigm is the prisoner’s dilemma, created by Dresher and Flood in the late 1940s. The prisoner’s dilemma captures the nature of cooperation: A relatively good outcome is accomplished only by virtue of two parties deciding to collaborate, even though each individual will be better off if they defect. Since its creation, the prisoner’s dilemma has been used in hundreds of experiments, leading to critical insights about the conditions that scaffold prosocial tendencies. For example, communication between parties facilitates cooperation (Sally, Reference Sally1995; Wichman, Reference Wichman1970), cooperation is harder to maintain as the termination horizon gets closer (e.g., coalition governments in an election year; Embrey et al., Reference Embrey, Fréchette and Yuksel2018); and people routinely exhibit tit-for-tat strategies when cooperating (Axelrod & Hamilton, Reference Axelrod and Hamilton1981).
An early game-theoretic analysis of human interaction argued that being presented with the possibility of an economic gain linked to an antisocial option would likely lead to selfish behavior. However, one behavioral paradigm – called the ultimatum game (UG), originally designed by Güth et al. (Reference Güth, Schmittberger and Schwarze1982) to test how people negotiate – revealed the exact opposite. In the UG, there are two players, a proposer and a responder. The proposer is endowed with an amount of money and must decide how to divide it between themselves and the other party. The responder can then accept or reject their offer. If the responder accepts, both parties receive the monetary split as proposed. If the responder rejects, neither party receives any money. What should we expect? Fifty years ago, an economist would have argued that the proposer will make the smallest possible offer (say, one cent), and the responder will accept this offer since some money is better than no money. As any social psychologist would have told you, this prediction is not borne out. First, proposers make relatively generous offers (somewhere between 30 and 50 percent of the monetary pie; Camerer, Reference Camerer2003), with an equal 50/50 split being the modal response. These proposers are not acting irrationally; responders who receive less than 20 percent of the pie typically reject the offer, punishing the proposer for behaving in a selfish manner (Cooper & Dutcher, Reference Cooper and Dutcher2011).
That people were making offers that deviated substantially from the smallest possible amount rocked the proverbial boat. Why were people being so generous? One explanation is that the threat of punishment – and fear of not receiving any money – drives this altruistic behavioral pattern. Were proposers unwilling to make low offers out of fear of being rejected and thus not receiving any money? To account for this possibility, Forsythe et al. (Reference Forsythe, Horowitz, Savin and Sefton1994) created the dictator game (DG), which tests this punishment account under stricter conditions. In the DG, there are also two parties. This time, however, the dictator makes the first – and only – move in the game. Because there is only one mover making an offer, the dictator need not fear rejection or punishment, since the receiver has no say in how the money is distributed. What happens in such an unfettered decision space? Proposers give, on average, 20 percent of the original endowment (Camerer, Reference Camerer2003; Engel, Reference Engel2011). Referencing this often cited number, however, misses the fact that offers in the DG are bimodally distributed, which tells a very different story: People either give a great deal (half of the endowment) or do not give at all (Engel, Reference Engel2011). In essence, a good chunk of humanity appears to be spontaneously generous.
One critical component of behaving prosocially is that the actor typically incurs some cost. This can come in many forms, from monetary costs we have outlined to psychological vulnerabilities, which means that, at least for some period of time, the actor is exposed to exploitation from another. This type of moral hazard is encapsulated in the trust game (Berg et al., Reference Berg, Dickhaut and McCabe1995). In this game, one participant (the trustor) is endowed with money and has to decide whether to give money to the other participant (the trustee). Any amount that is given to the trustee is multiplied by a given factor (e.g., tripled or quadrupled). The trustee then has the option to reciprocate by giving some money back, or they can pocket all the money for themselves, leaving the trustor with nothing. In contrast with prior paradigms, the highest possible payoff is only obtained if the trustor invests all their money in the trustee – although only if the trustee reciprocates. The majority of people entrust the other person with at least a portion of their original endowment (Fehr & Fischbacher, Reference Fehr and Fischbacher2003). Critically, the trustee tends to reciprocate, even in one-shot games where there is no chance of repeated play, and the percentage of money returned increases as a function of the amount of money originally trusted (Fehr et al., Reference Fehr, Kirchsteiger and Riedl1993; Hayashi et al., Reference Hayashi, Ostrom, Walker and Yamagishi1999). The more people trust, the more that trust is reciprocated.
Money allocation tasks are another popular method for exploring altruistic tendencies or other-regarding preferences (Charness & Rabin, Reference Charness and Rabin2002). In the classic variant of this task, participants face a list of options that describe two monetary outcomes, one for themselves and one for a peer. Options vary in how good or bad they are for each person in the dyad, from giving all the money to the participant and nothing to the peer, to giving all the money to the peer and nothing to the participant. Results seem to align with stable dispositional tendencies, where altruists mostly give, and egoists mainly care about their own payoff (Murphy & Ackermann, Reference Murphy and Ackermann2014). Moreover, people tend to allocate money in ways that benefit the in-group over the out-group (Balliet et al., Reference Balliet, Wu and De Dreu2014; Ng, Reference Ng1986). Together, research using these economic games paints a rosier picture about humanity’s prosocial tendency than some of the classic early social psychology research (e.g., Milgram and Zimbardo’s studies).
Although there is elegance in the simplicity of these games (mechanistically narrowing the possible motives that can cause prosocial behavior), such reductionist approaches also come with downsides. First, each economic game measures only concerns related to distributive justice – how goods are divided among parties. However, prosocial behaviors can involve other moral motives that stretch beyond distributive justice, including concerns about the well-being of others. Efforts have been made toward creating richer paradigms that better capture other types of prosocial behaviors across contexts (Darley & Batson, Reference Darley and Batson1973), such as modifications to the ultimatum game (FeldmanHall et al., Reference FeldmanHall, Sokol-Hessner, Van Bavel and Phelps2014; Shalvi et al., Reference Shalvi, Handgraaf and De Dreu2011) or the creation of entirely new paradigms that sidestep economic game theory altogether (FeldmanHall et al., Reference FeldmanHall, Mobbs, Evans, Hiscox, Navrady and Dalgleish2012; Lockwood et al., Reference Lockwood, Abdurahman, Gabay, Drew, Tamm, Husain and Apps2021; Lockwood et al., Reference Lockwood, Hamonet, Zhang, Ratnavel, Salmony, Husain and Apps2017). Second, in the original economic testbeds, researchers have artificially narrowed the option set (typically two options are presented: accept/reject and reciprocate/defect, etc.). For example, in the UG, a responder who receives an unfair offer has only one option to signal her disapproval – punish the perpetrator by rejecting the offer. Otherwise, accepting the offer signals that the responder is okay with receiving an unfair offer. Offering a participant only two options certainly simplifies the problem; however, this forced-choice paradigm ignores many contextual factors and social tensions – tensions that we know are critical to shaping behavior in real life (FeldmanHall et al., Reference FeldmanHall, Mobbs, Evans, Hiscox, Navrady and Dalgleish2012; Galizzi & Navarro-Martinez, Reference Galizzi and Navarro-Martinez2019; Levitt & List, Reference Levitt and List2007).
Outside the laboratory, punishment is rarely the only available option for a victim of a transgression. Although punishment is often considered to be anti-social, it can also be conceptualized as prosocial when it prevents immoral actions or enhances moral actions. For example, classic work in behavioral game theory illustrates how the simple threat of punishment in an economic game (i.e., being allowed to punish other players in the game, versus not having the option to punish) enhances cooperation by a magnitude of two (Boyd et al., Reference Boyd, Gintis and Bowles2010; Fehr & Gächter, Reference Fehr and Gächter2000). In other research, allowing people the option to communicate with other players reduces the number of norm violations, such as treating people unfairly or free riding during a public goods game (Jolly & Chang, Reference Jolly and Chang2021). In essence, punishment directly or indirectly through gossiping, serves as an important social oil that helps to maintain or even boost moral actions.
However, research reveals that the literature’s singular focus on punishment as a form of rebalancing the scales of justice or preventing immoral behavior might have been misleading. For example, punishment rates decrease substantially when people have other ways to communicate their disapproval, such as writing an emotional message to the perpetrator (Xiao & Houser, Reference Xiao and Houser2005). Moreover, when the UG is modified to include a wide variety of options for responding to an unfair offer (i.e., the justice game, FeldmanHall et al., Reference FeldmanHall, Sokol-Hessner, Van Bavel and Phelps2014), participants rarely choose to reject (or accept), and there seems to be little desire for punishment altogether (FeldmanHall et al., Reference FeldmanHall, Sokol-Hessner, Van Bavel and Phelps2014; Stallen et al., Reference Stallen, Rossi, Heijne, Smidts, De Dreu and Sanfey2018). In fact, any option that incorporates punishment is not favored. Instead, compensation of the victim, which increases the victim’s payoff without punishing the perpetrator, is consistently chosen 65 percent of the time (Chavez & Bicchieri, Reference Chavez and Bicchieri2013; FeldmanHall, Otto, & Phelps, Reference FeldmanHall, Dunsmoor, Tompary, Hunter, Todorov and Phelps2018; Heffner & FeldmanHall, Reference Heffner and FeldmanHall2019; Mattan et al., Reference Mattan, Barth, Thompson, FeldmanHall, Cloutier and Kubota2020; Son et al., Reference Son, Bhandari and FeldmanHall2019). This strongly counters the classic notion that humans prefer to punish individuals who behave unfairly and instead illustrates that people will preferentially compensate themselves and others (Chavez & Bicchieri, Reference Chavez and Bicchieri2013; FeldmanHall et al., Reference FeldmanHall, Sokol-Hessner, Van Bavel and Phelps2014; Hershcovis & Bhatnagar, Reference Hershcovis and Bhatnagar2017) – without punishing the transgressor – even when punishment is completely free. In other words, so long as a victim’s needs are met, there seems to be little desire to punish a norm violator.
There are other dimensions of prosociality, such as avoiding harm or care for others’ well-being, that stretch beyond distributive and retributive justice concerns. Our lab created a paradigm called the Pain versus Gain paradigm (PvG) that asks participants (Deciders) to choose between benefiting oneself financially or preventing physical harm to another, the Receiver (FeldmanHall et al., Reference FeldmanHall, Mobbs, Evans, Hiscox, Navrady and Dalgleish2012). Deciders choose how much, if any, of a £20 endowment they would like to spend to prevent a series of painful electric shocks from reaching the Receiver – shocks they would observe being administered. The more money Deciders give up, the lower the shock inflicted. Deciders could even give up all their money to ensure no shocks reached the Receiver. They could also do the opposite and keep all the money, ensuring the highest shocks are administered. Deciders make this decision many times, knowing that any money remaining at the end of the experiment is randomly multiplied for a potential maximum payout of £200. Results reveal that, on average, participants forego half their endowment to prevent a stranger from receiving electric shocks (FeldmanHall et al., Reference FeldmanHall, Mobbs, Evans, Hiscox, Navrady and Dalgleish2012). Other labs that have subsequently used this paradigm find similar effects (Gallo et al., Reference Gallo, Paracampo, Müller-Pinzler, Severo, Blömer, Fernandes-Henriques, Henschel, Lammes, Maskaljunas, Suttrup, Avenanti, Keysers and Gazzola2018), and in some cases, individuals are even willing to take the shocks to prevent harm from reaching another (Crockett et al., Reference Crockett, Kurth-Nelson, Siegel, Dayan and Dolan2014).
Across all these testbeds, the emerging picture is that humans routinely engage in a variety of prosocial acts. One critical, but understudied, aspect of prosociality is the role of uncertainty. This is largely a consequence of a concerted effort to remove any motivational ambiguity from economic games (i.e., if one is trying to establish a behavioral effect, free of confounds, it is better to be able to restrict the number of possible explanations for the effect). For example, in the DG, there is little doubt that an individual who offers 10 percent of the pie is behaving selfishly. Alas, in the real world, intentions and outcomes are not always so clear-cut (FeldmanHall & Shenhav, Reference FeldmanHall and Shenhav2019). In fact, when there is the opportunity to plausibly deny immoral behavior, people readily do so. Cheating games offer a unique test of plausibly denying any wrongdoing (Fischbacher & Föllmi-Heusi, Reference Fischbacher and Föllmi-Heusi2013; Gerlach et al., Reference Gerlach, Teodorescu and Hertwig2019). In the most paradigmatic example, people are asked to roll a die, the outcome of which will determine their monetary payoff (1 = $1, 2 = $2 and so on). The experimenter is never present, and thus a participant can report any number. At the individual level, it is impossible to determine if a participant is lying. However, at the group level, we can examine if the average outcome deviates from those expected by chance. Results reveal that people cheat but only by a little (Mazar et al., Reference Mazar, Amir and Ariely2008). The degree to which one cheats is influenced by who is around: A corrupt partner can enhance cheating behaviors (Weisel & Shalvi, Reference Weisel and Shalvi2015) while a consistently honest out-group member can attenuate cheating behavior – it is desirable to appear more virtuous than a moral out-group member (Vives et al., Reference Vives, Cikara and FeldmanHall2022).
Paradigms have also been created to try to capture real-world problems in which prosociality is a necessary antecedent for successfully completing the task. Perhaps the most emblematic example of this dynamic is the current, large-scale societal problem of stymieing climate change. A surge of experiments has created novel tasks that capture important dimensions of the climate change problem, including the exploitation of resources and the need for widespread cooperation across generations to combat such exploitation (Barfuss et al., Reference Barfuss, Donges, Vasconcelos, Kurths and Levin2020; Hauser et al., Reference Hauser, Rand, Peysakhovich and Nowak2014; Milinski et al., Reference Milinski, Sommerfeld, Krambeck, Reed and Marotzke2008; Tavoni et al., Reference Tavoni, Dannenberg, Kallis and Löschel2011). By creating paradigms that map to real-world problems, researchers can gain necessary insight into how to build better interventions that can boost prosocial behaviors in the wild.
To summarize, the past few decades of research conducted in the laboratory have revealed just how common and widespread prosociality is, a behavior that occurs even when there are no possible gains – monetary or reputational (see Section 12.4 on motivations). It appears that humans are well equipped to behave prosocially. Where does this tendency come from? A common question has asked whether prosociality has an innate basis. We now turn to research that investigates the extent to which prosociality is present in other animals and also review the mechanisms posited to explain the evolution of prosociality.
12.2 Our Ancient Prosocial Roots: Apes and Other Mammals
Prosocial behaviors and the need to cooperate to solve a collective problem are not uniquely human. There are many other mammals that coordinate with one another to promote their survival (e.g., in hunting). Evolutionary biologists have looked to our biological ancestors, the great apes, as well as other species, to understand whether human prosociality has an evolutionary origin. An examination of primate behavior and their capacity to relate, connect, and cooperate suggests that characteristics of human prosociality are also found in many of our nonhuman relatives. Traditionally, investigations regarding the capacity for prosocial behaviors are based on observation and field studies (De Waal, Reference De Waal2008). However, one of the most striking examples of altruistic, self-sacrificial behavior occurred in the laboratory, with Masserman’s Reference Masserman, Wechkin and Terris1964 experiment on rhesus monkeys (Masserman et al., Reference Masserman, Wechkin and Terris1964). Fifteen monkeys were trained to pull a cord to receive a piece of food. After training, another monkey was put into an adjacent cage. When the original monkey now pulled the cord to receive a piece of food, a painful electric shock was delivered to the nearby monkey. A number of monkeys consistently preferred to go hungry rather than shock another, with one even refraining from eating for 12 days, presumably to ensure that no shock was administered to the conspecific.
A large and growing body of research indicates that nonhuman primates regularly practice prosocial actions (De Waal & Suchak, Reference De Waal and Suchak2010). Take, for instance, the remarkable helping nature of chimps. Without apparent thought to personal gain, chimps have shown they are more than willing to help not only kin but even unfamiliar humans (Warneken et al., Reference Warneken, Hare, Melis, Hanus and Tomasello2007). When researchers placed a stick beyond the reach of a human but within reach of the chimp, the chimps spontaneously helped the reaching human. Once the person stopped trying to reach for the stick and there was no longer a clear goal requiring a helping action, the chimps stopped assisting. Even costly helping revealed the same results: Chimps were willing to make the effort to climb high to help a person who was reaching for the stick. The difference between these examples and other well-documented prosocial behaviors – like food sharing, grooming, and coalition formation – is the presence of psychological altruism, or caring about another’s welfare enough to deliberately benefit the other simply for their sake. Food sharing, grooming, and other similar activities are all behaviors that fall under biological (reciprocal) altruism (i.e., if you scratch my back, I will scratch yours), behaviors common among both primates and humans (Trivers, Reference Trivers1971). However, it is not always rose-colored glasses, as chimps can also display aggressive behaviors, and most primates, including humans, exhibit both anti- and prosocial tendencies, depending on the context (e.g., aggression to a potential threat; Wilson & Wrangham, Reference Wilson and Wrangham2003).
A major question centers on the existence of shared prosocial motives between human and primate. Evidence of shared prosocial behaviors would suggest that humans acquired preferences for specific actions and motives – such as concerns for equality – throughout their evolutionary history. In humans, concerns for equality lead to cooperative actions, and people are even willing to pay a price to reduce inequality (Fehr & Schmidt, Reference Fehr and Schmidt1999). A similar concern for equality is observed in nonhuman primates, especially for disadvantageous inequality. There is a famous video of a capuchin monkey happily eating a cucumber, which was given as a reward for completing a task. Once a nearby capuchin receives a grape (a much more coveted treat if you are a capuchin) for finishing the same task, the first, formerly happy capuchin takes their cucumber and throws it in the face of the researcher (Brosnan & De Waal, Reference Brosnan and De Waal2003). This strong aversion to inequality has been shown in chimpanzees (Brosnan, Reference Brosnan2006) and even nonprimate species such as domestic dogs (Range et al., Reference Range, Horn, Viranyi and Huber2009).
The modern human capacity to behave prosocially may have even deeper evolutionary roots that go beyond relatively sophisticated motives, such as inequality aversion (Meyza & Knapska, Reference Meyza and Knapska2018; Meyza et al., Reference Meyza, Bartal, Monfils, Panksepp and Knapska2017). Spontaneous altruistic behaviors have been observed in other animals as well, including rats (Bartal et al., Reference Bartal, Decety and Mason2011; Hernandez-Lallement et al., Reference Hernandez-Lallement, Attah, Soyman, Pinhal, Gazzola and Keysers2020) and mice (Burkett et al., Reference Burkett, Andari, Johnson, Curry, De Waal and Young2016), who are willing to put off eating a tasty piece of chocolate to figure out how to free an entrapped cage mate. Even more surprising, the rodent then shares the chocolate with the newly freed cage mate, exhibiting a behavior that encapsulates a common refrain heard from many parents – sharing is caring.
Certain prosocial behaviors – especially those that remove the fitness cost associated with the action – are believed to stem from an evolutionary basis (Axelrod & Hamilton, Reference Axelrod and Hamilton1981; Nowak & Sigmund, Reference Nowak and Sigmund2005; Trivers, Reference Trivers1971). Kin selection, for example, postulates that helping someone who is genetically related comes at no cost since the person or animal who is helped will transmit (through reproduction) part of the genetic information of the one who helped (Hamilton, Reference Hamilton1964). From an evolutionary account, prosocial behaviors among strangers is often explained through the idea of direct reciprocity: There is no cost as long as there is an expectation of repayment (Trivers, Reference Trivers1971) – a theory that has been extended to strangers of strangers, where repayment is accomplished by someone other than the person who helped, a phenomenon known as indirect reciprocity (Nowak & Sigmund, Reference Nowak and Sigmund2005). More recent proposals have suggested other mechanisms, such the survival of the friendliest. This account suggests that Homo sapiens self-domesticated to be prosocial (Hare, Reference Hare2017). Even though there is an ongoing discussion regarding the level in which all these evolutionary adaptions are occurring (e.g., either at the individual or group level; see Rand & Nowak, Reference Rand and Nowak2013 on the mechanisms behind the evolution of cooperation), prosocial tendencies seen in primates appear to be some form of rudimentary “pre-adaptions” of modern-day human prosociality.
12.3 Development of Prosocial Behavior
Another useful insight into understanding how prosociality unfolds is to take a developmental lens, tracking the emergence of prosocial behaviors from infancy to young adulthood. In the past 20 years there has been an explosion of research focused on measuring whether infants express prosocial concerns. The data have converged on the idea that prosociality is present at some of the very earliest stages of life (see Chapter 18, this volume). For example, research reveals that 6- and 10-month-olds have a greater preference for prosocial individuals who help others than antisocial individuals who hinder others (Hamlin et al., Reference Hamlin, Wynn and Bloom2007), and by 10 months, babies can also distinguish between accidental and intentional help (Woo et al., Reference Woo, Steckler, Le and Hamlin2017). These preferences extend to learning about the moral character of others as well: 16-month-olds are sensitive to the food preferences of prosocial puppets that help others but are insensitive to the food preferences of an antisocial puppet that hurt another (Hamlin & Wynn, Reference Hamlin and Wynn2012) – revealing a selective attunement to moral agents. While these results lend support to a nativist account of humans being born with some semblance of a moral calculus, a growing number of mixed findings has led to a multisite collaboration that aims to replicate and clarify the robustness of these effects (Lucca et al., Reference Lucca, Capelier-Mourguy, Byers-Heinlein, Cirelli, Dal Ben, Frank, Henderson, Kominsky, Liberman, Margoni, Reschke, Schlingloff, Scott, Soderstrom, Sommerville, Su, Tatone, Uzefovsky, Wang and Hamlinin press).
At some point in the developmental trajectory, these preferences for prosocial individuals are transformed into prosocial actions enacted by the child itself. One study found that when in their natural settings, toddlers spontaneously comfort others starting around 12–16 months (Zahn-Waxler et al., Reference Zahn-Waxler, Radke-Yarrow and King1983). In the same vein, 14-month-olds help other unrelated adults accomplish a goal, even when the adult does not request help (Warneken & Tomasello, Reference Warneken and Tomasello2006). By 30 months, most toddlers appear to have the capacity (and desire) to spontaneously help others, especially when the altruistic action involves a degree of empathic understanding (Svetlova et al., Reference Svetlova, Nichols and Brownell2010). Around three years of age, children become more selective in their prosocial behavior. For example, children begin to help friends more than other individuals (Engelmann et al., Reference Engelmann, Haux and Herrmann2019) and preferentially help those who are poor versus rich (Essler et al., Reference Essler, Lepach, Petermann and Paulus2020). At about five years, children begin to feel sorry for the misbehavior of other in-group members and will apologize even when the child had nothing to do with the wrongdoing (Over et al., Reference Over, Vaish and Tomasello2016).
What motivates children to help others? There are two widely held views (Hepach et al., Reference Hepach, Engelmann, Herrmann, Gerdemann and Tomasello2022), one that stresses an intrinsic desire borne out of other-regarding preferences and another that highlights a more strategic approach (both of which are detailed in Section 12.4). The more popular of the two has investigated how the development of intrinsic motivators such as empathy act as a necessary precondition to express prosocial actions (Hoffman, Reference Hoffman2000). For example, even newborns show distress for another crying newborn (Dondi et al., Reference Dondi, Simion and Caltran1999). Other work has sought to understand whether this motivation is preserved across childhood, or whether other factors come into play during certain developmental milestones. Recent research demonstrates that while two-year-olds largely appear to be intrinsically motivated to help, by the age of five, other less intrinsically derived strategies begin to be incorporated (Hepach et al., Reference Hepach, Engelmann, Herrmann, Gerdemann and Tomasello2022). In Section 12.4, we review in more detail stable motives and temporary states that have been postulated to explain mechanistically why people behave prosocially.
12.4 Stable and Enduring Motives for Human Prosocial Behavior
While animal research reveals that prosociality can be tracked across species, and developmental work reveals when these tendencies arise, it cannot explain the psychological mechanisms that underlie human prosocial behaviors. Why do people give to charity, lend money to a friend, or, in some extreme cases, place themselves in harm’s way to prevent harm to another? Philosophers, and more recently social scientists, have tried to identify and explain the broad array of internal psychological mechanisms that govern prosociality. Adam Smith famously argued that cooperative behavior stems from the logic of self-interest, not from an intrinsic motivation to be cooperative (Smith, Reference Smith1776). This rather simplified account (which Smith himself contradicts in his book Theory of Moral Sentiments; Smith, Reference Smith1853) has been long echoed, mostly by economists. However, in the second half of the twentieth century, a cottage industry aimed to demonstrate that people behave prosocially even when there is no chance for possible gain. Once this was established (although there are great disparities across societies; Henrich et al., Reference Henrich, Boyd, Bowles, Camerer, Fehr, Gintis and McElreath2001), researchers turned to explaining why some people behave prosocially while others do not and identifying the contexts that help to amplify (or attenuate) prosocial behaviors.
12.4.1 A Risk Perspective
Given that humans are not privy to the intentions, motivations, or beliefs of another, almost every social decision we take is marked by uncertainty (FeldmanHall & Shenhav, Reference FeldmanHall and Shenhav2019). For example, a trustworthy or cooperative individual can easily be exploited by another, less trustworthy individual. It is therefore necessary to constantly estimate the possible risks associated with making a prosocial action. Given this natural mapping between prosociality and uncertainty, there has been some work that explores individual differences in risk attitudes and their effect on altruism, trust, and cooperation (Eckel & Wilson, Reference Eckel and Wilson2004; Fehr, Reference Fehr2009). Early research revealed mixed evidence, with some work demonstrating a positive relationship between prosocial actions and risk attitudes and other work showing no effect whatsoever (Fairley et al., Reference Fairley, Sanfey, Vyrastekova and Weitzel2016). One possible reason for these mixed results was that individual risk attitudes were measured in gambling contexts in which people had access to the known probability associated with a given outcome (e.g., knowing that the probability of winning a lottery is 80 percent). In the social world, we do not have access to the known (or exact) probability of whether another individual will exploit us. The “risk” we perceive in the social domain is more akin to ambiguity – a type of uncertainty in which the probability associated with each outcome is unknown. Once ambiguity, rather than risk, is measured at the individual level, there appears to be a robust relationship between those who can tolerate ambiguity and the tendency to behave prosocially (Vives & FeldmanHall, Reference Vives and FeldmanHall2018). This relationship vanishes once the ambiguity associated with other individuals is resolved, for example, by gaining access to a history of their behavior or by the existence of strong social norms.
12.4.2 A Matter of Reputation
Given that evolutionary forces reinforce the selection of selfishness, reputation concerns have been leveraged in an attempt to reconcile this finding with the existence of prosocial behaviors (Boyd & Richerson, Reference Boyd and Richerson1989; Nowak & Sigmund, Reference Nowak and Sigmund2005). According to the reputation account, prosocial behaviors occur to bolster one’s reputation and maximize payoffs – assuming repetitive interactions among the same people. The price an individual pays for being prosocial is offset by gaining a positive reputation, which likely increases the probability of being treated well by others in the future. In line with this notion, people do behave more prosocially when reputational concerns are at stake (i.e., when they think others are watching; Bradley et al., Reference Bradley, Lawrence and Ferguson2018). By deciding quickly, people can also signal to others that prosocial actions are their default behavior (Jordan et al., Reference Jordan, Hoffman, Nowak and Rand2016). This type of prosocial signaling is useful for encouraging repeated social interactions. People use speedy prosocial responses as a signal to figure out who they should interact with during economic games, preferring those who are “mindlessly” prosocial (Jordan et al., Reference Jordan, Hoffman, Nowak and Rand2016). Critics have argued, however, that reputation as a motivational concern cannot explain a number of findings, including the high rates of prosociality observed when reputation is not at stake (i.e., when nobody is watching or decisions are anonymous). To combat this critique, it has been argued that people internalize the value of prosociality under standard reputation concerns, and, as a consequence, prosocial behavior is automatically endorsed even when nobody is watching (Jordan & Rand, Reference Jordan and Rand2020).
12.4.3 Other-Regarding Motives
Perhaps the most accepted explanation regarding motives for prosocial behavior is the existence of what are considered pure, other-regarding concerns (Cooper & Kagel, Reference Cooper, Kagel, Kagel and Roth2016). The more an individual is concerned about the well-being of others, the more likely they are to act prosocially (Eisenberg et al., Reference Eisenberg, Fabes, Miller, Fultz, Shell, Mathy and Reno1989). This dovetails with common wisdom and everyday observations: People are more prosocial with those they care about (colleagues, friends, and family). It has been theorized that our capacity for prosociality first originated to take care of kin, and, over time, this motive has generalized and extended to other people outside our kinship circles (Christakis, Reference Christakis2019). Research further reveals that the degree to which people value others (as measured by the Social Value Orientation scale; see Murphy & Ackermann, Reference Murphy and Ackermann2014) predicts their prosocial behavior in the laboratory (Balliet et al., Reference Balliet, Parks and Joireman2009). In contrast, stable traits that reflect egoistic motives, such as narcissism, psychopathy, and Machiavellism, are inversely related to prosocial behavioral patterns (Thielmann et al., Reference Thielmann, Spadaro and Balliet2020).
12.5 Temporary States Shaping Prosocial Behavior: The Role of Emotion
The motivations listed earlier are enduring dispositional factors that tend, on the whole, to shape an individual’s relative engagement in prosocial behaviors. However, there are also more micro, temporary factors – such as one’s emotions – that are also capable of enhancing or attenuating an individual’s willingness to behave prosocially (Heffner & FeldmanHall, Reference Heffner and FeldmanHall2022; Keltner et al., Reference Keltner, Horberg, Oveis and Forgas2006). Ever since Hume argued that emotions help to motivate behavior and social coherence (Hume, Reference Hume1896), the idea that emotions act as a fundamental precursor of prosocial behavior has endured (Bailey et al., Reference Bailey, Brady, Ebner and Ruffman2020; Coulombe et al., Reference Coulombe, Rudd and Yates2019; Malti & Krettenauer, Reference Malti and Krettenauer2013; Penner et al., Reference Penner, Dovidio, Piliavin and Schroeder2005). A plethora of evidence indicates that emotion plays a critical role in prosocial behaviors. For example, broad affective states – such as positive moods – can amplify charitable giving (Bohner et al., Reference Bohner, Crow, Erb and Schwarz1992), and specific emotions like contempt, shame, disgust, and guilt play a special role in the establishment of response-dependent valuation, norm compliance, and the motivation to behave in prosocial ways (D’Arms & Jacobson, Reference D’Arms and Jacobson1994). Understanding how emotion and choice interact is crucial to understanding the basic motivations governing human prosociality.
12.5.1 Affective States
Broad affective states, such as positive affect, or transient moods that are not explicitly tied to the causes and consequences of a decision itself, can readily modify how individuals respond when faced with a prosocial dilemma. One early example of this, documented by Isen and Levin (Isen & Levin, Reference Isen and Levin1972), revealed that a positive, happy mood can lead people to help others (Curry et al., Reference Curry, Rowland, Van Lissa, Zlotowitz, McAlaney and Whitehouse2018). To induce a positive mood, half of the subjects discovered a dime in the return slot of a public telephone. Subjects who found the dime then encountered a stranger who dropped a heap of papers in front of them. Subtly inducing a good mood (by way of a found dime – it was the 1970s) led subjects to help the stranger pick up papers, more so than those who were not put in a good mood. In essence, the “glow of goodwill” promoted prosocial behavior. Follow-up studies reported a similar finding, whereby inducing positive affect through emotional comedy videos led people to make more prosocial decisions (DeSteno et al., Reference DeSteno, Bartlett, Baumann, Williams and Dickens2010; Valdesolo & Desteno, Reference Valdesolo and Desteno2006), and reducing negative emotions also increased prosocial behavior (Kemeny et al., Reference Kemeny, Foltz, Cavanagh, Cullen, Giese-Davis, Jennings, Rosenberg, Gillath, Shaver, Wallace and Ekman2012). Additionally, recent research from our lab found that emotion prediction errors play a large role in governing whether an individual will punish a norm violator, revealing that violations of emotional expectations play an outsized role in governing our willingness to act prosocially (Heffner et al., Reference Heffner, Son and FeldmanHall2021). The idea that affective states govern our perceptions, including our social interactions, is gaining traction in the field, with more formal models dissecting how exactly affective states shape how we navigate and interact with our world (Eldar et al., Reference Eldar, Rutledge, Dolan and Niv2016; Quoidbach et al., Reference Quoidbach, Taquet, Desseilles, de Montjoye and Gross2019).
12.5.2 Distress
Other affective states, such as distress, can be tied to the decision to display prosocial behaviors. Watching another experience physical or emotional pain, such as social rejection, is highly aversive and is known to cause distress in the observer (Decety et al., Reference Decety, Michalska and Akitsuki2008; Eisenberger & Lieberman, Reference Eisenberger and Lieberman2004; Singer et al., Reference Singer, Seymour, O’Doherty, Kaube, Dolan and Frith2004), which can in turn amplify prosocial actions. For example, in one study, subjects’ skin conductance responses (SCRs) were measured while receiving painful shocks, or when observing pain being inflicted on another (Hein et al., Reference Hein, Lamm, Brodbeck and Singer2011). Subjects could prevent pain being administered to another by choosing to endure the painful shocks themselves. Critically, the magnitude of the SCR – the body’s physiological response to stressful or arousing events, which was taken here as a proxy for distress – predicted whether a subject would altruistically take the painful shocks in lieu of another.
While the link between distress and altruistic behavior has been replicated in other contexts as well (FeldmanHall et al., Reference FeldmanHall, Dalgleish, Evans and Mobbs2015), there is early theoretical and empirical work that suggests that feelings of distress are also associated with the opposite behavioral response: avoiding helping altogether (Shaw et al., Reference Shaw, Batson and Todd1994). In this case, terminating one’s own distress is more pressing than helping another in need (Batson et al., Reference Batson, Fultz and Schoenrade1987; Eisenberg, Reference Eisenberg2000; Eisenberg et al., Reference Eisenberg, Fabes, Murphy, Karbon, Maszk, Smith, O’Boyle and Suh1994). An example is when a person crosses to the other side of the street after seeing a homeless person on the sidewalk, since avoiding the moral situation attenuates the discomfort that comes with seeing another in distress. This led some researchers to posit that there are two ways to experience distress, either in sharing the pain of another (Cialdini et al., Reference Cialdini, Brown, Lewis, Luce and Neuberg1997; Singer et al., Reference Singer, Seymour, O’Doherty, Kaube, Dolan and Frith2004) or in understanding that another is in distress but without necessarily sharing in the pain itself (Batson, Reference Batson2011; Masten et al., Reference Masten, Morelli and Eisenberger2011). Evidence that individuals who show little arousal to the distress of others also exhibit little physiological arousal when distressed themselves (Shirtcliff et al., Reference Shirtcliff, Vitacco, Graf, Gostisha, Merz and Zahn-Waxler2009), hints at the possibility that the capacity to respond prosocially may be the result of failing to understand what the other is feeling, rather than actually sharing in their distress, per se. However, taking a bird’s-eye view of the data, especially with a nod to more recent neuroimaging work (Zaki & Ochsner, Reference Zaki and Ochsner2012), reveals mixed evidence for this claim. Context, for example, appears to have a strong mediating effect on whether prosociality is driven by understanding or sharing in the pain of others (see Zaki & Ochsner, Reference Zaki and Ochsner2012, for a more in-depth discussion).
Populations who exhibit a reduced ability to understand the distress of another help to clarify this relationship between distress and prosocial choice (K. Blair et al., Reference Blair, Marsh, Morton, Vythilingam, Jones, Mondillo, Pine, Drevets and Blair2006). Alexithymia, a personality construct characterized by the inability to identify and describe emotions in oneself or others (Sifneos, Reference Sifneos1973), showcases the intimate relationship between understanding another’s distress and altruistic action (Bernhardt & Singer, Reference Bernhardt and Singer2012; Taylor & Bagby, Reference Taylor, Bagby, Bar-On and Parker2000). Our lab tested the link between the ability to identify and understand another’s distress and helping behavior. In our Pain versus Gain paradigm explained above, we found that those high on the alexithymia spectrum reported experiencing less distress when watching others in pain, which translated into keeping more money at the expense of the Receiver’s physical well-being (FeldmanHall et al., Reference FeldmanHall, Dalgleish and Mobbs2013). This effect was mirrored at the neural level, with diminished activity in key brain regions that encode for the experience of distress.
12.5.3 Empathy
Empathy is one avenue in which people can understand the emotional experiences of another (Decety, Reference Decety2011; and see Chapter 11 in this volume). The term empathy is applied to a large spectrum of phenomena (Zaki & Ochsner, Reference Zaki and Ochsner2012) and includes a range of emotional dimensions, such as feelings of concern, affect sharing, and cognitive perspective taking. In short, empathy is a multicomponent process that includes both affective and cognitive dimensions (Zaki, Reference Zaki2014). Early foundational research suggested that feeling empathic concern for another in need is the lynchpin of motivated helping (Batson et al., Reference Batson, Fultz and Schoenrade1987). Batson and colleagues tested this empathy-altruism hypothesis (Batson, Reference Batson2011) across a number of studies. They found that increasing the perceived similarity between participants and the person in need of help increases altruistic behavior (Toi & Batson, Reference Toi and Batson1982). Moreover, inducing a person to feel greater empathy in general increases cooperation in the prisoner’s dilemma, regardless of whether the economic game is framed as a social exchange or a business transaction (Batson & Moran, Reference Batson and Moran1999). Empathy might even explain a famous “in the wild” experiment, in which Darley and Batson demonstrated that those who were walking across a university green with ample (rather than no) time on their hands before their next meeting were more likely to offer help to another in need along the way (Darley & Batson, Reference Darley and Batson1973) – perhaps because the helper had enough time to experience or conjure up empathic feelings.
More recent research has focused on investigating empathy in the context of witnessing another in physical pain. The prevailing narrative is that empathy is a form of mirroring another’s distress with our own distress. When we experience physical pain, there are certain brain regions, including the anterior insula, that are routinely activated (i.e., the brain’s prototypical pain circuit). This neural network is also activated when seeing a loved one – or even a stranger – in physical pain (Morelli et al., Reference Morelli, Rameson and Lieberman2014; Singer et al., Reference Singer, Seymour, O’Doherty, Kaube, Dolan and Frith2004). Evidence of shared neural activity suggests that perceiving an emotion in another can activate the same emotion in the observer. This perception–action matching mechanism cannot, however, differentiate between whether the observer is aversively aroused and therefore sharing the distressing experience, or whether the observer is simply recognizing and understanding the distress of the other.
Work from our own lab suggests that understanding the distress of another, rather than sharing in their distress, is the mechanism that fuels the empathy–altruism relationship. Using the same Pain vs Gain paradigm, we found that a Decider’s trait empathy, their ability to express empathic concern toward another, correlated with the amount of money they were willing to forgo to attenuate the number of shocks administered to the Receiver (FeldmanHall et al., Reference FeldmanHall, Dalgleish, Evans and Mobbs2015). This relationship between trait empathy and costly altruism was not reflected by activity in the classic pain circuit (i.e., anterior insula and anterior cingulate). Instead, we found the empathy–altruism relationship to be indexed by activity in the caudate, ventral tegmental area, and subgenual anterior cingulate – regions critical for processing reward and social attachment.
12.5.4 Guilt
As with empathy, guilt can also motivate people to behave in prosocial ways, especially when trying to repair relationships (Donohue & Tully, Reference Donohue and Tully2019; Drummond et al., Reference Drummond, Hammond, Satlof-Bedrick, Waugh and Brownell2017; Vaish et al., Reference Vaish, Carpenter and Tomasello2016). Guilt is a painful feeling aroused by causing (or even anticipating causing) an aversive event (Ferguson & Stegge, Reference Ferguson, Stegge and Bybee1998; Vaish, Reference Vaish2018). Guilt proneness consistently correlates with measures of perspective taking and empathic concern and is inversely related to antisocial behavior (Tangney et al., Reference Tangney, Stuewig and Mashek2007). Neuroimaging studies reveal that when describing moral transgressions, feelings of guilt are associated with neural activation in a network that is also engaged when thinking about another’s feelings (Basile et al., Reference Basile, Mancini, Macaluso, Caltagirone, Frackowiak and Bozzali2011; Takahashi et al., Reference Takahashi, Yahata, Koeda, Matsuda, Asai and Okubo2004). This is taken to indicate that a key function of guilt is to promote perspective taking and increased social cohesion. In fact, someone who expresses feelings of guilt or shame is more likely to be perceived positively than someone who does not express such feelings (Stearns & Parrott, Reference Stearns and Parrott2012). Guilt has also been shown to promote social cooperation. For example, norm infringement is associated with greater experiences of guilt, especially when confronted by an angry face perceived to be criticizing the norm violation (Giner-Sorolla & Espinosa, Reference Giner-Sorolla and Espinosa2011). Because guilt is associated with breaches of moral rule and social standard, the existence of guilt (or even the anticipation of guilt) can foster trust and social reciprocity (Chang et al., Reference Chang, Smith, Dufwenberg and Sanfey2011; Vaish, Reference Vaish2018). Together, this research illustrates how emotions can powerfully influence prosocial behaviors. Simply put, emotions afford a useful and effective control system to stymie antisocial behavior and promote prosocial behaviors.
12.6 The Brain’s Prosocial Network
For the past 20 years, new insights on human prosociality have been gained by directly looking at brain activity through a variety of neuroimaging methods (Bellucci et al., Reference Bellucci, Camilleri, Eickhoff and Krueger2020). Some of the earliest imaging experiments detailed the neural mechanisms associated with specific types of prosocial behaviors. For example, Rilling and colleagues (Rilling et al., Reference Rilling, Gutman, Zeh, Pagnoni, Berns and Kilts2002) recorded brain activity while participants played iterative rounds of the prisoner’s dilemma with the same partner. Results revealed that mutual cooperation between partners was associated with higher activation in the nucleus accumbens, the caudate nucleus, the medial prefrontal cortex (mPFC), and the anterior cingulate cortex. In a follow-up study, the authors found that higher striatal activity was uniquely associated with cooperating with other humans, as opposed to computers (Rilling et al., Reference Rilling, Sanfey, Aronson, Nystrom and Cohen2004). Since the ventral striatum is typically associated with reward, these findings dovetail with the notion that successful cooperation is rewarding in itself, above and beyond just monetary gain (Fehr & Camerer, Reference Fehr and Camerer2007). Similarly, altruism is also associated with activation in reward regions, suggesting that behaving altruistically, without the possibility of economic gain, is experienced in a hedonically rewarding manner (Harbaugh et al., Reference Harbaugh, Mayr and Burghart2007).
As more research followed suit, the brain’s prosocial network began to be mapped in a more granular way. Two primary factors – emotional involvement and taking the perspective of another – appear to be fairly well localized within the brain when making prosocial decisions (Bellucci et al., Reference Bellucci, Camilleri, Eickhoff and Krueger2020). The amygdala emerged as a region sensitive to salient emotional stimuli (Anderson & Phelps, Reference Anderson and Phelps2001; Gangopadhyay et al., Reference Gangopadhyay, Chawla, Dal Monte and Chang2021; Inman et al., Reference Inman, Bijanki, Bass, Gross, Hamann and Willie2020; Phelps & LeDoux, Reference Phelps and LeDoux2005) and, in particular, harmful actions toward another (Berthoz et al., Reference Berthoz, Grèzes, Armony, Passingham and Dolan2006; R. J. R. Blair, Reference Blair2007; Harenski et al., Reference Harenski, Harenski, Shane and Kiehl2010; Kédia et al., Reference Kédia, Berthoz, Wessa, Hilton and Martinot2008). Interestingly, extraordinary altruists (kidney donators) were found to have enhanced volume in the right amygdala (Marsh et al., Reference Marsh, Stoycos, Brethel-Haurwitz, Robinson, VanMeter and Cardinale2014). Recent work linking neural lesions to antisocial behavior reveals the importance of the amygdala in processing the negative value associated with causing harm (Darby et al., Reference Darby, Horn, Cushman and Fox2018). The amygdala has also been implicated in a number of other, related domains, including indexing another’s untrustworthiness (FeldmanHall, Dunsmoor, et al., Reference FeldmanHall, Dunsmoor, Tompary, Hunter, Todorov and Phelps2018; Todorov & Engell, Reference Todorov and Engell2008) and monetary inequality (Haruno & Frith, Reference Haruno and Frith2010). Prosociality also necessitates taking the perspective of another. Here, the research has resoundingly revealed that the temporoparietal junction seems to uniquely serve the capacity for theory of mind (Bednya et al., Reference Bednya, Pascual-Leone and Saxe2009; Cikara et al., Reference Cikara, Jenkins, Dufour and Saxe2014; Gweon et al., Reference Gweon, Dodell-Feder, Bedny and Saxe2012; Richardson et al., Reference Richardson, Lisandrelli, Riobueno-Naylor and Saxe2018; Saxe & Kanwisher, Reference Saxe and Kanwisher2003; Telzer et al., Reference Telzer, Masten, Berkman, Lieberman and Fuligni2011; van Hoorn et al., Reference van Hoorn, Fuligni, Crone and Galván2016) – a finding that is robust to manipulation and context (Schreuders et al., Reference Schreuders, Klapwijk, Will and Güroğlu2018; van Baar et al., Reference van Baar, Halpern and FeldmanHall2021). Furthermore, using transcranial magnetic stimulation, a technique that temporally disrupts the neural activity of targeted brain regions, corroborates the link between experiencing pain and prosocial actions. If the region in the brain that processes pain (i.e., somatosensory cortex) is disrupted, then the relationship between feeling the distress of another in need and lending help to that person is reduced (Gallo et al., Reference Gallo, Paracampo, Müller-Pinzler, Severo, Blömer, Fernandes-Henriques, Henschel, Lammes, Maskaljunas, Suttrup, Avenanti, Keysers and Gazzola2018).
Other factors that can influence the degree to which a person acts prosocially have also been functionally identified within the brain. Factors such as being able to emotionally regulate (Buhle et al., Reference Buhle, Silvers, Wage, Lopez, Onyemekwu, Kober, Webe and Ochsner2014; Etkin et al., Reference Etkin, Büchel and Gross2015; Moll & de Oliveira-Souza, Reference Moll and de Oliveira-Souza2007), identify whether a perpetrator had intent (Berthoz et al., Reference Berthoz, Armony, Blair and Dolan2002; Koster-Hale et al., Reference Koster-Hale, Saxe, Dungan and Young2013; Young & Saxe, Reference Young and Saxe2009; Zapparoli et al., Reference Zapparoli, Seghezzi, Scifo, Zerbi, Banfi, Tettamanti and Paulesu2018), and experience a negative response to the harm of others (Chakroff et al., Reference Chakroff, Dungan, Koster-Hale, Brown, Saxe and Young2016; Kédia et al., Reference Kédia, Berthoz, Wessa, Hilton and Martinot2008) have all been neurally linked to the mPFC and amygdala, revealing a neurobiological road map for many of the behavioral manipulations that can turn the dial on the capacity for behaving prosocially. The mPFC in particular has become robustly associated with the process of weighing up the costs and benefits of partaking in a prosocial action (Hu et al., Reference Hu, Hu, Li and Zhou2021).
One question that framed much of the early imaging work, and which still endures today, is whether there is a distinct neural network specifically dedicated to indexing social interactions, under which prosocial behavior falls (Lockwood et al., Reference Lockwood, Apps and Chang2020). At present, this is a hard question to answer, since most imaging work has focused on linking certain psychological processes to specific brain regions – the question of where cognitive processes are indexed. At first blush, it appears that the same brain regions that encode the value of a cookie also encode the value of another human life. However, these types of analyses can only tell us the story of functional localization (which does seem to occur in a largely domain-general manner), but they are silent on the topic of representation (Kriegeskorte et al., Reference Kriegeskorte, Mur and Bandettini2008). Using recent advances in imaging methods (e.g., multivariate analysis) we can, however, look at how certain brain regions, such as the mPFC, might differentially represent the value associated with a cookie versus human life. Future work can clarify how the brain might uniquely represent social information, especially when deciding to act prosocially.
A shortcoming with imaging research is that it is entirely correlational. A way to fill this gap is to investigate patients with brain damage and track the relationship between impaired behaviors and specific damage within the brain (Fellows et al., Reference Fellows, Heberlein, Morales, Shivde, Waller and Wu2005). This type of lesion research has been instrumental for documenting which brain regions play a causal role in prosocial decision making. We now know that lesions to the dorsal lateral PFC lead to dishonest behavior (Zhu et al., Reference Zhu, Jenkins, Set, Scabini, Knight, Chiu, King-Casas and Hsu2014) and less cooperation (Wills et al., Reference Wills, FeldmanHall, Meager, Van Bavel, Blackmon, Devinsky, Doyle, Luciano, Kuzniecky, Nadkarni, Vazquez, Najjar, Geller, Golfinos, Placantonakis, Friedman, Wisoff and Samadani2018), while lesions to the mPFC increase decisions to punish (Koenigs & Tranel, Reference Koenigs and Tranel2007). Moreover, lesions to the amygdala affect the ability to build interpersonal trust (Koscik & Tranel, Reference Koscik and Tranel2011) and to assess the uncertainty endemic to the social situation (FeldmanHall & Shenhav, Reference FeldmanHall and Shenhav2019). In tandem with more traditional imaging approaches, lesion research paints a fuller picture of just how critical the prefrontal-amygdalar network is for governing prosocial behaviors.
12.7 A Few Caveats
Successful cooperation, interpersonal trust, and altruism depend on a variety of factors, including the environment, the people we are interacting with, our mood, and the personal tendency to be biased toward behaving prosocially. The complexity and number of factors that bias this process make studying prosociality in the lab an enduring challenge, especially because there is no agreed-upon method for examining prosocial behaviors. For example, recent research has cast doubts on the ecological validity of the classical economic games used in the literature, since they are less successful at predicting prosocial behavior outside the laboratory (Galizzi & Navarro-Martinez, Reference Galizzi and Navarro-Martinez2019). In an attempt to move toward more ecologically valid paradigms, there have been a number of recent, creative efforts aimed at investigating behaviors in the real world. For example, one recent study recorded what happens when thousands of wallets stuffed with money are left around the globe. People tend to return the lost wallets and, surprisingly, the rate of returned wallets increases as the amount of money in the wallet increases, spiking to 60 percent when there is more than $90 in the wallet (Cohn et al., Reference Cohn, Maréchal, Tannenbaum and Zünd2019).
While findings like this highlight just how widespread and robust the tendency to act prosocially really is, they also reveal a shortcoming. Most research exploring prosocial behaviors involves money. There are good reasons for this: Money is universally rewarding and easily divisible. Yet, many real-world prosocial acts do not involve money per se but instead require an individual to give their time or effort. Recently, researchers have begun to explore the interplay between effort and prosocial behaviors (Lockwood et al., Reference Lockwood, Abdurahman, Gabay, Drew, Tamm, Husain and Apps2021; Lockwood et al., Reference Lockwood, Hamonet, Zhang, Ratnavel, Salmony, Husain and Apps2017). However, time still manages to be a neglected factor, probably because it is difficult to run experimental longitudinal studies. Individuals do not become prosocial overnight but instead learn to be prosocial over time and through repeated interactions. In just the last few years, however, research has moved toward characterizing how individuals leverage trial-by-trial learning to titrate their level of prosocial behavior so that it is both useful to others but not too costly to themselves (FeldmanHall, Dunsmoor, et al., Reference FeldmanHall, Dunsmoor, Tompary, Hunter, Todorov and Phelps2018; Lamba et al., Reference Lamba, Frank and FeldmanHall2020).
12.8 Concluding Remarks
This chapter provided a general overview of the study of prosociality in the laboratory, the predecessors of prosocial behaviors observed in the animal kingdom and during the early formative years in humans, the different motives and psychological constructs associated with behaving prosocially, and a summary of the insights neuroscience has provided us with on the neurobiological mechanisms subserving this complex process. The human capacity to be kind, generous, and altruistic with one another has historically surpassed their capacity for selfish, harmful behavior. Indeed, as we enter the “golden age of social sciences” (Buyalskaya et al., Reference Buyalskaya, Gallo and Camerer2021), the future success of the inquiry of prosociality relies on our capacity for widespread cooperation.
This chapter critically reviews the existing literature on antisocial behavior and morality from a moral psychology perspective, aiming to integrate disparate strands of prior research into an overarching perspective. Throughout this chapter, particular attention is given to psychopathy, a mental disorder characterized by the propensity for immoral, antisocial behavior. Following a brief introduction, this chapter poses the initial question of whether individuals on the antisocial spectrum do indeed show an impairment in moral reasoning or, alternatively, they have emotional impairments that interfere with their ability to act in a morally appropriate manner. We then turn to the neural foundations of moral decision making and antisocial behavior. A neurobiological model of both antisocial and moral behavior is presented, together with its relevance to various subforms of antisociality. Interventions aimed at enhancing the moral sense of the individual and reducing antisocial behavior are briefly outlined. Finally, directions for future research are provided.
13.1 The Antisocial Spectrum of Disorders
Antisocial personality disorder (APD) was initially described by Phillipe Pinel (Reference Pinel1801) to characterize patients with severely impulsive and destructive behavior but who otherwise did not show any evidence of mental disorder or psychosis. He called the condition manie sans délire (insanity without delirium). J. C. Prichard (Reference Prichard1835) later described the condition as “moral insanity,” a gross disturbance in social behavior without any impairment in mental functioning (Livesley et al., Reference Livesley, Schroeder, Jackson and Jang1994). This label is intended to convey the notion that although those with this condition are not legally insane, they lack self-control and have a propensity to harm others, a moral equivalent of insanity. As late as the turn of the twentieth century, “psychopathic” or “psychopathological” referred to any mental disorder, and the term “psychopathic inferiority” was introduced to refer to various deviations of personality (Koch, Reference Koch1891). Kahn (Reference Kahn1931), Kraepelin (Reference Kraepelin1913), and Schneider (Reference Schneider1923) proposed an alternative classification of personality disorders that included people classified with destructive personalities who cause problems for themselves and the wider society they live in. The term “psychopathy” was used, albeit inconsistently, to refer to any type of personality disorder or antisocial or aggressive personality. The American Psychiatric Association (APA, 1952) added the term “sociopathic personality disturbance” to the first edition of the Diagnostic and Statistical Manual of Mental Disorders (DSM-I) in the 1950s and the term “antisocial personality” was introduced in later editions.
According to the fifth edition of the DSM (DSM-V), all 10 personality disorders are grouped into three clusters (A, B, and C). Antisocial personality disorder (APD) falls into cluster B, along with borderline, narcissistic, and histrionic. All these disorders normally present with dramatic, emotional, and unpredictable interactions with others (Fisher & Hany, Reference Fisher and Hany2021). The unique essential feature of APD is the persistent violation or neglect of the rights of others. The seven subfeatures of APD are law-breaking, lying and deceiving, impulsivity, irritability and aggression, disregard for the safety of others, irresponsibility, and lack of remorse for actions. A diagnosis of APD requires that the person be at least 18 years old. A distinctive feature of APD that separates it from all other personality disorders is its developmental progression. Of all the personality disorders in the DSM-V, APD is the only one that stipulates a developmental course among its diagnostic criteria, including a history of conduct disorder (CD) before age 15 (APA, 2013). In turn, CD is invariably preceded by oppositional defiant disorder (ODD), which can have symptoms appearing as early as the preschool years (APA, 2013). Therefore, one can expect development over time from hostility, uncooperativeness, and symptoms of ODD in early childhood to more significant aggression, impulsivity, deceitfulness, and rule violations of CD in later childhood and adolescence, progressing from there to more serious APD in early adulthood (Caspi et al., Reference Caspi, Moffitt, Newman and Silva1996).
Psychopathy is commonly viewed as a disorder lying on the extreme of the antisocial spectrum (Hare & Neumann, Reference Hare and Neumann2008). Individuals with psychopathy are notorious for their amoral behavior, characterized by lack of empathy and guilt, shallow affect, manipulation of other people, and severe, premeditated, and violent antisocial behavior (Hare & Neumann, Reference Hare and Neumann2008). Although individuals with criminal psychopathy share features such as impulsivity with other antisocial offenders, the factor which psychopathy researchers often highlight to set them apart from others is a relative absence of guilt for their harmful behavior (Hare, Reference Hare1999). They show little concern for the suffering of their fellow human beings, form shallow affiliative bonds, and rarely show loyalty unless it is in their own interest (Kiehl, Reference Kiehl2015).
Psychopathy is one of the best predictors of criminal offending and reoffending, with psychopathic individuals being three times as likely as other offenders to recidivate. Males are disproportionately affected, with an estimated four males with psychopathy to one female (e.g., Vitale et al., Reference Vitale, Smith, Brinkley and Newman2002). The most extensively used measure of psychopathy is the Psychopathy Checklist–Revised (PCL-R; Hare, Reference Hare1991). This “gold standard” assessment tool consists of a semistructured interview combined with a review of correctional institution charts. The PCL-R contains 20 items assessed on a three-point scale, with item scores summed to obtain a total dimensional score ranging from 0 to 40, where higher scores indicate more psychopathic features. It provides a broad assessment of several dimensions of psychopathy: 1) the affective and interpersonal traits key to classic definitions of psychopathy (e.g., shallow affect, superficial charm, manipulativeness, lack of empathy), and 2) the lifestyle and antisocial dimension (e.g., criminal versatility, impulsiveness, irresponsibility, poor behavior controls, juvenile delinquency). Extensive research has confirmed that the PCL-R is a reliable and valid measure of psychopathy (Hare, Reference Hare2003).
In explaining the moral abnormality of psychopathy, some authors suggest that the antisocial behavior characteristics of the psychopathic individual derive from a failure of moral rationalization (e.g., Blair, Reference Blair1995), which may result from a lack of the cognitive capacity to tell right from wrong (e.g., Fiedler & Glöckner, Reference Fiedler and Glöckner2015). In contrast, others see affective processes as the main driver of moral decisions and suggest that while psychopathic individuals demonstrate fairly normal moral reasoning abilities (e.g., Aharoni et al., Reference Aharoni, Sinnott-Armstrong and Kiehl2012; Cima et al., Reference Cima, Tonnaer and Hauser2010), aberrant emotional processing plays a crucial role in their moral incompetence (e.g., Harenski et al., Reference Harenski, Harenski, Shane and Kiehl2010; Newman et al., Reference Newman, MacCoon, Vaughn and Sadeh2005). Which of these accounts is correct constitutes a key question in the literature, which this review aims to resolve. (For a related discussion of psychopathy and moral reasoning, see Chapter 3, this volume.)
13.2 Can the Psychopathic Individuals Tell Right from Wrong? The Rational Deficit Hypothesis
A prominent explanation for psychopathic antisocial behavior is that individuals cannot distinguish cognitively between right and wrong (Blair, Reference Blair1995, Reference Blair1997; Garrigan et al., Reference Garrigan, Adlam and Langdon2018). This proposition bears on the legal and philosophical debate about the psychopathic individual’s moral responsibility and whether an assessment of psychopathy should qualify as a mitigating circumstance in criminal cases (see Aharoni et al., Reference Aharoni, Funk, Sinnott-Armstrong and Gazzaniga2008; Blair, Reference Blair2008; Fine & Kennett, Reference Fine and Kennett2004; Levy, Reference Levy2007; Litton, Reference Litton, Kiehl and Sinnott-Armstrong2013; Morse, Reference Morse2008; Pillsbury, Reference Pillsbury, Kieh and Sinnott-Armstrong2013). This issue is critical because “rational capacity” – knowing what you are doing at the time of the act and that the act in question is wrong – forms the legal basis for culpability.
The moral judgment of the psychopathic individual has been examined using a variety of measurement tools. The four most commonly used approaches are: 1) measures of Kohlbergian moral reasoning (Kohlberg, Reference Kohlberg1963), which typically ask people to rank the reasons they believe are most relevant to deciding how to respond to moral dilemmas; 2) the moral-conventional transgressions distinction (Nucci & Nucci, Reference Nucci and Nucci1982; Nucci & Turiel, Reference Nucci and Turiel1978; Turiel, Reference Turiel1979, Reference Turiel1983), which challenges respondents to draw a line between acts that are considered wrong merely by social convention versus those that are morally wrong; 3) sacrificial moral dilemmas (Greene et al., Reference Greene, Sommerville, Nystrom, Darley and Cohen2001), in which participants decide how one would act in a situation that entails harming one or more people to avoid harming a larger group of individuals, such as authorizing the death of one person to save several others; and 4) moral foundations questions (Haidt & Graham, Reference Haidt and Graham2007), in which participants indicate which moral concerns, such as purity and harm, are most relevant to their moral judgment.
These four methods of measuring moral judgment, while conceptually overlapping, also provide distinctive insights into how the psychopathic individual may differ in their moral judgment. The Kohlbergian measures focus on the reasons why one chooses a certain course of action and are thus characterized as indices of different stages of moral reasoning. Moral-conventional distinctions determine whether moral judgment relies on conventional rules that are contingent, local, and enable social coordination through communal understandings, or by moral rules that could be applied universally and hold independently of the expectations and commands of political or social authorities. Sacrificial moral dilemmas are a type of ethical problem used to study how individuals make tough moral decisions, particularly when they must choose between competing ethical principles. Moral foundation measures are based on the factors people consider when deciding whether an action is morally permissible. Next we examine each perspective in turn to clarify the mixed state of the literature on psychopathy and moral cognition.
13.2.1 Evidence from Kohlberg’s Test of Moral Development
Initial research in this area focused exclusively on assessing the psychopathic individual’s developmental level of moral reasoning, as per Kohlberg’s (Reference Kohlberg1963) theory of moral development. This theory outlines three overarching moral stages: 1) preconventional, 2) conventional, and 3) postconventional, with each stage subdivided further into two substages. In each overarching stage, Kohlberg proposed that people invoke various considerations when making moral decisions. Individuals at the preconventional stage decide on moral issues based on the immediate consequences for themselves, while those at the postconventional stage make decisions based on relatively abstract moral principles independent of existing social rules, laws, and authority. Those in the intermediate, conventional stage, make moral decisions based on the expectations of social groups and society. One well-validated and widely used questionnaire measure within Kohlberg’s moral development framework is the defining issues test (DIT; Rest et al., Reference Rest, Copper, Coder, Masanz and Anderson1974), which presents five moral dilemmas. One well-known example, the Heinz drug dilemma, asks participants whether a man should steal a drug to save his dying wife’s life when there are no other options, although doing so is against the law. The DIT yields an overall moral development score, with higher scores representing a higher level of postconventional reasoning.
Researchers have used Kohlbergian measures of moral reasoning to understand which elements of a psychopathic individual’s moral cognition may be compromised. Given that psychopathy is marked by self-centeredness, some authors (e.g., Campbell et al., Reference Campbell, Doucette and French2009) have hypothesized that psychopathic individuals tend to prioritize self-interest over abstract moral principles when making moral decisions and, therefore, do not progress beyond the lower levels of moral development (Campbell et al., Reference Campbell, Doucette and French2009; O’Kane et al., Reference O’Kane, Fawcett and Blackburn1996). Surprisingly, the research evaluating this argument is dated and has yielded inconsistent results. Some studies have supported the hypothesis of lower levels of moral reasoning in psychopathic individuals (Fodor, Reference Fodor1973; Jurkovic & Prentice, Reference Jurkovic and Prentice1977), some argue for higher levels of moral reasoning (Link et al., Reference Link, Scherer and Byrne1977), while others reveal no significant group differences on moral development (Lose, Reference Lose1977; O’Kane et al., Reference O’Kane, Fawcett and Blackburn1996). Supporting the hypothesis that psychopathic individuals reason differently from others about moral issues, Campbell and colleagues (Reference Campbell, Doucette and French2009) found that such individuals tend to both prioritize lower moral considerations (e.g., potential threats to personal interest) and de-emphasize higher moral reasons (e.g., attempts to protect human rights). Adding to the confusion, Link and colleagues (Reference Link, Scherer and Byrne1977) found that psychopathic individuals received significantly higher moral reasoning scores than did nonpsychopathic individuals on the Kohlberg (Reference Kohlberg1958) test. However, later researchers raised the criticism that Kohlberg’s test relies heavily on good analytical skills and completely neglects the evidence for the powerful emotional and nonverbal determinants of morality (Garrigan et al., Reference Garrigan, Adlam and Langdon2018). On balance, these findings provide only a cloudy and inconclusive picture of the moral-developmental level of the psychopath.
13.2.2 Evidence from the Moral-Conventional Distinction
Another form of the moral reasoning hypothesis relies on the findings of Blair et al. (Reference Blair, Jones, Clark and Smith1995, Reference Blair, Jones, Clark and Smith1997) based on the moral/conventional task (MCT). In these studies, investigators assessed the ability of adults and children who were low or high in psychopathic tendencies to correctly classify eight hypothetical actions based on their moral content. Using the MCT, Blair et al. (Reference Blair1995, Reference Blair1997) asked participants to judge whether 1) the hypothetical action was acceptable, and 2) whether it would still be acceptable even if an authority figure permitted it under free response conditions. Conventional transgressions presuppose that the perceived wrongfulness is reliant on what the authority says (e.g., wearing pajamas at work). In contrast, moral transgressions are defined as acts that society considers wrong even if there were no rules, customs, or laws against them (e.g., hurting others). Participants low in psychopathy correctly distinguished moral scenarios that were authority-independent from conventional scenarios. However, participants with high psychopathy made no such distinction. It is notable that psychopathic participants who failed to make moral-conventional distinctions in Blair’s (Reference Blair1995) studies did not rate both types of scenario as acceptable. Rather, they tended to rate both types as unacceptable, regardless of the conventions laid down by authority, suggesting that these participants believed all the scenarios to involve moral violations. Blair (Reference Blair1995) concluded that individuals with psychopathy exhibited this counterintuitive effect as the result of attempting to appear socially desirable or “faking good” (see Blair, Reference Blair1995, p. 23; Blair et al., Reference Blair, Jones, Clark and Smith1995, p. 749) and that in reality they may have difficulty detecting or rating merely conventional transgressions.
Recent studies on psychopathic individuals and the moral-conventional distinction have not reproduced Blair’s original results. Aharoni et al. (Reference Aharoni, Sinnott-Armstrong and Kiehl2012, Reference Aharoni, Sinnott-Armstrong and Kiehl2014) tested psychopathic individuals’ ability to distinguish between moral and conventional transgressions in a forced-choice setting, telling participants that exactly half the scenarios were considered morally wrong by typical members of society. Respondents were expected to make a forced choice between predefined moral and conventional transgressions, meaning overclassification would not be an effective strategy and allowing a purer assessment of moral reasoning. Under this methodology, if psychopathic participants truly lack moral understanding they should show less accuracy in making moral-conventional classifications than nonpsychopathic controls. Conversely, if psychopathic individuals are no less accurate in these forced-choice classifications, we could conclude that they do indeed understand moral wrongfulness. In both studies, no association was established between psychopathy scores and the percentage of acts correctly classified as moral or conventional. In sum, when factors relating to social desirability are removed, psychopathic participants in moral-conventional distinction studies perform as well as controls, providing additional evidence that individuals with psychopathy do indeed understand wrongfulness (Aharoni et al., Reference Aharoni, Sinnott-Armstrong and Kiehl2012, Reference Aharoni, Sinnott-Armstrong and Kiehl2014).
13.2.3 Evidence from Sacrificial Moral Dilemmas
Another commonly used method for measuring moral decision making draws upon the seminal work of Greene and colleagues (Reference Greene, Sommerville, Nystrom, Darley and Cohen2001). They used hypothetical moral dilemmas in which participants must decide whether to sacrifice the life of one person in order to save the lives of a greater number (for a broader discussion of moral dilemmas, see Chapter 5, this volume). In a version of the classic trolley problem (Thomson, Reference Thomson1976), the switch dilemma, participants must choose whether or not to pull a lever that would divert a train and save five others who are lying on the track ahead but also kill one person at the end of the diversion track. In another sacrificial dilemma, the footbridge dilemma, participants must choose whether or not to push a heavyset man off a bridge in order to stop the train, which would kill him but thereby save five others on the track ahead. Although the participants in both cases have to decide whether to cause the death of one person to save several, endorsement of the action in either situation indicates a willingness to engage in utilitarian decision making (i.e., saving five people instead of one), while avoiding action in either case indicates a deontological decision (i.e., considering it immoral to inflict harm). However, while avoiding action in the footbridge dilemma clearly aligns with deontological principles (i.e., considering it immoral to directly inflict harm), the deontological decision in the switch case is more complex. Pulling the lever can be seen as consistent with the maxim of universalizability and a duty to save lives, making the deontological judgment in this scenario less clear-cut.
Interestingly, most individuals deem the “pulling the lever” – the switch dilemma – to be more morally acceptable than pushing a person onto the tracks, despite both choices having identical outcomes in terms of numbers of lives lost (Cushman et al., Reference Cushman, Young and Hauser2006). One explanation for this discordance in moral decision making across these dilemmas postulates that most people think pulling the lever does not cause direct harm to someone else whereas pushing the man off the footbridge causes direct physical harm to another individual (Cushman et al., Reference Cushman, Gray, Gaffey and Mendes2012). This distinction is crucial because causing direct harm typically elicits stronger emotional responses, such as empathy and guilt. People are more likely to experience and be influenced by these emotions when contemplating direct harm. Psychopathic individuals, who have deep-rooted deficits in empathy and guilt (Berg et al., Reference Berg, Lilienfeld and Waldman2013), may not be as affected by the emotional weight of causing direct harm. As a result, they might adopt more utilitarian reasoning, focusing on the outcomes (i.e., saving more lives) rather than the means of achieving those outcomes. The reduced influence of emotions in individuals with psychopathy may enhance the more “rational” aspects of their moral decision making over those of nonpsychopathic individuals. The reduced influence of emotions in individuals with psychopathy may serve to enhance the more “rational” aspects of moral decision making over those of nonpsychopathic individuals (Pletti et al., Reference Pletti, Lotto, Buodo and Sarlo2017).
Supporting this prediction, Bartels and Pizarro (Reference Bartels and Pizarro2011) found that self-reported psychopathy was positively associated with utilitarian decision making about sacrificial dilemmas among undergraduates. Koenigs et al. (Reference Koenigs, Kruepke, Zeier and Newman2012) partially replicated this correlation in a sample of prisoners, albeit specifically in the case of the classical switch dilemma where only indirect harm is inflicted. Tassy et al. (Reference Tassy, Deruelle, Mancini, Leistedt and Wicker2013) replicated the study with French university students and showed that a high level of psychopathic traits predicted a greater likelihood of a utilitarian response to hypothetical choices (i.e., “Would you … in order to …?”) but not for the judgment question (i.e., “Is it acceptable to … in order to …?”). This finding confirmed the link between psychopathic traits and utilitarianism in choice of action but not in moral judgment.
13.2.4 Moral Foundations Theory
A final way in which researchers have examined the moral abnormality of psychopathy emerges from work on moral foundations. Six moral domains are argued to represent the various ethical considerations of decision making (Fernandes et al., Reference Fernandes, Aharoni, Harenski, Caldwell and Kiehl2020; Haidt, Reference Haidt2012): 1) preventing harm, which involves concern about the suffering of others; 2) preserving fairness, which involves following the norms of reciprocity, equality, and justice; 3) respecting authority, which involves moral obligations to hierarchies; 4) exhibiting in-group loyalty, which involves moral obligations to members of an identified group without betrayal and giving preferential treatment to in-group over out-group members; 5) practicing purity, which involves living in a noble manner with purity of body, mind, and soul; and 6) protecting liberty, which involves the feelings of reactance and resentment people feel toward those who dominate others and restrict their freedom.
In the context of moral judgments, moral foundations research is focused primarily on preferences in moral values as opposed to identifying deficits in moral judgment patterns. A study conducted by Blair (Reference Blair2007) showed that while individuals scoring high in psychopathy demonstrated a lack of concern about harming others, they had similar levels of concern within other domains, such as purity and loyalty in the MFQ (Graham et al., Reference Graham, Nosek, Haidt, Iyer, Koleva and Ditto2011). This finding was replicated in a sample of 222 incarcerated males, where it was found that psychopathic traits were negatively associated with concerns about harm and fairness (Aharoni et al., Reference Aharoni, Antonenko and Kiehl2011). Others, in contrast, have found that psychopathy is related to diminished moral concern in all five moral domains (Jonason et al., Reference Jonason, Strosser, Kroll, Duineveld and Baruffi2015). Using a large online sample, Glenn et al. (Reference Glenn, Iyer, Graham, Koleva and Haidt2009) found that individuals who scored higher on the psychopathic traits reported less concern about preventing harm and being fair than those who scored lower on psychopathy. The psychopathic individuals also showed slightly increased concerns about in-group loyalty, which may be due to their reduced concern about nongroup members and their desire for their in-group to socially dominate other groups. Although the body of work relating psychopathy to moral foundations theory (MFT) is relatively small, overall these studies suggested that psychopathy is associated with moral abnormality.
13.2.5 Meta-analysis on Moral Judgment and Psychopathy
Taking different perspectives and using various measures of moral judgment, the previous sections have tried to address the question of whether individuals with psychopathy can distinguish cognitively between right and wrong. Although the results of these findings are mixed, they generally point to the conclusion that psychopathic individuals have the capacity to make moral judgment (Borg & Sinnott-Armstrong, Reference Borg, Sinnott-Armstrong, Kiehl and Sinnott-Armstrong2013). Recently, Marshall et al. (Reference Marshall, Watts and Lilienfeld2018) conducted a meta-analysis of 27 studies, systematically analyzing the results of previous research to derive conclusions about the relationship between psychopathy and moral reasoning. Drawing from these studies (N = 4,376) of psychopathy using measures including self-reporting and common moral tasks (i.e., sacrificial moral dilemmas, Kohlbergian moral reasoning, and the moral foundations questionnaire [MFQ]), they detected a small but statistically significant relationship between psychopathy scores and the commonly used measures of moral decision making (rW = 0.16) and moral reasoning (rW = 0.10). However, very little evidence of “pronounced and overarching” moral deficits in conjunction with psychopathic traits is presented. It is possible that reports of marked deficits are overstated due to publication bias (Marshall et al., Reference Marshall, Watts and Lilienfeld2018). The authors suggested that their results raise the distinct possibility that the mental health profession and laypeople have underestimated the capacity of psychopathic individuals to appreciate moral responsibility.
Six studies within the meta-analysis used the MFQ created by Graham et al. (Reference Graham, Nosek, Haidt, Iyer, Koleva and Ditto2011). The results suggest that, like nonpsychopathic individuals, psychopathic individuals emphasize nearly all moral foundations they relate to equally. At the same time, while the slightly (but nonsignificantly) stronger scores on the harm subscale of the MFQ do not provide strong statistical support for the hypothesis (Glenn et al., Reference Glenn, Iyer, Graham, Koleva and Haidt2009) it hints at a trend that warrants further investigation. Overall, the meta-analysis from Marshall et al. (Reference Marshall, Watts and Lilienfeld2018) provides evidence against the idea that psychopathic individuals exhibit a pronounced or predominant deficit of moral reasoning (Furnham et al., Reference Furnham, Daoud and Swami2009), suggesting instead that psychopathic individuals exhibit subtle differences in moral judgment and emphasize slightly different moral foundations for making decisions than nonpsychopathic individuals.
In summary, as most studies fail to identify consistent or systematic differences in judgment capabilities between individuals with psychopathy and nonpsychopathy, there is good reason to believe that psychopathic individuals possess the capacity for making normative moral judgments. Though psychopathic individuals appear to have the ability to form a rational judgment as to whether an act is right or wrong, at a behavioral level they lack the capacity to follow through with what they reason to be right or wrong (or are prone to acting against moral norms out of self-interest). In other words, they are incapable of holding genuine moral concepts as a guide to behavior and thus fail to translate moral judgments into action. In Section 13.3, we move on from the role of reasoning and deliberation in moral judgment to the role of emotion processing and discuss how emotions contribute to moral decision making.
13.3 Can Individuals with Psychopathy Act on What Is Right? The Role of Emotion Processing
Some researchers believe that there is a strong connection between emotion and judgment of moral behavior (see Horne & Powell, Reference Horne and Powell2016). Such processing involves not only the ability to experience negative emotions around the understanding of right and wrong but also the ability to recognize others’ affect (e.g., empathy; Nichols, Reference Nichols2004; Tuvblad et al., Reference Tuvblad, Bezdjian, Raine and Baker2013). In the following subsections, we explore how emotion processing differs in individuals with psychopathy and how such differences likely affect moral behavior.
13.3.1 Shallow Emotional Experience
A fundamental reason why most individuals refrain from committing crimes relates to moral emotions. Feelings of anticipated shame, guilt, and remorse that are aversive and unpleasant serve to buffer individuals from offending (Rebellon et al., Reference Rebellon, Piquero, Piquero and Tibbetts2010; Tangney et al., Reference Tangney, Stuewig and Hafez2011). But what if an individual is unable to experience these emotions? Psychopathic personality is characterized by shallow emotional experiences and an inability to experience the full range and depth of normal emotions (Hare & Neumann, Reference Hare and Neumann2008; Ribeiro da Silva et al., Reference Ribeiro da Silva, Rijo and Salekin2012). Such emotional incapacity in psychopathic individuals may prevent them from relating empathically to others.
In an explication of early theories of psychopathy, Blackburn (Reference Blackburn and Patrick2006) put forward the idea that psychopathic individuals have a deficit in participating in role-taking within social groups and are insensitive to the needs of others. He argued that they are prone to ignore the rights of others and show a general inability to form lasting emotional bonds, all relevant to this key failure to assume the normal range of role-taking responsibilities that most people do quite easily. Without concern for others, individuals with psychopathy lack the capacity to make authentic moral judgments based on empathy and can only rationalize such judgments in the abstract. This basic ability to empathize with another can be viewed as foundational to all other emotions that one experiences in reaction to what people do to others. Essentially, the reason why we think it is wrong to harm others is because when we contemplate such harm, we feel part of the pain of the victim through an empathic affective reaction. This basic propensity therefore gives rise to other, more recognizable moral emotions, such as anger at injustice. A deficit in this ability to experience empathy among psychopathic individuals is thought to disrupt their emotional connectedness to others and interfere with the development of a moral conscience (DeLisi et al., Reference DeLisi, Vaughn, Gentile, Anderson and Shook2013; Poon & Ho, Reference Poon and Ho2015) and the internalization of moral standards of behavior (Blair, Reference Blair1995).
13.3.2 Facial Affect Recognition Deficits
Studies of emotion processing have shown that individuals with psychopathy not only display shallow emotional experience but also exhibit poor accuracy in their recognition of others’ emotions. Processing facial affect is crucial for socialization and normal social interaction (Corden et al., Reference Corden, Critchley, Skuse and Dolan2006; Fridlund, Reference Fridlund1991). Antisocial and aggressive behaviors may stem from an inability to correctly interpret the social and emotional cues of others (Blair, Reference Blair2003; Montagne et al., Reference Montagne, van Honk, Kessels, Frigerio, Burt, van Zandvoort, Perrett and de Haan2005; Poon, Reference Poon2016; Walker & Leister, Reference Walker and Leister1994). Some studies have suggested that this deficit appears to be specific to the emotions of fear and sadness, as demonstrated in a “violence inhibition” paradigm proposed by Blair and colleagues (Blair et al., Reference Blair, Mitchell, Peschardt, Colledge, Leonard, Shine, Murray and Perrett2004; Marsh & Blair, Reference Marsh and Blair2008).
In this model, the absence of fear and sadness when faced with others’ distress promotes violent behavior (Blair et al., Reference Blair, Budhani, Colledge and Scott2005). The core idea of Blair’s model (Blair, Reference Blair1995, Reference Blair2006) lies in classical conditioning theory. People typically find the fear and sadness cues of others to be innately aversive, so that when an antisocial behavior is followed by fear or sadness, the antisocial action itself becomes aversive and is inhibited. It follows that if an individual is impaired in recognizing fear and sadness, this allows them to behave in a more self-gratifying manner without the aversive effect of “feeling bad.” This explanation is supported by a meta-analysis conducted by Marsh and Blair (Reference Marsh and Blair2008), which suggests that psychopathy is associated with a specific deficit in recognizing fear and sadness, rather than a universal deficit in recognizing all emotions.
Some studies have found that psychopathic individuals showed impairments in the recognition of fear (Fairchild et al., Reference Fairchild, Van Goozen, Calder, Stollery and Goodyer2009) and sadness (e.g., Fairchild et al., Reference Fairchild, Stobbe, Van Goozen, Calder and Goodyer2010; Fairchild et al., Reference Fairchild, Van Goozen, Calder, Stollery and Goodyer2009; Hastings et al., Reference Hastings, Tangney and Stuewig2008) based on facial expressions. Others have also found deficits in recognizing fear by measuring callous-unemotional (CU) traits (Leist & Dadds, Reference Leist and Dadds2009; Muñoz, Reference Muñoz2009). There are, however, also numerous studies that did not find psychopathic traits to be associated with specific deficits in recognizing facial expressions of fear (e.g., Eisenbarth et al., Reference Eisenbarth, Alpers, Segrè, Calogero and Angrilli2008; Fairchild et al., Reference Fairchild, Stobbe, Van Goozen, Calder and Goodyer2010; Hastings et al., Reference Hastings, Tangney and Stuewig2008) or sadness (Del Gaizo & Falkenbach, Reference Del Gaizo and Falkenbach2008; Hansen et al., Reference Hansen, Johnsen, Hart, Waage and Thayer2008; Leist & Dadds, Reference Leist and Dadds2009; Muñoz, Reference Muñoz2009). Moreover, Del Gaizo and Falkenbach (Reference Del Gaizo and Falkenbach2008) found that psychopathy was associated with better recognition of fear, and a trend in this direction was observed by Woodworth and Waschbusch (Reference Woodworth and Waschbusch2008). Other studies have also found evidence that psychopathic traits are also associated with deficits in recognizing other emotions such as disgust (Hansen et al., Reference Hansen, Johnsen, Hart, Waage and Thayer2008), anger (in relation to CU traits; Muñoz, Reference Muñoz2009), happiness (Hastings et al., Reference Hastings, Tangney and Stuewig2008), and surprise (Fairchild et al., Reference Fairchild, Van Goozen, Calder, Stollery and Goodyer2009), calling into question whether psychopathic deficits really are specific to fear and sadness.
These mixed findings raise the possibility that deficits in facial expression recognition associated with psychopathic traits may be more pervasive than previously believed. A meta-analysis of facial expression recognition by Wilson et al. (Reference Wilson, Juodis and Porter2011) discovered that psychopathy was significantly associated with deficits in multiple emotional domains, although the effects were numerically small (N-weighted r = 0.06–0.12) and trended toward being largest for fear and sadness. These findings suggest that emotion recognition impairments in psychopathy are more pervasive than the more specific deficits outlined in Marsh and Blair’s (Reference Marsh and Blair2008) model.
13.3.3 Physiological Evidence of Emotion Deficits in Psychopathic Individuals
Psychophysiological research has convincingly supported the hypothesis that psychopathy involves a fundamental deficit in emotion processing. Studies in this area relate primarily to automatic affective responses consisting of a peripheral physiological response mediated by the autonomic nervous system such as the skin conductance response, eye blink reflex, and the approach-avoidance response (Levenson, Reference Levenson2014; Patrick, Reference Patrick2018). Early studies examined the change of participants’ skin conductance level to stimuli associated with the distress of others. In one paradigm, participants observed confederates who they thought were receiving electric shocks while their own skin conductance responses were recorded. When witnessing the distress of these confederates, offenders with a high level of psychopathy were found to show weaker autonomic arousal (i.e., lower skin conductance responses) than those without psychopathic traits (Aniskiewicz, Reference Aniskiewicz1979; House & Milligan, Reference House and Milligan1976).
Since then, numerous other studies have been conducted using the startle eye-blink response to test the hypothesis that psychopathic individuals are deficient in fear reactivity. The affective startle paradigm involves recording startle eye-blink responses to incidental noise bursts presented with emotional or neutral foreground stimuli. For the normal population, the startle blink response is an unconscious defensive reaction to aversive stimuli (Lang et al., Reference Lang, Bradley and Cuthbert1990), which is more potentiated during exposure to unpleasant images (e.g., violence, mutilation, murder) compared to neutral images (Vaidyanathan et al., Reference Vaidyanathan, Patrick and Bernat2009). A considerable body of research using this paradigm provides consistent and compelling evidence that, compared to normal individuals, psychopathic individuals are less sensitive (i.e., exhibit reduced or absent startle eye-blink responses) to aversive cues, indicating blunted emotions (e.g., Patrick, Reference Patrick2018) and heightened thresholds for reactivity to an aversive affective state (Levenston et al., Reference Levenston, Patrick, Bradley and Lang2000).
As one example, Patrick et al. (Reference Patrick, Bradley and Lang1993) reported that high-psychopathic male offenders have a heightened threshold for the startle blink during aversive picture viewing, indicating that it takes a stronger or more imminent threat to activate their defensive motivational system. Similar results were reported by Anderson et al. (Reference Anderson, Stanford, Wan and Young2011) in a sample of 76 undergraduate women assessed with high levels of psychopathic traits. The authors measured both affective startle blink responses and the P3 wave (i.e., an event-related potential component elicited in the process of decision making) using an affective picture-viewing task. Results showed that those scoring high on psychopathic traits lacked startle blink potentiation and demonstrated larger P3 amplitudes when faced with aversive stimuli. The P3 wave, is associated with attention and cognitive processing, with larger amplitudes indicating more cognitive resources devoted to the stimulus. In aversive contexts, larger P3 amplitudes suggest heightened arousal or emotional response, indicating greater engagement. For those with high psychopathic traits, larger P3 amplitudes in response to aversive stimuli suggest atypical processing, reflecting enhanced focus or heightened arousal that differs from typical responses. The study supports the generalizability of deficient startle potentiation of nonincarcerated females with psychopathic traits and adds to a growing body of research suggesting that psychopathic traits are associated with distinctive information-processing characteristics, as indexed by P3 amplitude.
In addition to the above paradigms, the approach-avoidance task (AAT) is another physiological reaction task used in psychopathy studies. In a study conducted by von Borries et al. (Reference von Borries, Volman, de Bruijn, Bulten, Verkes and Roelofs2012), the AAT was used to assess avoidance reactions to stimuli of potential threat among psychopathic individuals. During the task, stimulus photographs display facial expressions of actors with either angry, happy, or neutral expressions. Participants push a joystick either away or toward themselves as the faces increase or decrease in size. In this task, healthy individuals show a general tendency to move away from angry expressions and approach happy faces as in previous findings (e.g., Volman et al., Reference Volman, Toni, Verhagen and Roelofs2011). Psychopathic individuals, in contrast, show a total absence of avoidance tendencies toward angry faces (von Borries et al., Reference von Borries, Volman, de Bruijn, Bulten, Verkes and Roelofs2012), suggesting a lack of defensive responding consistent with findings from the startle blink paradigm. Moreover, their study found that avoidance rates were negatively correlated with aggression rates, suggesting that psychopathic individuals who show the lowest avoidance tendencies have the highest level of aggression. In summary, an absence of emotional reactivity seems to be a central deficit in psychopathy, leading to increased antisocial behavior due to diminished aversive arousal from punishment.
To conclude, intact emotional processes are theorized to be essential to moral behavior, providing immediate and salient feedback on behavior. Psychopathic individuals have difficulty responding empathetically, showing impaired processing of distress in others. Hence, we suggest that, while individuals with psychopathy understand in a rational sense what is right or wrong, they fail to experience this cognitive knowledge at an affective level because they lack the necessary emotional reactions to witnessing the harm caused to others. Nonetheless, how such deficits account for their deficient or absent moral sense still requires further investigation. We next take this physiological evidence further by addressing the neural foundation of the morality impairments found in antisocial and psychopathic individuals.
13.4 The Neuromoral Theory of Antisocial, Violent, and Psychopathic Behavior
Assuming that antisocial individuals exhibit some form of moral impairment, what accounts for such a deficit? In this section we argue that there are fundamental differences in the brains of psychopathic, violent, and antisocial individuals that can account for their moral aberrations. We first consider structural and functional brain deficits that have been observed in these antisocial populations. We then consider the neural foundations of moral decision making in normal populations. Finally, we review these two very different literatures in order to draw some broad conclusions on reasons why criminal offenders behave as they do. Our key proposition is that the brain impairments found in offenders correspond to those parts of the brain that govern moral decision making and that, lacking that neural moral compass, some individuals can step into a zone of moral transgressions (Crockett et al., Reference Crockett, Siegel, Kurth-Nelson, Dayan and Dolan2017).
Regarding brain impairments in offenders, the past 25 years have established beyond reasonable doubt that antisocial individuals have brains that are structurally and functionally different from the rest of us (Raine, Reference Raine2013). Early research documented that murderers compared to normal controls have reduced functioning in the prefrontal cortex, as well as impairments to other brain regions that include the angular gyrus and the amygdala (Raine et al., Reference Raine, Buchsbaum and LaCasse1997; Raine et al., Reference Raine, Buchsbaum, Stanley, Lottenberg, Abel and Stoddard1994). More recent research has confirmed and extended this proposition, documenting also that murderers have reduced gray matter in the prefrontal cortex, as well as in other regions that include the anterior cingulate and insula (Sajous-Turner et al., Reference Sajous-Turner, Anderson, Widdows, Nyalakanti, Harenski, Harenski, Koenigs, Decety and Kiehl2020). Such brain impairments are not specific to extreme groups, however. Structural and functional brain differences have been documented in a wide range of antisocial populations (those with conduct disorder or antisocial personality disorder, violent and nonviolent criminals, psychopathic individuals, life-course persistent offenders) from adolescence to adulthood (Carlisi et al., Reference Carlisi, Moffitt, Knodt, Harrington, Ireland, Melzer, Poulton, Ramrakha, Caspi, Hariri and Viding2020; Raine & Yang, Reference Raine and Yang2006), with prefrontal deficits being the most replicated finding. Additionally, impairments to the temporal lobe have also been highlighted in both past and current research on antisocial populations (Carlisi et al., Reference Carlisi, Moffitt, Knodt, Harrington, Ireland, Melzer, Poulton, Ramrakha, Caspi, Hariri and Viding2020; Raine, Reference Raine1993; Yang et al., Reference Yang, Raine, Colletti, Toga and Narr2009; Yang et al., Reference Yang, Wang, Baker, Narr, Joshi, Hafzalla, Raine and Thompson2015). Despite structural and functional brain differences, it is important to acknowledge the difficulty in demonstrating whether aberrant cognitive task performance by individuals with psychopathy represents a lack of capacity of understanding what is right and wrong or an actual difference in values.
An entirely different line of enquiry concerns the neural basis of moral decision making in normal individuals. At the turn of the century, a pioneering brain imaging study of moral decision making employing both personal and impersonal moral dilemmas at the turn of the century documented greater activation of the prefrontal cortex, posterior cingulate, and angular gyrus in response to personal moral dilemmas compared to impersonal dilemmas (Greene et al., Reference Greene, Sommerville, Nystrom, Darley and Cohen2001). Since then, a considerable body of knowledge has been built on the neural network subserving moral decision making in healthy individuals. One meta-analysis of studies employing either moral decisions (deciding whether to accept a proposed solution) or moral evaluations (judging the appropriateness of another’s actions) has highlighted frontal and temporal regions in addition to the cingulate cortex (Garrigan et al., Reference Garrigan, Adlam and Langdon2016). A second ALE (activation likelihood estimation) meta-analysis in the same year documented common neural denominators to both right–wrong moral judgments and reasoning about moral dilemmas in the medial frontal gyrus, the middle frontal gyrus, and the left middle and superior temporal gyri (Bryant et al., Reference Bryant, Wang, Deardeuff, Zoccoli and Nam2016). A third meta-analysis highlighted the ventromedial prefrontal cortex, orbitofrontal cortex, cingulate cortex, temporoparietal junction including the angular gyrus, amygdala, and the temporal pole being activated during moral-related tasks compared to control conditions (Han, Reference Han2017). As such, while different reviews draw somewhat different conclusions, overall there is a consensus that the prefrontal cortex (particularly ventral, polar, and medial prefrontal sectors), posterior cingulate, temporal cortex, and angular gyrus are particularly involved in moral decision making in healthy controls.
A broad question concerns whether or not these two disparate fields of enquiry – one on the neural basis of antisocial behavior and the other on the neural basis of moral judgments – are related in any way. This issue was addressed more than 15 years ago in a review that for the first time generated the hypothesis that there exists a common neural denominator to both antisocial behavior and moral decision making (Raine & Yang, Reference Raine and Yang2006). With respect to antisocial behavior, it was argued that key areas found to be functionally or structurally impaired in antisocial populations included dorsal and ventral regions of the prefrontal cortex, the amygdala, hippocampus, angular gyrus, posterior cingulate, and subregions of the temporal cortex including anterior and superior gyri. For moral decision making, it was suggested that extant research implicated a neural circuit consisting of polar, medial, and ventral regions of the prefrontal cortex, the superior temporal sulcus, and the angular gyrus, with initial findings additionally implicating the temporal pole and the amygdala. It was concluded that there exists a substantial overlap between the brain mechanisms implicated in antisocial/psychopathic behavior on the one hand and those involved in moral decision making on the other (Raine & Yang, Reference Raine and Yang2006).
Recently this neuromoral theory of antisocial behavior has been updated to include more recent findings in these two neuroimaging fields (Raine, Reference Raine2019). This revised model is illustrated in Figure 13.1. Key revisions in this update consist of the insula and anterior cingulate being added as areas common to both antisociality and morality. The striatum (caudate, putamen, globus pallidus, nucleus accumbens – only the caudate and putamen are illustrated) were added as areas specific to antisocial behavior, although, as we have noted, there is increasing support for its role in both antisocial and moral processing. The angular gyrus remains a common area but should be taken more widely to represent the associated temporoparietal junction.

Figure 13.1 The neuromoral model of antisocial behavior. Brain regions impaired only in antisocial groups include the striatum, hippocampus, temporal lobe, and dorsolateral prefrontal cortex. Regions activated only in moral decision making include the posterior cingulate. Regions common to both antisocial behavior and moral decision making include the anterior cingulate, fronto-polar/medial prefrontal cortex, ventral prefrontal cortex, amygdala, insula, superior temporal gyrus, medial prefrontal cortex, and angular gyrus.
The original neuromoral theory (Raine & Yang, Reference Raine and Yang2006) and its update (Raine, Reference Raine2019) are essentially based on comparing and contrasting different patterns of brain imaging findings from two different research fields, essentially a correlational approach. However, recent neurological research places this model on a more causal footing. Researchers identified 17 cases in which lesions to the brain were followed by criminal behavior (Darby et al., Reference Darby, Burke and Fox2018). Lesions in these 17 cases were scattered throughout the brain – there was not one single focus, although in more than half the cases lesions occurred in the frontal cortex. Using lesion network mapping, it was established that all of these locations were connected to one neural network. It was then documented that this specific network overlapped considerably with the functional network underlying moral decision making, consisting of the ventromedial prefrontal cortex, the inferior frontal gyrus, and the temporal cortex. From different scientific perspectives, therefore, there is converging evidence that damage to brain areas that result in criminal offending have their causal effect by impairing the neural basis of moral decision making.
The neuromoral theory of antisocial behaviors constitutes a functional neuroanatomical model positing that a foundational cause of disparate antisocial behaviors resides in dysfunction to a network of brain areas that constitute the infrastructure to moral behavior. As a causal model, its core proposition is that dysfunction to one or more areas of the neuromoral circuit results in impairment to feeling, thinking, and behaving in a moral way, which in turn lays the foundation for antisocial, violent, and psychopathic behavior. There are, nevertheless, unresolved issues. For example, which moral component is more impaired in offenders – the cognitive or the emotional? The neuromoral theory has argued that it is the emotional feeling of what is morally wrong that constitutes the primary deficit, with cognitive components of morality being secondary (Raine & Yang, Reference Raine and Yang2006). In contrast, research on brain lesion cases suggests that emotional components such as empathy and cognitive control are not involved, whereas cognitive components of morality such as theory of mind and reward-based decision making are involved (Darby et al., Reference Darby, Burke and Fox2018). The distinction between cognition and emotion may be a false dichotomy, as both could be implicated in moral decision making. These two competing perspectives on the primacy of affective versus cognitive attributes of morality in relation to the neuromoral theory requires further resolution.
13.5 Interventions for Enhancing Moral Sensitivity and Reducing Antisocial Behavior
Despite many years of study, the efficacy of traditional psychosocial treatments for psychopathy such as cognitive behavioral therapy (Carroll & Kiluk, Reference Carroll and Kiluk2017), psychodynamic interventions (Juni, Reference Juni2010), behavioral modification (Pearson et al., Reference Pearson, Heilbronner, Barack, Hayden and Platt2011), and the therapeutic community approach (Hecht et al., Reference Hecht, Latzman, Lilienfeld, David, Lynn and Montgomery2018) is still controversial. Reviews tend to yield somewhat inconsistent conclusions, ranging from optimistic (Salekin, Reference Salekin2002; Salekin et al., Reference Salekin, Worley and Grimes2010) to pessimistic (Harris & Rice, Reference Harris, Rice and Patrick2006). Moreover, the main objective of most of these interventions is reducing recidivism, while changes in moral cognition and affect are usually neglected (Caldwell et al., Reference Caldwell, Skeem, Salekin and Van Rybroek2006; Salekin et al., Reference Salekin, Worley and Grimes2010; Skeem et al., Reference Skeem, Polaschek, Patrick and Lilienfeld2011).
Recent studies have attempted to address this research gap by studying treatment efficacy using empathic improvement as an outcome measure (Romero-Martínez et al., Reference Romero-Martínez, Lila, Gracia and Moya-Albiol2019; Romero-Martínez et al., Reference Romero-Martínez, Lila, Martínez, Pedrón-Rico and Moya-Albiol2016). In one study, Romero-Martinez and colleagues (2019) examined whether violent perpetrators experienced changes in emotion-decoding abilities and showed improvement in empathic responses after completing a standardized court-ordered intervention program incorporating cognitive restructuring, emotion management skills, and problem-solving training with motivational strategies. Results revealed that perpetrators receiving this program were more accurate at decoding emotional facial signals and presented with better cognitive empathy (perspective taking) than those who did not receive this intervention. Nevertheless, the authors were unable to identify which part of the intervention accounted for improvement in these abilities.
In contrast to psychosocial interventions, a completely different approach has been explored recently using brain stimulation techniques with the goal of enhancing moral decision making in order to reduce criminal behavior. In a randomized controlled trial, Choy et al. (Reference Choy, Raine and Hamilton2018) randomized human adults to receive transcranial direct current stimulation to the prefrontal cortex and others to receive a sham condition where participants believed they were getting stimulation but ultimately did not. Following the stimulation, participants were given vignettes which placed them in social situations in which they were provoked and given the opportunity to commit a violent act. Stimulation to the prefrontal cortex reduced the intention to commit a violent act by 47.8 percent relative to the sham condition. In addition, prefrontal stimulation increased the participants’ sense of the moral wrongfulness of perpetrating a retaliatory violent act. An increased sense of moral wrongfulness was, in turn, associated with a reduced likelihood of acting violently. Mediation analysis documented that the moral enhancement (i.e., perceptions of greater moral wrongfulness) produced by prefrontal stimulation accounted for 31 percent of the treatment effect. We caution that not all studies have shown positive treatment effects using transcranial direct current stimulation (Ling et al., Reference Ling, Raine, Choy and Hamilton2020), but the study does further our understanding of causation, and these recent results provide a new vista on how moral behavior could potentially be enhanced in a way to reduce violence.
13.6 Future Directions and Summary
This review has aimed to clarify a mixed literature on the relationship between antisociality and morality and to provide insight into the roles of cognition and emotion in the moral behavior of individuals on the antisocial spectrum. One broad conclusion that we can draw is that psychopathic individuals are probably more capable of forming reasoned moral judgments than has been traditionally assumed. A second conclusion is that, at a physiological and neurocognitive level, psychopathic individuals have difficulty responding empathically and show an impairment in the processing of distress in others, affective deficits that may predispose them to immoral, uncaring behavior. A third conclusion is that the immoral behavior of antisocial individuals is in part predicated on disruption to neural processes that subserve moral decision making.
Notwithstanding these broad conclusions, there are several issues that are unresolved and require further investigation. First, we have suggested that psychopathic individuals have emotional impairments that interfere with their ability to act in a morally appropriate manner, specifically highlighting deficits in empathizing with the negative emotional states of others. Recent evidence however suggests that psychopathic individuals are significantly less impaired in empathizing with others’ positive emotional states (Raine et al., Reference Raine, Chen and Waller2022). A future avenue for treatment could capitalize on this islet of positive emotional processing in psychopathic individuals to enhance their moral sense.
Second, it is important to note that antisociality and aggressivity are broad and multifaceted constructs, and there are different subgroups within this population that may have different moral capacities. For example, while children who are proactively aggressive (using aggression to achieve desired goals) are less differentiated in moral and conventional concepts, children who are reactively aggressive (aggressive in response to provocation) have greater differentiation in these concepts (Jambon & Smetana, Reference Jambon and Smetana2018). Future research could explore whether this provocative finding could apply to adult reactive and proactive aggression.
Relatedly, while it is argued that there is substantial evidence for the involvement of dysfunctional neural structure in producing both impaired moral decision making and antisocial behavior, the heterogeneity of antisocial behavior suggests that this perspective may not apply to all criminal offenders. For example, while the moral cognitions, emotions, and behavior of “primary” psychopathic individuals, proactive aggressive individuals, and life-course persistent offenders may in large part be explained by brain impairment to the neural moral circuit, this may be less true of more “emotionally charged” reactively aggressive individuals, “secondary” psychopathic individuals, and drug offenders where other causal factors may better account for their antisocial and aggressive behavior (Raine, Reference Raine2019).
Furthermore, there is increasing evidence that different antisocial subgroups within this antisocial spectrum have distinct cognitive-affective deficits (Baskin-Sommers et al., Reference Baskin-Sommers, Curtin and Newman2015). For instance, in the area of emotion regulation, individuals within the antisocial spectrum often display significantly different patterns of stress tolerance. Specifically, individuals with psychopathy with a callous, fearless, and irresponsible disposition have been found to have an emotionally “cold” style and to demonstrate superior distress tolerance, while antisocial individuals with externalizing behavior are usually associated with an emotionally “hot” style and demonstrate poor distress tolerance (Baskin-Sommers & Newman, Reference Baskin-Sommers, Newman, Robinson, Watkins and Harmon-Jones2013; Sargeant et al., Reference Sargeant, Daughters, Curtin, Schuster and Lejuez2011). Similarly, in the area of cognitive control, individuals with psychopathy demonstrate better cognitive control and working memory than an antisocial subgroup with externalizing traits (Endres et al., Reference Endres, Rickert, Bogg, Lucas and Finn2011). Baskin-Sommers et al. (Reference Baskin-Sommers, Curtin and Newman2015) demonstrated that training designed to treat the distinct deficits of different antisocial subtypes resulted in differential improvements on both behavioral and psychophysiological measures. Given the heterogeneity of antisocial subtypes, future research needs to identify better methods of characterizing subgroups with relatively distinct and homogeneous cognitive-affective traits, especially in the area of moral capacity, to develop more effective mechanism-based interventions for antisocial subpopulations. In addition, a recent review suggested that there are four classes of moral judgment: blame judgments, wrongness judgments, norm judgments, and evaluations (Malle, Reference Malle2021). Not only is the concept of psychopathy heterogeneous, but moral judgments are also varied. Therefore, successful examination of the moral capacity of psychopaths will require new empirical research studies that take these differences into account.
In closing, it is clear that research into the moral values of antisocial and psychopathic populations has a long and venerable past. What must be recognized, however, is that this body of knowledge has struggled to develop any practical utility in enhancing the moral sense in antisocial individuals to reduce their offensive behavior. A major future challenge lies in translating our theoretical knowledge of deficits in moral judgment making into practice.
It goes without saying that parties in intergroup conflict tend to hold negative views of one another: jaundiced at best, vilifying at worst. It has been said that we always hurt the ones we love, but those we hurt often seem to be the ones we hate. In theory, conflicts could arise between groups that are mutually appreciative and respectful but that just happen to have different interests. In fact, of course, lasting conflicts tend to be marked by shared beliefs that our opponents are wicked and untrustworthy.
This commonsense view that mutual derogation is fundamental to conflict is backed up by generations of social psychological theory and research. Studies of intergroup relations present as axiomatic the view that people evaluate their own group (in-group) more positively than other groups (out-groups), and even if the discrepancy primarily represents in-group love rather than out-group hate (Brewer, Reference Brewer1999), negative perceptions of out-groups that picture them as morally suspect and us as morally elevated rapidly emerge from it when conflict arises. Studies of stereotyping commonly draw the same conclusion: We have the virtues of warmth and competence, but they have a collection of vices. Studies of prejudice examine how negative attitudes toward out-groups drive antagonistic behavior and who tends to hold these attitudes more than others.
If the idea that intergroup conflict is grounded in negative group perceptions is common sense, it has become close to common sense in some circles that the most atrocious conflicts also tend to involve dehumanizing group perceptions. There is widespread historical awareness of genocides and wars in which human groups have been likened to subhuman animals or have had their personhood denied in egregious ways. These examples seem to suggest that in intergroup conflicts, groups of others are not only perceived as more negative but also as less human. Although this view has entered the zeitgeist as an almost self-evident fact, it has become a good deal more controversial than many people suspect.
Just as social psychology has clarified the cognitive, emotional, and motivational processes that underpin negative group perceptions within intergroup conflict, it has also examined the processes and dynamics that underpin dehumanizing perceptions in these contexts. Especially in the past two decades, psychologists have developed theoretical accounts of dehumanization and methodological tools that allow it to be studied in accordance with the field’s scientific norms. Hundreds of empirical studies have documented dehumanizing perceptions of an assortment of groups, from Roma in Hungary to cyclists in Australia to American Republicans and Democrats, and explored the correlates, consequences, and cures of these perceptions. Dehumanization has come to be seen as another way in which out-groups and enemies can be viewed, perhaps secondary to out-group derogation and prejudice but important in its own right.
In this chapter I will present an overview of the burgeoning literature on dehumanization with the aim of clarifying its roles in intergroup conflict. The chapter will present an overview of the main psychological accounts of dehumanization, an analysis of the many forms that dehumanization has been theorized to take, and a discussion of its possible functions in intergroup conflict in general and genocide in particular. It will conclude with some remarks on recent challenges to the role of dehumanization in intergroup conflict and some speculations on its boundary conditions. The aim of the chapter is to provide a panoramic view of the field that will allow moral psychology researchers to orient themselves within it and plan new lines of investigation.
Two prefatory points must be made before proceeding further. First, although most dehumanization scholarship has addressed its role in intergroup contexts, it is important to note that some more recent work has examined dehumanization processes in perceptions of self and in interpersonal relationships. This work explores how people may come to view themselves as losing or lacking humanness and how people in close relationships may – fleetingly or lastingly – perceive their loved ones in ways that fail to humanize them. This work demonstrates that dehumanization may not belong exclusively in the intergroup domain. Readers interested in an example of this genre or a review of dehumanization research in a more expansive frame might consult Pizzirani et al. (Reference Pizzirani, Karantzas, Roisman and Simpson2021) and Haslam and Loughnan (Reference Haslam and Loughnan2014), respectively.
Second, this chapter will focus primarily on social psychological work on dehumanization. It does so in full awareness that there is a lively and important body of philosophical research on the topic (e.g., Kronfeldner, Reference Kronfeldner and Kronfeldner2021a). The reason for attending primarily to the psychological literature, with the exception of an analysis of the work of David Livingstone Smith, is twofold. For one thing, social psychology is easily the most substantial disciplinary contributor to the reemergence of dehumanization scholarship over the past three decades (Haslam, Reference Haslam and Kronfeldner2021). For another, philosophical work on the subject has mainly addressed fundamental conceptual issues in the field, generally with a critical stance toward psychological formulations and without contributing to or motivating empirical work. As a result, it has largely developed in parallel to the much larger corpus of psychological work, rather than exerting a strong influence on it. Readers wishing for a more multi-disciplinary overview of dehumanization studies, with philosophical voices well represented, should consult Kronfeldner (Reference Kronfeldner2021b).
14.1 Theoretical Accounts of Dehumanization
Before we explore how social psychologists have fashioned their theoretical accounts of dehumanization, it is instructive to consider how “dehumanization” tends to be employed in everyday discourse. First, it tends to be seen as an extreme phenomenon, most often invoked in the contexts of the Holocaust, genocides, or armed conflicts. Second, in these extreme settings dehumanization is usually understood to promote and enable violence and atrocity. Third, it is typically exemplified by the blatant use of offensive animal metaphors, such as references to Jews as rodents in Nazi propaganda, to Tutsis as cockroaches during the Rwandan conflict, or to people of the African diaspora as apes on racist websites or in European soccer stadiums. Commonly it is assumed that people using these demeaning animal metaphors believe that the people slurred in this way are literally less than human.
Recent psychological accounts of dehumanization depart significantly from every one of these typical understandings. Most of them present dehumanization as a phenomenon that falls on a spectrum from the extreme to the everyday. Much of the recent body of dehumanization research examines banal forms of dehumanization that can occur in the absence of intergroup conflict and that constitute entirely normal processes of group perception. Although it is usually assumed and sometimes corroborated that dehumanization fosters aggression, most accounts do not present dehumanization primarily as an enabler of violence. In addition, contemporary psychological theories of dehumanization do not conceptualize it as the explicit use of animal metaphors. Instead, dehumanization is understood and measured as a subtler phenomenon that may not be symbolized verbally or visually and may have nothing to do with nonhuman animals even indirectly. Rather than amounting to an ontological claim that members of a group are nonhuman or less than human, psychological accounts typically claim it when the group is ascribed human attributes to a lesser degree than others. It is essential to remember that dehumanization has a substantially broader meaning in psychological accounts than it does in the prototypical cases from Nazi rhetoric and other extreme examples that define the common understanding among laypeople.
14.1.1 Moral Disengagement and Moral Exclusion
The first psychological writers on dehumanization began to publish on the subject in the 1970s. Their work made important contributions to the theory of dehumanization but relatively modest contributions to its empirical base. Perhaps as a result of this emphasis, the topic did not catch on in social psychology with the vigor that characterized the second wave of dehumanization scholarship, which arose around the millennium. Nevertheless, their work often foregrounded the moral dimensions of dehumanization in a way that later work did not. Among the key contributors of this period are Kelman (Reference Kelman, Kren and Rappoport1976) and Staub (Reference Staub1989), who wrote conceptually nuanced accounts of the role of dehumanization in war and in mass killing. However, the work of Bandura, Opotow, and Bar-Tal, and their colleagues, is of particular relevance to a chapter on moral psychology.
Bandura, one of the most influential psychologists of the twentieth century, is noteworthy for bringing dehumanization into the laboratory, for developing a theoretical formulation that emphasized how dehumanization overcomes inhibitions against aggressive behavior, and for examining individual differences in the tendency to dehumanize others. According to Bandura’s model, people are ordinarily restrained from behaving aggressively toward others by moral self-sanctions. When the potential targets of aggression are perceived as less than human, these self-sanctions are disengaged and those inhibitions are weakened. This “moral disengagement” account therefore presents an intrapsychic mechanism for enabling aggression, viewing it essentially as releasing the brakes of moral compunction and guilt. Bandura and colleagues carried out a series of studies that demonstrated greater punitiveness toward people described in a dehumanizing manner (e.g., Bandura et al., Reference Bandura, Underwood and Fromson1975; see Bandura, Reference Bandura1999, for an overview).
Whereas Bandura formulated dehumanization’s role in aggression in terms of loosened inhibitions on hostile behavior, Opotow’s somewhat related work on “moral exclusion” offered a more cognitive account that was not tied specifically to aggressive behavior. By her account (Opotow, Reference Opotow1990), individuals or groups can be shifted outside the moral sphere, where normal considerations of justice, fairness, and compassion apply. This active process of excluding persons from the moral domain effectively disqualifies them from moral standing. The result of that exclusion might be passive neglect or forms of active harming that extend beyond aggression in the narrow sense. Opotow did not privilege dehumanization as a phenomenon in immoral action or inaction but presented it as merely one among many forms of moral exclusion.
In a similar vein, Bar-Tal (Reference Bar-Tal, Bar-Tal, Grauman, Kruglanski and Stroebe1989) presented dehumanization as one of several forms of “delegitimization,” by which people withdraw legitimacy from groups they oppose, rendering aggressive behavior normatively acceptable. Legitimacy is commonly denied to opponents in violent conflicts, who are frequently denigrated as animals, demons, or monsters who are undeserving of concern and unfit for anything but the harshest punishments. Delegitimization and moral exclusion are therefore alternative ways of expressing the claim that people treat one another morally by default, but immoral treatment is enabled when their status as humans is denied or diminished.
14.1.2 Infrahumanization
Theories of moral disengagement, moral exclusion, and delegitimization understand dehumanization as one of a collection of related processes or mechanisms that serve to release moral restraints on aggression and other forms of ill-treatment. These theories typically understand dehumanization to be overt (e.g., revealed in language), its effects to be harmful, and its settings to be conflictual. The theories of moral exclusion and disengagement in particular also did not tie dehumanization primarily to intergroup contexts: One might exclude another from consideration on grounds other than their out-group status, and Bandura investigated whether delinquent schoolchildren’s aggressiveness might spring from greater tendencies to disengage morally. Research on “infrahumanization” (Leyens et al., Reference Leyens, Demoulin, Vaes, Gaunt and Paladino2007; Leyens et al., Reference Leyens, Rodriguez-Torres, Rodríguez-Pérez, Gaunt, Paladino, Vaes and Demoulin2001) differed from these earlier theories in all of these respects. The infrahumanization phenomenon was subtle and covert, was not necessarily associated with harmful behavior, occurred in the absence of major conflict, and was fundamentally an intergroup process.
Leyens and colleagues initiated the concept of infrahumanization not as part of a theory of dehumanization – they were at pains to differentiate the two concepts – but as a way to understand ethnocentrism, the tendency for ethnic groups to see their own cultural kind as the true humans. Pilot work suggested that humanness, conceptualized as contrastive with nonhuman animals, was understood by laypeople as constituted primarily by the capacities for language, intelligence, and uniquely human emotions (sentiments in French). Narrowing their focus to emotions, the researchers reasoned that if people see their in-group as more human than out-groups, they should reserve uniquely human emotions such as nostalgia and guilt for the in-group but not differentially attribute nonuniquely human emotions such as happiness and fear. This basic finding of differential attribution of uniquely human emotions, independent of emotion valence (positive versus negative), was replicated in many intergroup comparisons, including comparisons between groups with little or no contemporary conflict (e.g., Peninsular and Canary Island Spaniards). The ease with which infrahumanization could be studied in laboratory and field settings, using familiar survey and implicit social cognition tasks, ensured that it generated a substantial research literature.
Leyens and colleagues presented infrahumanization as a subtle but fundamental phenomenon that had more to do with how people view their in-group than how they view out-groups. Whereas dehumanization might normally be understood as a denial of humanity to others, generally accompanied by a negative view of them, infrahumanization was conceptualized as reserving humanity for the in-group, with the out-group thereby perceived as less human by this standard of reference. Infrahumanization does not entail dehumanization in a strong sense, but it might plausibly prepare the ground for it when conflict is stoked. If the out-group is already judged as less human than the in-group – further down the continuum from humans to animals than the in-group – then exacerbation of tensions might tip implicit deficiency in humanness into explicit denial. Leyens and colleagues showed that infrahumanization was not without consequences – it reduced out-group helping, for example, and deepened intergroup divides – but they gave less attention than earlier writers to the moral psychological dimensions of the phenomenon. Nevertheless, the fact that uniquely human emotions tend to be seen as more diagnostic of people’s “moral nature” (Demoulin et al., Reference Demoulin, Leyens, Paladino, Rodriguez‐Torres, Rodriguez‐Perez and Dovidio2004) than other emotions implies that infrahumanization is not a morally neutral bias in group perception.
14.1.3 The Dual Model
Infrahumanization theory offers a novel account of intergroup dehumanization that presents it as subtle, everyday, and falling on an implied continuum with more severe variants. Covertly perceiving out-groups as deficient in some of the emotional capacities that distinguish humans from nonhuman animals is a few incremental steps removed from overtly judging them to be animals. The theory also departs from prior accounts by understanding dehumanization processes in terms of human attributes rather than categories. Instead of understanding “human” as a sphere from which people could be categorically removed, as in moral exclusion, or “nonhuman” as a derogatory label such as “monster,” as in delegitimization research, Leyens and colleagues characterized infrahumanization as a lack of specified human characteristics relative to an in-group. Humanness was formulated as a dimension of psychological content along which groups can be located, such that dehumanization is a matter of people being judged less human or lesser humans, rather than as categorically nonhuman.
Haslam’s (Reference Haslam2006) dual model represents an extension of the approach pioneered by Leyens and colleagues. It retained their crucial idea that a satisfactory theory of dehumanization needs a theory of humanness, the quantity that is denied to people when they are dehumanized. It also retained their basic assumption that dehumanization can be subtle and should be conceptualized as a matter of degree. The dual model differed from infrahumanization theory in two key respects. First, it proposed that humanness could not be adequately defined only in contrast to nonhuman animals. Second, it expressly argued that dehumanization occurs on a continuum of severity or blatancy and that the contrast categories for defining humanness are implicated in that continuum.
On the first point, Haslam (Reference Haslam2006) argued that although humanness can be defined vis-à-vis animals as a set of uniquely human attributes, it can also be defined in a second salient way vis-à-vis inanimate objects. What makes us human is not only the uniquely human qualities of intellect, rationality, civility, self-control and the like that distinguish us from beasts but also the qualities of emotionality, warmth, and mental flexibility that people believe distinguish us from robots or automatons. Haslam referred to these latter qualities collectively as “human nature” and demonstrated empirically that the two senses of humanness were independent, were negatively associated as theorized with animals and objects, respectively, and had different correlates (e.g., human nature attributes were believed to emerge early in development and be widely prevalent in the population and were essentialized, whereas uniquely human attributes were believed to emerge later in development and to be uncommon, and were not essentialized). Haslam presented the two senses of humanness as encompassing dimensions that could be examined empirically by examining a range of psychological attributes, including emotions and personality traits.
On the second point, Haslam (Reference Haslam2006) argued that a theory of dehumanization must not only refer to these distinct senses of humanness but also remember that they are anchored by different kinds of nonhuman entity. Milder forms of dehumanization are likely to involve subtle denials of humanness to out-groups, but harsher forms will tend to involve explicit likening of those groups to the corresponding nonhuman. The standard infrahumanization effect is a subtle judgment that the out-group has a deficient capacity for experiencing refined, uniquely human emotions, but it lies at the mild end of a spectrum of animalistic dehumanization that extends to the use of vilifying animal metaphors at the other end.
Haslam’s dual model aimed to provide an integrative framework for dehumanization research in psychology by describing two such spectra. “Animalistic” dehumanization captures forms of dehumanization in which people are denied uniquely human attributes and/or likened to nonhuman animals and tends to occur when groups are seen as primitive, bestial, crude, uncivilized, barbaric, or stupid. “Mechanistic” dehumanization occurs when groups are denied human nature attributes and/or likened to inanimate objects. It tends to occur when people are judged to be cold, unfeeling, fungible, robotic, shallow, or literally objectified. An assortment of studies has found support for elements of this two-spectrum model, including the content and cross-cultural consistency of the humanness dimensions (Bain et al., Reference Bain, Vaes, Haslam, Kashima and Guan2012), and applied it to understanding dehumanizing perceptions of a range of groups (e.g., Loughnan & Haslam, Reference Loughnan and Haslam2007).
14.1.4 Dehumanization and Mind Perception
The dual model conceptualizes dehumanization as the denial of humanness to people along two dimensions. Although these dimensions would appear to have moral significance and should, in principle, capture the moral standing of humans relative to nonhumans, most research has assessed them using personality traits. Traits, and the emotions studied in infrahumanization research, are clearly germane to moral psychologists (e.g., Bastian et al., Reference Bastian, Laham, Wilson, Haslam and Koval2011). However, the moral psychological implications of dehumanization became clearer in another recent tradition of research on dehumanization that grounded it in mind perception. In this work (e.g., Gray et al., Reference Gray, Gray and Wegner2007; Waytz et al., Reference Waytz, Gray, Epley and Wegner2010), the account of dehumanization largely falls out of a primary focus on understanding the nature, causes, and consequences of perceiving mind in other entities, human and nonhuman. Dehumanization is understood as dementalization.
A key insight of this work is that perceptions of mind may not be unitary and different entities may be ascribed different kinds of mind. Work by Gray and colleagues (Reference Gray, Gray and Wegner2007), for example, found evidence of two dimensions of mind, which they labelled Agency and Experience. Agency represents the degree to which entities were judged to have the capacity for thinking, planning, and self-control, with adult humans ranking high and nonhuman animals ranking low. Experience, in contrast, represents the degree to which entities are judged capable of feeling, motivation, and subjectivity, with humans and other mammals ranked high and inanimate objects, robots, dead persons, and persons in a persistent vegetative state ranking low. The mapping of these dimensions onto the dual model’s human/animal and human/object contrasts is striking, but Gray and colleagues (Reference Gray, Gray and Wegner2007) drew out their moral implications more powerfully by linking their mind dimensions to the dimensions of moral agency and patiency. The degree to which people are seen as moral agents, capable of responsibility and blameworthiness, is driven by their perceived capacity for Agency, and the degree to which they are seen as moral patients, capable of having moral and immoral acts perpetrated on them, tracks their perceived capacity for Experience. On this understanding, dehumanization involves mind denial, and the form of that denial may have distinct implications for the moral credentials of the target.
Understanding dehumanization in terms of mind perception has benefits beyond providing clarity to moral psychologists. Research on the cognitive and motivational influences on mind perception (Waytz et al., Reference Waytz, Gray, Epley and Wegner2010) has direct relevance for understanding the causes and effects of dehumanization, connects it to important related literatures on empathy and theory of mind, and also helps to make dehumanization processes tractable for cognitive neuroscientists. Harris and Fiske (Reference Harris and Fiske2006), for example, evaluated dehumanization of an assortment of social groups by examining the absence of activation of social cognition centers in the brain, in the process supporting their proposal that groups stereotyped as low in warmth and competence, such as drug addicts and homeless people, are especially likely to be cognized as objects rather than social beings. This work has inspired several later studies of the neuroscience of dehumanization (e.g., Bruneau et al., Reference Bruneau, Jacoby, Kteily and Saxe2018; Jack et al., Reference Jack, Dawson and Norr2013). Although the specific validity of Gray et al.’s (Reference Gray, Gray and Wegner2007) dimensions of mind perception has been called into question (Malle, Reference Malle, Goel, Seifert and Freksa2019; Weisman et al., Reference Weisman, Dweck and Markman2017), and alternative structures might therefore imply different accounts of dehumanization, the idea of grounding an account of plural forms of dehumanization in an account of plural components of perceived mind has been a highly generative one. (See Chapter 9 in this volume for further discussion of the relationship between mind perception and dehumanization.)
14.1.5 David Livingstone Smith’s Account
We have examined several prominent psychological accounts of dehumanization, but it is also essential to review the most influential philosophical account, which has been set out by David Livingstone Smith in a powerful series of books (Smith, Reference Smith2011, Reference Smith2020, Reference Smith2021). Smith’s work places a high premium on applying his theoretical account of dehumanization to historical examples of racism, genocide, and enslavement, and it aims to explain the apparently paradoxical conceptual aspects of the phenomenon rather than to drive forward an empirical research program. In these respects, it tends to take a critical perspective toward psychological accounts of dehumanization, which in their efforts to bring the concept into the laboratory and make it tractable for social psychology researchers have often lacked historical depth, theoretical ambition, and philosophical sophistication.
At the core of Smith’s account is the recognition of an ambiguity or contradiction at the heart of well-known examples of dehumanization. He argues that perceiving others as humans is automatic (Smith, Reference Smith2021, p. 359). Dehumanization occurs when perceivers are led to believe by people with epistemic authority – whether they be scientists, ideologues, or others, and regardless of whether that authority is legitimate – that this human appearance is deceptive and the others are in reality subhuman. As a result, Smith (Reference Smith2021, p. 359) writes:
[W]hen people dehumanize others, they are saddled with two contradictory mental representations of them. And because these are starkly contradictory, they cannot both be salient simultaneously. The mind of the dehumanizer foregrounds the humanity of the other and backgrounds their subhumanity at some moments, and foregrounds their subhumanity and backgrounds their humanity at others.
Smith uses the concept of “psychological essentialism” – laypeople’s tendencies to believe that certain categories are grounded in causal essences that define their members’ true nature – to explain this contradictory tendency for people to perceive others as simultaneously human and less than human. At the level of appearance, people are invariably perceived as human-like, but when they are dehumanized these “human-seeming beings” are ascribed a subhuman essence. Dehumanization therefore represents a belief that although someone might look human, they fundamentally are not: The target of dehumanization is both human and subhuman.
Smith takes issue with how some psychological writers employ the concept of psychological essentialism in their work. He argues, for example, that Leyens and colleagues’ work on infrahumanization, which formulates the phenomenon as a denial to out-groups of the “human essence” (represented by uniquely human emotions inter alia), is problematic because infrahumanization is conceptualized as a matter of degree whereas essences are either present or absent. To be infrahumanized should therefore constitute a categorical loss of humanness, not merely a partial or relative attenuation of humanness. Similarly, Smith criticizes Haslam’s account, which argues (and finds; Haslam et al., Reference Haslam, Bain, Douge, Lee and Bastian2005) that human nature attributes are essentialized and that mechanistic dehumanization therefore involves a denial of fundamentally human qualities to people. To Smith, phenotypical attributes such as traits are not the relevant locus of essentialist thinking in dehumanization and such attributes cannot qualify as causal essences regardless.
Ultimately, these critiques hinge on unresolved theoretical disagreements about the nature of psychological essentialism. At root, Smith’s account holds that essentialism operates at the level of the human category, that people construe essences in a categorical fashion, and that dehumanization is therefore a categorical perception that the target is nonhuman by virtue of having a subhuman essence. Leyens’ account also holds that essentialism operates at the level of the human category but conceptualizes the essence as a matter of degree: People are perceived to have the human essence to the extent they are ascribed uniquely human attributes. Haslam’s account places essentialism at the level of human attributes, not the category as a whole, and understands essentialist beliefs about human attributes to vary by degrees: The extent to which particular human qualities are seen as deep-seated, fixed, either/or, and defining falls on a continuum (Haslam et al., Reference Haslam, Bastian and Bissett2004).
The upshot of these disagreements is that Leyens and Haslam construe dehumanization as implicating essentialist thinking but occurring on a spectrum, a default assumption in quantitative social science if not in philosophy. By their accounts, dehumanization amounts to seeing others as less human, or as lesser humans, rather than as nonhuman. In contrast, Smith construes it as a categorical phenomenon: Either someone is dehumanized or they are not. This view raises challenging questions: If dehumanization is either/or, why does it appear to vary in severity or blatancy, and how can a person be dehumanized and yet still be perceived as human in some respects? Smith’s account resolves these questions by proposing that two simultaneous representations exist, one human (appearance) and one subhuman (essence). This account therefore has to posit dual representations of fluctuating relative salience to accommodate varying levels or severities of dehumanization, or it has to use “dehumanization” to refer only to the starkest examples and exclude subtler but otherwise very similar phenomena. In contrast, the psychological accounts can accommodate these variations more parsimoniously by invoking single representations that vary by degrees on a continuum from subtle to severe. Whether that parsimony is gained at the cost of a misunderstanding of psychological essentialism and an overly expansive definition of dehumanization remains open to debate.
14.2 The Diversity of Dehumanization: Clarifying Forms and Definitions
This extended review of psychological theories of dehumanization emphasizes the diversity of their descriptive and explanatory approaches. It sets aside the large body of research that applies these theories to such questions as which people are most likely to dehumanize others, which people are most likely to be dehumanized, which situational and psychological factors promote dehumanization, what are its effects, and how might it be reduced. This now very extensive literature is out of scope for the present chapter but is reviewed in Kteily and Landry (Reference Kteily and Landry2022) and Haslam and Loughnan (Reference Haslam and Loughnan2014). For present purposes it is more important to clarify the range of meanings that dehumanization has acquired in recent psychological work. Definitional issues are sometimes ignored, but the concept of dehumanization is intrinsically slippery, the term is used in disparate ways by different writers, and discrepant conceptualizations may contribute to misunderstandings between scholars who approach the topic from different disciplinary backgrounds. In this section of the chapter, I unpack several dimensions along which understandings of dehumanization vary within psychological theory and research.
14.2.1 Blatant versus Subtle
The most commonly discussed examples of dehumanization from history fall at one extreme on a dimension of blatancy or explicitness. When a human group is directly likened to a reviled animal and that likening carries an obvious implication for how the group should be treated, there can be no mistaking the severity of the dehumanization. As we have seen, however, some accounts of dehumanization refer to much subtler and milder phenomena. When a group is infrahumanized, there is no explicit likening to a nonhuman entity and both perceivers and targets may be unaware that targets are being perceived as less human humans. The phenomenon itself may simply be an unconscious and unmotivated tendency to not ascribe some emotions, both negative and positive, to the out-group. Other research on dehumanization deliberately examines unconscious or automatic processes by assessing implicit associations between groups and either humanness-related traits or nonhuman words, using the implicit association task or similar procedures (e.g., Loughnan & Haslam, Reference Loughnan and Haslam2007; Vaes et al., Reference Vaes, Paladino and Puvia2011). Research of this kind may reveal that people mentally associate out-groups with animals or do not associate them with uniquely human traits, but those associations are likely to be entirely unknown to the research participant and sincerely deniable by them. In cases such as these, dehumanization has occurred according to some existing accounts, even if they fall well short of prototypical cases from world history.
Although researchers have tended to emphasize the more subtle forms of dehumanization over the past two decades, blatant forms have also received some attention. The important work of Kteily, Bruneau and colleagues is especially relevant (e.g., Kteily & Bruneau, Reference Kteily and Bruneau2017; Kteily et al., Reference Kteily, Bruneau, Waytz and Cotterill2015). They have shown repeatedly that many people are willing to adjudge some groups as less than fully human when asked to rate them on the well-known “Ascent of man” scale, which depicts an evolutionary sequence from ape (0) to modern human (100). This blatant form of dehumanization, which appears to be strongest toward groups with which study participants are experiencing conflict, such as Muslim terrorists in the US and Roma in Hungary, is only moderately correlated with prejudice toward those groups, demonstrating that dehumanization and humanness are not reducible to hatred and negativity. Equally importantly, blatant dehumanization measured by the scale predicts a range of aggressive, punitive, and callous responses to its targets independently of prejudice. Blatant forms of dehumanization have also been examined in studies of animal metaphors (e.g., Haslam et al., Reference Haslam, Loughnan and Sun2011), which make the important point that not all animal metaphors are judged to be dehumanizing when applied to people, and that animals whose metaphorical use is judged to be offensive tend either to be seen as disgusting or seen as hierarchically beneath humans (e.g., apes; Brandt & Reyna, Reference Brandt and Reyna2011) in a way that implies a dehumanizing comparison.
14.2.2 Absolute versus Relative
Just as contemporary research on dehumanization encompasses subtle as well as blatant expressions, it also encompasses cases in which an out-group is seen as lacking humanness in itself and also when that lack is simply relative to the in-group. The prototypical cases of historical dehumanization are absolute in this sense, as they viewed ethnic or racial groups not merely as less human than other groups but as categorically and intrinsically subhuman. However, some forms of dehumanization recognized by contemporary psychological researchers are entirely relative in this sense. The infrahumanization effect involves the lesser attribution of uniquely human emotions to the out-group relative to the in-group and therefore need not involve any absolute or wholesale denial of the out-group’s capacity to experience refined emotions. Misunderstanding this point has implications for some challenges to the role of dehumanization in violence, to which I will return later in this chapter.
14.2.3 Animal versus Object
Prototypical examples of dehumanization invariably involve animal metaphors. Although animals may constitute the most salient contrast to the human category, and hence a primary basis for dehumanizing comparisons, Haslam’s (Reference Haslam2006) dual model insists that dehumanization can occur on a second register. The human/object contrast may not provide as many ready-to-hand dehumanizing metaphors as the human/animal distinction. Nevertheless, people spontaneously define humanness relative to unfeeling machines, have cross-culturally consistent understandings of human nature that are orthogonal to their understandings of human uniqueness, and differentiate humans from objects on a dimension of perceived mind that is more powerful than the one that differentiates humans from animals, according to Gray et al. (Reference Gray, Gray and Wegner2007). Failing to recognize how humans can be denied human nature and likened to objects would overlook many proposed manifestations of dehumanization that do not fit the animalistic mold. These examples include the literal objectification of women (Loughnan et al., Reference Loughnan, Haslam, Murnane, Vaes, Reynolds and Suitner2010), the reduction of patients to unfeeling bodies in some medical practices (Haque & Waytz, Reference Haque and Waytz2012), and the treatment of employees as fungible instruments in organizations (Brison et al., Reference Brison, Stinglhamber and Caesens2022). The concept of dehumanization is necessarily broader than a single framing of humanness in opposition to animals allows.
14.2.4 Simple versus Complex
A final contrast is between simple models of dehumanization and more complex ones. The simple case that dominates standard conceptualizations of dehumanization is that dominant group A holds and expresses a dehumanizing view of subordinate group B. Recent psychological work on the topic complicates this picture in several ways: 1) A may dehumanize itself; 2) A and B may dehumanize one another reciprocally; and 3) A and B may both believe that the other dehumanizes them.
The first of these complexities may seem implausible, but researchers have documented several cases in which people hold transient dehumanizing self-perceptions, sometimes in response to externally inflicted events such as social exclusions (Bastian & Haslam, Reference Bastian and Haslam2010), sometimes as a side effect of engaging in violent behavior (Bastian et al., Reference Bastian, Jetten and Radke2012), and sometimes as an enabler of immoral action (Kouchaki et al., Reference Kouchaki, Dobson, Waytz and Kteily2018). A perceived need to restore humanity to the self or group may motivate moral conduct.
The second complexity, reciprocal dehumanization, has been studied primarily by Kteily and colleagues in recognition that dehumanizing perceptions and the actions they drive may engender a matching response. Kteily and Bruneau (Reference Kteily and Bruneau2017), for example, demonstrated that Latinos and Muslims in the US held dehumanizing perceptions (assessed on the Ascent scale) of Republicans and Republican candidates during the 2016 Republican Primaries, in response to accurately perceived dehumanization by them, and the former perceptions drove reduced willingness to assist with counterterrorism efforts and increased desires to engage in violent protest. Their work demonstrated that dehumanization can be cyclical, mutual, and self-reinforcing and that it may be symmetrical even in asymmetrical conflicts.
The third complexity is linked to the second: A major trigger of reciprocated dehumanization is the perception that one group has been dehumanized by the other. This “meta-dehumanization” – feeling or learning that one’s group is not seen as fully human by an out-group – has become a major focus of research, starting with the work of Kteily et al. (Reference Kteily, Hodson and Bruneau2016). Their studies found, for example, that feeling dehumanized was distinct from feeling disliked or hated (meta-prejudice) and that meta-dehumanization was associated with support for aggressive actions and policies toward out-groups independently of meta-prejudice. Recent work underscores the vital importance of meta-dehumanization, showing that it predicts hostility more strongly than meta-prejudice (Landry et al., Reference Landry, Ihm and Schooler2021) and that meta-humanization – feeling or learning that one’s group is seen as fully human – may effectively counteract prejudice (Pavetich & Stathi, Reference Pavetich and Stathi2021).
To summarize this section of the chapter, the recent psychology of dehumanization has extended its meanings well beyond the standard understanding embodied in well-known historical examples from the Holocaust and the Rwandan genocide. The prototype of dehumanization is blatant, categorical, and one-directional in applying an animal metaphor to a target group. However, dehumanization extends to cases that are subtle and unconscious, that are matters of degree defined relative to an in-group rather than absolute denials of humanity, that assimilate humans to inanimate objects rather than animals, and that occur in reciprocal cycles driven by meta-perceptions.
14.3 The Roles of Dehumanization in Intergroup Conflict
Researchers have identified an extensive range of adverse correlates or consequences of dehumanization, in relation to a very extensive range of social groups (see Haslam & Loughnan, Reference Haslam and Loughnan2014, for a review). Studies have shown that people who dehumanize a group are especially likely to respond to that group with dislike, disgust, callous neglect, racism, and social rejection. Dehumanizers demonstrate greater willingness to harm and greater unwillingness to help those they dehumanize. They are more likely to support policies and practices of torture, harsh punishment, retaliatory violence, military intervention, and forced population displacement. They show greater opposition to immigration and the integration of minority groups and support active discrimination against them. Men who dehumanize women demonstrate a greater propensity to sexually harass them, to endorse sexist attitudes, and to score high on measures of proclivity to rape. People who dehumanize a group report less guilt and moral concern over injustices that befall it.
The span of dehumanization’s potential consequences appears to be broad indeed. Even so, precisely how those consequences occur is not always evident. Much of the research evidence is correlational, unable to determine to what degree dehumanization plays a causal role in the harms with which it is associated or to clarify its mechanisms or functions. Although a great deal of further research is required to elucidate these mechanisms and functions, it is helpful to distinguish them temporally, according to whether they precede, accompany, or follow a harmful action or inaction.
14.3.1 Before Harm
The role of dehumanization in conflict has often been thought of as preparatory. Dehumanizing a group is seen as creating the conditions under which violence flourishes. Several lines of theory and research help to comprehend how that preparation occurs. The most relevant approaches to dehumanization for this purpose are those that conceptualize it as an enduring set of beliefs or perceptions rather than as an episodic phenomenon or one that motivates specific actions.
The infrahumanization effect, for example, implies that ethnocentrism is a bedrock phenomenon in which groups perceive themselves to be more human than others. At the most basic level, Leyens and colleagues would claim, groups are predisposed to judge others as less good and less human than themselves. Intergroup relations start from a position in which a subtle degree of dehumanization is present even before conflict arises, and infrahumanization may make conflict more likely to develop than if that imbalance in perceived humanness did not exist.
Approaches to dehumanization grounded in trait-like dimensions of humanness or mind or in models of stereotype content are similar in this regard. In the dual model, out-groups can be located on dimensions of human uniqueness and human nature according to their members’ perceived attributes, and those locations may influence how more severe forms of dehumanization may arise under conditions of conflict. For example, out-groups perceived as lacking in uniquely human qualities (e.g., perceived as unintelligent and lacking in self-control) are at risk of being seen as bestial and barbaric during conditions of conflict, and the more they are seen as lacking those qualities the greater that risk may be. The same may be inferred from the mind perception account of dehumanization. Groups that are judged to lack one dimension of mind may be especially likely to be mistreated on account of their associated (perceived) deficiencies. A group believed to lack Experiential mental capacities should be especially likely to be treated badly when conflict arises, as it might be seen as lacking the capacity to suffer. This link between group stereotypes and vulnerability to mistreatment is clearest in Harris and Fiske’s (Reference Harris and Fiske2006) model of dehumanization, which identifies groups stereotyped as low in warmth and competence as at particular risk.
14.3.2 During Harm
Dehumanization in these varied theoretical models can help to explain how and why intergroup conflict may erupt in harmful behavior. Other models emphasize how dehumanization enables aggression in the present by releasing restraints on harming others. This element is most prominent in Bandura’s early work on moral disengagement (e.g., Bandura et al., Reference Bandura, Underwood and Fromson1975). In effect, Bandura proposes that perceiving the target of potential harm as less than fully human overcomes normal inhibitions that block it, such as anticipatory guilt and moral compunction. In the heat of conflict, dehumanizing the out-group may enable harmful behavior and dissolve objections to it if it is being carried out by members of one’s own group.
Delegitimization and inflammatory metaphors can also be understood to play a part in enabling harm or violence in the present. Bar-Tal’s (Reference Bar-Tal, Bar-Tal, Grauman, Kruglanski and Stroebe1989) work on intergroup conflict emphasized how withdrawing the legitimacy of an enemy group can take place in the emotional heat of conflict and is commonly expressed on vilifying labels. In this line of work, dehumanizing language expresses and reinforces a view of the other group as undeserving of concern in the here and now.
14.3.3 After Harm
Although dehumanization is usually examined as a process that prepares the ground for future harm and enables the commission of present harm, researchers have also examined how people dehumanize victims after harm has been perpetrated against them. In this work, people have several motives for dehumanizing harmed out-groups, including rationalizing the harm, minimizing its magnitude, and reducing guilt and collective responsibility for committing the harm. For example, Castano and Giner-Sorolla (Reference Castano and Giner-Sorolla2006) found that people whose nations had killed indigenous inhabitants of their colonies denied those inhabitants the capacity for uniquely human emotions more when reminded of past atrocities. Similarly, Cehajic et al. (Reference Cehajic, Brown and Gonzalez2009) showed that Serbians reminded of atrocities committed against Bosnian Muslims during the Bosnian War infrahumanized them. Other research has suggested that dehumanization may serve not only to whitewash the past but also to be an obstacle to future reconciliation. Tam and colleagues (Reference Tam, Hewstone, Cairns, Tausch, Maio and Kenworthy2007) showed that Northern Irish participants on the Protestant and Catholic sides who denied uniquely human emotions to their adversaries were less inclined to forgive, to abandon historical grievances, and to seek rapprochement.
14.3.4 The Case of Genocide
These discussions indicate that dehumanization may be thoroughly enmeshed in several aspects of intergroup conflict rather than merely setting the stage for it. The multiple involvements of dehumanization across the life history of conflict can be combined to offer an expanded framing of the role of dehumanization in genocide in particular (see Haslam, Reference Haslam and Newman2019, for a more fully developed analysis). According to one prominent model of the genocidal process (Stanton, Reference Stanton1998), dehumanization is a factor that is specific to the third of eight stages. Us–them classification of in-group and out-group (stage 1) is succeeded by the symbolizing of group identities by labeling or visible signs such as distinctive dress (stage 2). Dehumanizing expressions, taking the form of hateful animal metaphors, are then disseminated by perpetrator groups (stage 3) and promote the formation of organized militias (stage 4). Extremist voices polarize intergroup relations (stage 5) and extreme actions such as segregation of the victim group (stage 6) prepare the way for mass killings (stage 7), which are covered up after the fact by concealment of evidence and denials (stage 8).
This stage model may be accurate as far as it goes, but its delimited role for dehumanization rests on a narrow understanding of the phenomenon as blatant, absolute, animal metaphor-based, and one-directional. As I have shown, contemporary dehumanization theory and research provide a much more expansive account of dehumanization’s varied forms. Taking up that account allows for fuller understanding of the role that dehumanization might play in multiple stages of the genocide process.
For example, infrahumanization research indicates that even prior to conflict, in-groups tend to perceive out-groups as less human than they are, ensuring that even at stage 1 the us–them distinction incorporates a tacit assumption that the out-group is composed of lesser humans. Similarly, research indicates that the derogatory group labels and stereotypes of Stanton’s stage 2, which in theory occur prior to the overt adoption of dehumanizing animal metaphors, are often accompanied by covert and perhaps unconscious dehumanizing associations. The work of Goff et al. (Reference Goff, Eberhardt, Williams and Jackson2008), for example, showed that White Americans implicitly associate Black Americans with apes and that unconscious metaphorical association is implicated in hostile and punitive responses to Black targets. Implicit forms of dehumanization may therefore be consequential aspects of the symbolization of out-groups, before dehumanizing metaphors are spread by organized propaganda.
The processes of planning, polarization, segregation, and killing (stages 4–7) may also be propelled by the dynamics of dehumanization rather than merely set up by them, as Stanton’s stage model asserts. This view is shared by genocide scholars Haagensen and Croes (Reference Haagensen and Croes2012), who write: “Dehumanization … does not only play a role in preparing the ground for genocide; it also drives genocidal campaigns forward, sustaining killing operations over lengths of time” (p. 225). For example, Kteily et al.’s (Reference Kteily, Hodson and Bruneau2016) work demonstrates that the blatant dehumanization of out-groups is associated with endorsement of extreme or polarized policy positions, a finding also obtained by Maoz and McCauley (Reference Maoz and McCauley2008) in a study of the Israel-Palestine conflict. Moral exclusion surely plays a role in driving popular support for segregation of a targeted minority, and moral disengagement has been shown to release inhibitions on violent action, potentially including acts of mass killing. Stage 8 denial may also implicate dehumanization, which has been shown to help reduce collective guilt, rationalize past in-group violence, diminish support for providing reparations to victims, and minimize perceptions of their suffering in contexts including indigenous–settler relations, civil war, and religious strife.
In sum, it is possible that out-group dehumanization plays a variety of roles through the genocide process rather than being confined to a single stage. This speculation may even be too timid, as other forms of dehumanization may also be implicated. Self-dehumanization (Kouchaki et al., Reference Kouchaki, Dobson, Waytz and Kteily2018) may play a causal role in enabling the commission of violence or other immoral actions. Some descents into genocide may also be accelerated by reciprocal dehumanization between conflicting groups. When dehumanization is understood in an expansive way that is grounded in recent research and scholarship, the processes through which it might contribute to particular cases of violence are many and varied, extending well beyond the traditional emphasis on hateful propaganda.
14.4 Critiques
Over the past decade, several scholars have challenged the role of dehumanization in intergroup conflict in general or genocide in particular. Their interconnected critiques constitute a backlash of sorts to the growing dehumanization literature, which has the potential to refine, qualify, and enhance that literature. Three main challenges can be identified: that the role of dehumanization in conflict has been overstated; that some violence attributed to dehumanization only makes sense if the victim is perceived as human; and that some supposed cases of dehumanization are better understood in terms of valence rather than humanness. These challenges are interlocking but they will be examined separately.
The argument that dehumanization’s role in collective violence has been exaggerated has been articulated cogently by Lang (Reference Lang2010, Reference Lang2020). Lang maintains that genocidal violence is grounded much more in hatred than in prejudice and that the phenomenology of group-based cruelty often makes sense only in the context of a desire to punish and hurt despised fellow humans. Lang and colleagues set up their critique in opposition to a putative “dehumanization hypothesis” that claims that “dehumanization of the victims is a necessary pre-condition for killing them” (Brudholm & Lang, Reference Brudholm, Lang and Kronfeldner2021, p. 346).
Lang’s challenge is a crucial one. Dehumanization scholars should resist the urge to inflate the role of dehumanization in collective violence and should evaluate carefully how or when dehumanization, hatred, their toxic combination, or something else entirely accounts for it. Nevertheless, the challenge attacks a straw man, as contemporary dehumanization researchers – in psychology at least – make no strong claims about the necessity of dehumanization as a precondition for killing. Instead, they have proposed a variety of ways in which dehumanization might contribute to violence. Recent scholarship offers an account of dehumanization that is broader and more textured than the narrow-sense account held by some early genocide scholars, and it offers an expanded range of dehumanization processes that might be implicated in violence. Determining whether or to what degree these processes operate in actual genocides is a task for further investigation. That task will be difficult because so many of the proposed processes occur in the conscious and unconscious minds of the protagonists during the conflict rather than in the public record. A narrow understanding of dehumanization that only sees evidence of it when hateful animal metaphors are used in public fora will overlook the many subtle ways that dehumanizing perceptions operate in private minds and hearts and in interpersonal and intergroup dynamics. Ideally, future studies will avoid maximalist claims that treat the centrality of dehumanization to genocide as self-evident and minimalist accounts that see it as marginal, based on constricted understandings of dehumanization.
A second challenge to dehumanization scholarship is the claim that many acts of violence ascribed to dehumanization only make sense if the perpetrator perceives the victim as human. Efforts to inflict pain and humiliation depend on the conviction that victims are not insensate objects and that they have the uniquely human capacity to feel shame and the degrading loss of human dignity. Arguments of this kind about the “dehumanization paradox” have been proposed by Appiah (Reference Appiah2006) and Manne (Reference Manne2016), among others. If dehumanization were the total, categorical denial of a person’s humanity, such that they were perceived as entirely equivalent to objects or beasts, this argument might have merit. As we have seen, however, dehumanization is not conceptualized in this fashion in contemporary theory and research. In most psychological accounts it is understood to vary in degrees of severity, including cases where the person’s humanness is perceived as attenuated rather than annihilated. In another recent psychological account, Kteily and Landry (Reference Kteily and Landry2022) defend the related view that multiple aspects of humanness may be denied in variegated ways, rather than all being denied at once. In Smith’s account, the categorical denial of humanness is accompanied by a simultaneous representation of the dehumanized person’s humanity. These conceptualizations all allow that a person can be perceived as unequivocally human – albeit a lesser human – and therefore capable of pain, mortification, and other moral emotions, though perhaps to a lesser degree than perpetrators arrogate to themselves. As Kronfeldner (Reference Kronfeldner2021b, p. 15) argues, the paradox of dehumanization “can easily be dissolved since the core of it rests on an equivocation of ‘being human’: many cases of dehumanization seem to involve a recognition of the bare humanity of the targets, while (parts of) their humanness and/or their moral standing is ignored or destroyed.”
The third main challenge to recent dehumanization research queries whether dehumanization is in fact redundant with negative valence: whether it is genuinely distinct from hatred, dislike, or prejudice. This challenge has recently been posed by Over and colleagues (e.g., Enock et al., Reference Enock, Flavell, Tipper and Over2021; Over, Reference Over2021), who argue that some findings that claimed to reflect infrahumanization or dehumanization in fact reflect intergroup preference – the subject of a vast social psychological literature – rather than denials of humanness. Their work, the empirical details of which are beyond the scope of this brief review, is significant in drawing critical attention to the conceptual claims and evidence base of social psychological research on dehumanization. Although some writers have stressed the importance of differentiating dehumanization from negative evaluation both conceptually and empirically (e.g., Haslam, Reference Haslam, Bain, Vaes and Leyens2013), researchers have not consistently done so in their studies, and it is therefore probable that some findings ascribed to the dehumanization of a group may partially or wholly reflect prejudice or dislike toward it.
Nevertheless, the argument that dehumanization as studied in social psychological research is nothing but negative evaluation is highly implausible. Emotion- and trait-based measures of humanness have deliberately controlled for the valence of their emotions and traits and demonstrated effects of humanness in group perception tasks that are separable from valence effects, including studies that statistically control for valence (e.g., Haslam et al., Reference Haslam, Bain, Douge, Lee and Bastian2005). Studies using the “ascent of humans” measure of dehumanization (e.g., Kteily et al., Reference Kteily, Bruneau, Waytz and Cotterill2015) show, unsurprisingly, that the extent to which groups are dehumanized correlates moderately with the extent to which they are evaluated negatively, as assessed by a feeling thermometer. However, they also show that dehumanization predicts responses to groups independently of prejudice. In some cases, dehumanization-based measures predict the harshness of responses more strongly than prejudice-based measures (Landry et al., Reference Landry, Ihm and Schooler2021). Dehumanization and negative evaluation even appear to have distinct neural signatures (Bruneau et al., Reference Bruneau, Jacoby, Kteily and Saxe2018). Findings such as these indicate that dehumanization and negative valence are often entwined – the groups we see as less than fully human are often the ones we dislike – but they are incompatible with the view that one can be reduced to the other.
14.5 Conclusions
Scholarship on dehumanization has burgeoned in the past two decades, generating an assortment of new theoretical accounts and a voluminous empirical literature. The rapid growth of this work has produced substantial evidence for the malign implications of dehumanization in the context of intergroup conflict. It has also generated an array of alternative conceptualizations and often poorly specified definitions that has triggered substantial debates and disagreements and motivated a range of conceptual and empirical critiques. Humanness is surely a central concept in moral psychology, and the stage is set for a new generation of scholars to clarify it and to determine how it can best be deployed to understand the continuing fact of human inhumanity.
15.1 Morality and Its Regulation
To regulate group living, humans have developed two related social-cultural tools: a system of norms and complex social practices of regulating those norms. Central to norm regulation are responses to violations of norms. They include conciliatory ones, such as tolerance or forgiveness, and corrective ones, such as blame and punishment. Blame is often treated as parallel to punishment, as two currencies of moral sanction or norm enforcement (Ames & Fiske, Reference Ames and Fiske2013; Berger & Hevenstone, Reference Berger and Hevenstone2016; Cushman, Reference Cushman2008). Indeed, both blaming and punishing can be pedagogical (Cushman, Reference Cushman2015; Malle et al., Reference Malle, Guglielmo and Monroe2014), and both impose costs on the offender. However, the two are also importantly distinct (Baumard, Reference Baumard2011; Buckholtz et al., Reference Buckholtz, Martin, Treadway, Jan, Zald, Jones and Marois2015; Malle, Reference Malle2021). Blaming a transgressor is both a particular type of moral judgment (Malle, Reference Malle2021) and, when socially expressed, an overt act of moral criticism. Punishment is always an overt act – and not merely one of criticism but one of damage. Because it would, in other circumstances, violate a person’s rights, punishment must be warranted or legitimized, typically by roles or institutions, such as parent, teacher, or state. Blame, too, must be warranted (Coates & Tognazzini, Reference Coates, Tognazzini, Coates and Tognazzini2012; Malle et al., Reference Malle, Guglielmo, Voiklis and Monroe2022), but blame can be withdrawn when found unwarranted whereas punishment cannot. Once inflicted, the damage of punishment is rarely reversible.
This is just a rough comparison. In this chapter, I investigate blame and punishment from two perspectives: their cultural history and their psychology. I also hope to make plausible the hypothesis that the distinct cultural histories of blame and punishment have shaped their distinct psychology.
I should acknowledge at the outset that some literatures use the term punishment as another word for all forms of norm enforcement or sanction and then subsume blame (as verbal criticism) under the label punishment (for a more nuanced form of this position, see Chapter 16, this volume). I agree that both blame and punishment are forms of norm enforcement, but I hope to convince the reader that distinguishing between the phenomena of blame and punishment as well as between the corresponding terms helps clarify the important differences among the various tools of human norm enforcement.
15.2 Cultural History of Blame and Punishment
In human cultural evolution, responses to an individual’s norm violations have undergone enormous changes, often characterized as overcoming a brutal past (Farrington, Reference Farrington1996) to more “humane” present times (Pinker, Reference Pinker2011). However, if we look closely, there were really two distinct phases of human social living (Dubreuil, Reference Dubreuil2010) that gave rise to two different responses to norm violations: blame and punishment.
The first phase was a biological evolution from often violent dominance hierarchies in our primate ancestors to more egalitarian forms of human social living (Boehm, Reference Boehm2000), beginning perhaps as early as homo erectus, 1 million years ago (Dubreuil, Reference Dubreuil2010). As humans evolved in small communities of hunter-gatherers, they enforced social-moral norms with sanctions that were primarily informal, interpersonal, and relatively mild (Wiessner, Reference Wiessner2005) – much like today’s informal acts of blame as moral criticism.
The second phase began with human settlements after 10,000 BCE, leading to the formation of chiefdoms and early states, rapid population growth, and eventually culminating in nations and empires with large populations. Crucially, these societies reintroduced dominance hierarchies but legitimized them, not through our primate ancestors’ raw strength and aggression, but through wealth, inheritance, military, and religious structures. Many norms became codified into law, shaping a system of institutionalized punitive sanctions imposed on the rest of the community by those at the top of the dominance hierarchies (e.g., chief, king, state).
I now flesh out this sketch of how two phases of human evolution gave rise to two forms of sanctioning, which form the origins of today’s tools of blame and punishment.
15.2.1 Hunter-Gatherer Communities and Moral Criticism
Until about 10,000 years BCE, humans lived as hunter-gatherers in small bands of 25–50 (Boehm, Reference Boehm1999; Knauft, Reference Knauft, Sponsel and Gregor1994). We know this from some archeological finds (Bandy, Reference Bandy2004; Enloe, Reference Enloe2003), population genetic analyses (Atkinson et al., Reference Atkinson, Gray and Drummond2008; Henn et al., Reference Henn, Gignoux, Jobin, Granka, Macpherson, Kidd, Rodríguez-Botigué, Ramachandran, Hon, Brisbin, Lin, Underhill, Comas, Kidd, Norman, Parham, Bustamante, Mountain and Feldman2011), but predominantly from ethnographic research of hunter-gatherer societies over the past 100 years (Lee & Daly, Reference Lee and Daly1999; Wilson, Reference Wilson1988; Woodburn, Reference Woodburn1982). According to this accumulated evidence, hunter-gatherer communities were nomadic and thus highly mobile, changing camp every few weeks to months (Lee, Reference Lee and Spooner1972), mostly lacking possessions, and having little sense of territorial boundaries (Wilson, Reference Wilson1988). They were highly egalitarian and, without one supreme ruler, lawmaker, or judge, leadership was provided by different members for different tasks (Service, Reference Service1966).
The hunter-gatherer norm system was grounded in reciprocity, which regulated both community living and vital activities such as food acquisition and consumption. When a hunter shared his yield with another hunter one day, he could expect to receive part of the other’s yield another day. Indeed, sharing large hunts counted as a “virtually universal rule” among hunter-gatherer societies (Wilson, Reference Wilson1988, p. 37), and violating this sharing norm was met with sanctions.
In these highly interdependent bands, sanctioning was interpersonal. Most transgressions were easy to detect (Silberbauer, Reference Silberbauer, Leacock and Lee1982) because life was public and transparent. But sanctions of norm violators rarely consisted of punishment (Baumard, Reference Baumard2010; Guala, Reference Guala2012), since these egalitarian groups disdained assertations of power and coercion, and debilitating penalties would have hurt the community at least as much as the transgressor. Instead, communities favored communication, criticism, sometimes ridicule (Wiessner, Reference Wiessner2005), and gossip (Dunbar, Reference Dunbar1996). Repeat offenders may have been isolated or, as a last resort, expelled from the group (Woodburn, Reference Woodburn1982). But the typical way of responding to norm violations was a form of public criticism, a threat to the person’s social standing and reputation – which, in small, interdependent groups, largely ensured norm compliance.
These civil forms of sanctions, however, occurred within groups and did not prevent highly punitive actions between groups. How much warfare occurred in early human hunter-gatherer communities is debated. But few scholars doubt that intergroup violence gradually increased with population density and with the correlated expansion from extended families to clan-based networks (Flannery & Marcus, Reference Flannery and Marcus2012). These growth patterns set the stage for a transformation in the social organization of human societies.
15.2.2 The Emergence of Punishment Post Settlement
The second phase in the cultural evolution of sanctioning methods occurred after humans settled down, beginning around 12,000 years ago. Norm enforcement turned from an interpersonal process of criticism and reputation threat to an institutionalized process that confronted the perpetrator with the damaging actions of a ruling entity and its henchmen. The transition occurred in different places at different times, but the endpoint was, with few exceptions, a hierarchical society that centralized and legitimized harsh punishment. A few essential facts about sedentary life help explain this change (Lee, Reference Lee and Spooner1972; Peregrine et al., Reference Peregrine, Ember and Ember2007; Redman, Reference Redman1978).
Property. As hunter-gatherers, people possessed only tools, tents, and clothes; food was distributed and consumed right away; and land had no boundaries of ownership (Wilson, Reference Wilson1988). After settling down, more and more people had permanent housing, household effects, and eventually land, livestock, and crops. Quickly, some gained more property than others. Such inequalities were enlarged by inheritance rights, intermarriage, as well as reciprocal trade and protection among property owners (Peregrine et al., Reference Peregrine, Ember and Ember2007). The have-nots desired others’ property, and the crimes of robbery and burglary emerged. The haves feared losing their property, so the need for protection, law enforcement, and punishment increased.
Population Growth. Between 10,000 and 8,000 BCE, population growth was spurred by environmental changes (the end of the Last Glacial) and geographic opportunities (e.g., the fertile Near East; Aurenche et al., Reference Aurenche, Kozłowski and Kozłowski2013), turning camps into villages and towns (Atkinson et al., Reference Atkinson, Gray and Drummond2008; Gignoux et al., Reference Gignoux, Henn and Mountain2011). Settlements also increased the rates of offspring. Whereas nomad communities had to carry their newborns for thousands of kilometers a year, limiting child birth to once every 3–4 years (Lee, Reference Lee and Spooner1972), settled communities were able to dramatically increase the frequency of pregnancies (Buikstra et al., Reference Buikstra, Konigsberg and Bullington1986). Furthermore, plant and animal breeding experiments as well as improved food production and storage technologies fed more people. In a positive feedback loop, the resulting population growth put pressure on food production and demand for labor. Consequently, a second population growth turned towns into cities, several of which grew to 30,000 inhabitants or more between 2000 BCE and 1200 BCE. Another feedback loop then emerged between rapidly increasing built structures (Wilson, Reference Wilson1988) – from private homes to royal monuments and sacral edifices – and the massive opportunities for work to sustain this building boom (Boserup, Reference Boserup1965). That work was likely done by those who didn’t own property – the poor, foreigners, and conquered populations; thus, dependent labor and class stratification accelerated.
Intergroup Conflict. By and large, hunter-gatherer societies were peaceful (Kelly, Reference Kelly2000; Silberbauer, Reference Silberbauer, Leacock and Lee1982). There was little reason to attack another band (what do they have that we don’t?). There was also little reason to fight over territory that constantly shifted (Woodburn, Reference Woodburn1982). But once there were settlements, some land was more fertile, some villages more productive, and fast-growing populations had a need to either intensify food production (Carneiro, Reference Carneiro1970) or expand their territory (Dumond, Reference Dumond and Spooner1972), especially because food resources may not have kept pace with population growth (Larsen, Reference Larsen1995). Thus emerged incentives for organized warfare (Milner et al., Reference Milner, Anderson and Smith1991; Redman, Reference Redman1978) to gain land, food, raw materials, tools, and to acquire further labor by subjugating other communities. As potential targets for attack and raids, settled communities bolstered their safety by developing weapons and defensive facilities, which further demanded an expanded labor force and eventually a professional army.
Hierarchical Social Structure. All these changes – expanded and unequally distributed property, population growth, dependent labor, and intergroup hostilities – demanded organization and administration, taken on by a few that turned into elites and authorities (Redman, Reference Redman1978). New crimes emerged when this authority was violated, such as treason, social revolt, or tax evasion. Leadership structures were also necessary to maintain within-group cohesion in light of new role differentiation and resulting conflict potential (White, Reference White2007). Settled humans were arranged from slaves to freemen to landowners, eventually aristocrats and priests, and almost always a supreme ruler. In the hundreds of thousands of years prior to settlement, primate hierarchies were transformed into egalitarian, cooperative human bands; within just a few thousand years, a new society was born, returning to hierarchical and often brutally competitive primate roots.
15.2.3 Escalation of Punishment
In small hunter-gatherer communities, a norm violator had equal-status relationships with other group members, and these relationships had to be maintained for continued group success. Sanction severity for norm violations was therefore restrained because harsh penalties might invite equally harsh responses in the future and harm community relationships. By contrast, in larger settlements, ties between norm violators and the ruling class became weak and asymmetric, promising scant reciprocal future interaction and setting few limits on sanction severity. Over thousands of years of growing social hierarchies, we therefore see both increasingly centralized and increasingly brutalized forms of punishment. Several causal processes contributed to such escalation of punishment.
Punishment became more severe when rulers defended themselves against threats to their position. For example, in eighteenth-century Tahiti, the chief left it to family heads to enforce rules about property or marriage but applied severe punishment when offenses were “sacred,” typically those against his interests (Claessen, Reference Claessen and Feldbrugge2003). In nineteenth-century despotic African Buganda, sacred crimes, especially those against the ruler, were met with punishments of maiming, cutting into pieces, and burning alive (Claessen, Reference Claessen and Feldbrugge2003). Political rulers in China’s hierarchical societies BCE (Bozan et al., Reference Bozan, Shao and Hu2001) also relied on stern rules and punishment, especially during times of civil unrest. The death penalty applied to many offenses against the emperor (e.g., treason, rebellion) and the state (e.g., malpractice, bribery, illicit sales). In the early Roman Republic, laws were in tight control of the ruling class and were even kept secret from the lower class. When the laws were finally publicized in the Twelve Tables (c. 450 BCE), they contained detailed civil statutes but also prescribed capital punishment for those who defamed others in song (VIII.1b) – to avert any chance of inciting the proletariat’s sentiments against the ruling elite. Over the centuries, wars and uprisings caused tyrannic rulers to take the reins of the Roman Republic, and punishments became excessively cruel, even public spectacles, to deter any threat to the elites (Moore, Reference Moore2001).
War is a driver of severe punishment. It brutalizes relations between communities, and it brutalizes relations within. When people are exposed to war, violence is part of life, and aggressive behavior becomes disinhibited (MacManus et al., Reference MacManus, Dean, Al Bakir, Iversen, Hull, Fahy, Wessely and Fear2012). In wartime, rulers and state governments also more easily justify violent punishment within the state, especially for crimes of treason or cowardice. For example, before and during World War II, Nazi Germany took capital punishment to incomparable extremes, with a huge increase in death warrants and executions (Messerschmidt, Reference Messerschmidt2005). The death penalty was applied to a wide range of criminal acts, various forms of military disobedience, and many forms of political resistance.
Despite the brutality of punishment in many early societies, it did not succeed in deterring crime, which then further incited harsher punishments. We see this cycle in modern times as well. For example, in the United States’ “war on crime” of the 1980s (Pratt et al., Reference Pratt, Brown, Brown, Wallsworth and Morrison2005), when punishment trends hardened in response to (perceived) rising crime rates but failed to loosen when crime rates declined. To this day, increased institutional punitiveness does not result in crime reduction (Cullen et al., Reference Cullen, Jonson and Nagin2011).
We also see detrimental escalation processes in today’s school environment (Darling-Hammond et al., Reference Darling-Hammond, Ruiz, Eberhardt and Okonofua2023). As teachers increasingly punish students for even minor misbehavior – often because they regard the students as “troublemakers” (Okonofua & Eberhardt, Reference Okonofua and Eberhardt2015) – students become defiant and misbehave even more (Pesta, Reference Pesta2022), which cyclically escalates punishment. This escalation occurs primarily for minoritized students (Amemiya et al., Reference Amemiya, Mortenson and Wang2020).
Thus, both past and present reveal that punishment is a questionable method of moral regulation; it often expresses power, social hierarchies, and their dominance-maintaining mechanisms (Sidanius & Pratto, Reference Sidanius and Pratto1999).
15.3 Psychology of Punishment and Blame
I now shift from a cultural history of blame and punishment to the contemporary psychology of these phenomena. I begin with an outline of their broad differences and then devote more detailed attention to our current understanding of each form of sanctioning.
15.3.1 Distinguishing Punishment and Blame
Everyday moral sanctions range from raised eyebrows to muttered disapproval, from emotional complaints to thoughtful criticism; but there are also acts of physical aggression, withholding of resources, public shaming, and social exclusion. The first set is more representative of moral criticism (blame, for short) whereas the second set is more representative of genuine punishment.
Even though blame and punishment are not categorically distinct, a bundle of features allows us to distinguish them. Blame has a dual nature as a particular kind of moral judgment and as communicated moral criticism; punishment is always an observable act. When expressed, blame often requests or demands that the violator respond to the criticism and change their (future) behavior; punishment coercively imposes costs on the transgressor, often relying on power or status to legitimize the imposition of costs. Blame is typically reversible – one can admit that one unfairly blamed the other and take back the criticism – whereas the tangible costs of punishment can rarely be taken back. Finally, blame can be contested, leaving room for justification and negotiation; punishment leaves little room for negotiation and is often the last resort when moral criticism has failed. (For a different conception of punishment, incorporating the features of blame, see Chapter 16, this volume.)
In light of these distinguishing properties, we would expect that, in informal, everyday interactions, humans prefer to use moral criticism over punishment. Indeed, using daily surveys to study naturally occurring norm violations, Molho et al. (Reference Molho, Tybur, Van Lange and Balliet2020) found that people were generally unmotivated to engage in punishment as physical confrontation and somewhat more motivated to gossip or exclude the transgressor. Field studies show that people enforce norms with subtle verbal or nonverbal interventions but not active punishment (e.g., Przepiorka & Berger, Reference Przepiorka and Berger2016). In a cross-cultural study, rates of responding to violations with cost-imposing sanctions (e.g., yelling, insulting, hitting) were extremely low, between 1 percent and 6 percent (Pedersen et al., Reference Pedersen, McAuliffe, Shah, Tanaka, Ohtsubo and McCullough2020). Even though a number of behavioral economics experiments have suggested that people are willing to punish norm violators, this form of punishment is mild and indirect (taking a little money from a stranger). Moreover, when given a choice, people prefer not to punish but to criticize the violators to their face or to others (Feinberg et al., Reference Feinberg, Cheng and Willer2012; Kriss et al., Reference Kriss, Weber and Xiao2016; Xiao & Houser, Reference Xiao and Houser2005), to compensate the victim (Chavez & Bicchieri, Reference Chavez and Bicchieri2013; Hershcovis & Bhatnagar, Reference Hershcovis and Bhatnagar2017), or even to restore order by correcting the negative outcome (e.g., cleaning up other people’s litter; Berger & Hevenstone, Reference Berger and Hevenstone2016). As we move into more hierarchical relationships – parent and child, teacher and student, boss and employee – punishment becomes more likely, more consequential, and is often seen as more legitimate (Gershoff & Lee, Reference Gershoff and Lee2020; Mooijman & Graham, Reference Mooijman and Graham2018). Embedded in such hierarchies, “[p]unishment is not a behavior, but an institution” (Binder, Reference Binder2002, p. 321). Indeed, punishment is the predominant sanctioning instrument used by the state (Tonry, Reference Tonry2009).
15.3.2 Understanding Punishment
If interpersonal punishment is relatively infrequent in daily life, why is it so widespread in its institutionalized form? The brief cultural history of punishment has taught us that punishment has been used to maintain hierarchies, defend the powerful, and remove the unwanted. The severity of punishment has declined over time, but to this day, punishment is an instrument of power, oppression, and the maintenance of dominance hierarchies (Redford & Ratliff, Reference Redford and Ratliff2018; Sidanius et al., Reference Sidanius, Mitchell, Haley and Navarrete2006). And not only is punishment discriminatory, it is also ineffectual as an instrument of deterrence (Cullen et al., Reference Cullen, Jonson and Nagin2011). If punishment tends to fail in so many ways, what maintains the public support for institutional punishment?
15.3.3 Forces that Maintain Support for Punishment
Retribution. A first hypothesis is that people are fundamentally retributivists – desiring proportional and painful punishment of norm violators, even when it has no deterrent function (Carlsmith et al., Reference Carlsmith, Darley and Robinson2002; Goodwin & Gromet, Reference Goodwin and Gromet2014). But rather than explaining common support for institutional punishment, people’s retributive tendencies may be themselves explained by their exposure to institutions of punishment. Research on lay retributivism always probes whether people recommend or endorse retributive punishment for crimes, such as assault, rape, or murder. For those violations, people may endorse severe punishment because that, they have learned, is the proper response. However, people are not merely reflecting back the severity of punishment in legal reality; they are often manipulated by political rhetoric, more so from the right, that crime is rampant and society needs to be tougher on crime (Muhammad et al., Reference Muhammad, Gottschalk, Thompson, Travis, Western and Redburn2015; Tonry, Reference Tonry2009).
However acquired, what is it that human retributive tendencies really support? The current literature often defines retributive commitments as referring to “deserved punishment of a guilty offender” (Goodwin & Gromet, Reference Goodwin and Gromet2014, p. 562). People are retributivists, then, if they endorse a statement such as “by punishing offenders we give them what they deserve” (Nadelhoffer et al., Reference Nadelhoffer, Heshmati, Kaplan and Nichols2013, p. 244). But in what sense is deserved punishment “retributive” but a deserved award is not (Cottingham, Reference Cottingham1979, p. 239)? In both cases, the response is considered appropriate or justified, given what the person actually did. This notion of desert stands in contrast to the extreme consequentialist position that what the transgressor did (or thought while doing) is irrelevant; that the only thing that matters for punishment is its beneficial consequence (e.g., deterrence of crime, calming public outrage). Rejecting this position, people endorse retribution in the sense of paying back (Cottingham, Reference Cottingham1979), which is what the original twelfth-century English word referred to. Such retribution was a symmetric practice of moral accounting – both rewards and punishments were forms of paying back.
If retribution is the mirror image of the human disposition toward reciprocity for good deeds (Gouldner, Reference Gouldner1960), then being retributive need not imply cruel enjoyment of a wrongdoer’s suffering. Instead, a scale is being balanced between the transgressor’s behavior and the community’s norms; this scale, of course, is what the goddess Justitia holds in her hand to symbolize justice. Whether the repayment is cruel or civil – mutilations and executions, or monetary fines, apology, and atonement – depends on the historically and geographically specific currency in which moral debts are repaid. Thus, even if ordinary people’s support for retributive punishment is grounded in a general principle of reciprocity, it has undoubtedly been nourished by long-standing patterns of political and religious influence (Griffith, Reference Griffith2020).
Delegation. A second contributor to support for institutional punishment stems from the dangers of personal punishment. Punishment is typically an act of aggression and inflicts damage on the perpetrator, who may therefore retaliate. A perpetrator who has committed a serious transgression has already shown little commitment to the community’s norms; when punished, the person is likely to avenge the inflicted damage. In hunter-gatherer societies, personal interventions were more pressing and more likely in response to extreme violations (Kelly, Reference Kelly2000, p. 27). However, they came with significant risk. Lee (Reference Lee1979) reported that almost half of the !Kung community’s fatalities over several decades were bystanders or “peacemakers,” who died in the attempt to end a conflict. Retaliation against norm-enforcing punishment is even more likely to be committed by people outside one’s group because the perpetrators have no relationship with the norm enforcer or the community and thus no reason to show remorse, correct their behavior, and protect their reputation. Indeed, in human history, intergroup conflicts have often escalated from individual revenge to family- or group-based revenge, spiraling into blood feuds (Boehm, Reference Boehm2011).
In all these cases, the costs to directly punish a perpetrator are high, but the risks of retaliation can be reduced or even eliminated by handing the task of punishment to specific agents or institutions (Cushman, Reference Cushman2015). Such delegation of punishment protects ordinary people and allows them to safely condemn violators and support the punishment meted out by the authorities – even excessively harsh punishment. It is far easier to cheer on the torturer or vote for “tough on crime” policies than it is to break the prisoner’s arm oneself or stand guard at a solitary confinement cell. Delegated punishment, however, requires justification, which a ruler or state has in general and can award to particular officials, such as judges, prison guards, or executioners. Institutionalized punishment guarantees legitimacy without having to establish it anew each time (Binder, Reference Binder2002; Garland, Reference Garland1990). As long as the institution is seen as reasonably consistent and fair, the community will support it and grant it the permission to punish some of norm-violating members.
15.3.4 Does Punishment Foster Cooperation?
Leaving the institutionalized context, punishment in small groups can foster cooperation by decreasing the number of free riders and increasing the contributions individuals make to the group (Fehr & Gächter, Reference Fehr and Gächter2000; Yamagishi, Reference Yamagishi1986). Punishment works best when it articulates community norms (Andrighetto et al., Reference Andrighetto, Brandts, Conte, Sabater-Mir, Solaz and Villatoro2013; Xiao, Reference Xiao, Teitelbaum and Zeiler2018), teaches the violator (Cushman, Reference Cushman2015), brings about a response from the violator (Funk et al., Reference Funk, McGeer and Gollwitzer2014), and when it is gentle and deemphasizes power differences (Kochanska & Aksan, Reference Kochanska and Aksan2006).
But punishment can go awry. When cooperators are punished, costly punishment no longer promotes cooperation (Rand et al., Reference Rand, Armao, Nakamaru and Ohtsuki2010), and the option to punish can lead to retaliation (Hopfensitz & Reuben, Reference Hopfensitz and Reuben2009). When social communities view punishment as unwarranted, it is less effective at correcting social behavior (Herrmann et al., Reference Herrmann, Thöni and Gächter2008). Furthermore, impressions of third-party punishers are less than favorable (Dhaliwal et al., Reference Dhaliwal, Patil and Cushman2021), and even victims who punish are not seen very positively; instead, victims who forego punishment are seen as moral, trustworthy, and altruistic (Heffner & FeldmanHall, Reference Heffner and FeldmanHall2019).
Punishment may not even be necessary for cooperation (Baumard, Reference Baumard2010). Punishment’s apparent power to increase cooperation in mixed-motive games is matched (Feinberg et al., Reference Feinberg, Willer and Schultz2014) or surpassed (Wu et al., Reference Wu, Balliet and Van Lange2016) by that of gossip, and it is also matched by communicated disapproval (Masclet et al., Reference Masclet, Noussair, Tucker and Villeval2003). When we compare overall payoffs in these game contexts, both gossip and disapproval lead to better collective outcomes than does punishment (Dugar, Reference Dugar2010; Wu et al., Reference Wu, Balliet and Van Lange2016).
15.3.5 Punishment Judgments
Research has examined punishment not only as a behavior of imposing costs on another person but as ordinary people’s recommendations for the legal punishment a transgressor should suffer. Researchers have studied the kinds of information people are sensitive to and the various judgment biases they may fall prey to. On the information side, there is consistent evidence that people assign different degrees of punishment (e.g., length of prison term or monetary fines) as a function of violation severity, intentionality, foreseeability, and justification (Kneer & Machery, Reference Kneer and Machery2019; Robinson & Darley, Reference Robinson and Darley1995), largely consistent with what the law prescribes. In addition, however, studies document that punishment recommendations are susceptible to factors that should arguably not influence legal judgments – for example, the defendant’s ideology (Sood & Darley, Reference Sood and Darley2012), physical or social attractiveness (e.g., Nemeth & Sosis, Reference Nemeth and Sosis1973), stereotypically Black appearance (Eberhardt et al., Reference Eberhardt, Davies, Purdie-Vaughns and Johnson2006), or the juror’s prejudice (Gamblin et al., Reference Gamblin, Kehn, Vanderzanden, Ruthig, Jones and Long2021).
But while several studies do confirm biased recommendations for degrees of punishment, numerous other studies failed to find similar biases in people’s verdicts – that is, judgments of whether a person is guilty or not (Block, Reference Block1991; Stewart, Reference Stewart1985). Why might people be more calibrated in their judgments of guilt than in their punishment recommendations? First, punishment and sentencing recommendations are inherently more discretionary, comparative, and subjective than are guilty verdicts (Tata, Reference Tata1997), creating more room for errors and biases. Second, at least in the United States, people don’t normally make punishment recommendations – as jurors, their role is to come to a verdict, whereas judges decide on punishment. Thus, research that measures legal punishment recommendations may ask people for unusual, unfamiliar judgments, which may be more susceptible to biasing information. Third, previous studies often treated past transgressions, character, and motive as “biasing” factors when in fact they often are legally admissible for sentencing, if not for guilty verdicts (Guglielmo, Reference Guglielmo2015); so by and large, people’s judgments are consistent with the law.
In no way does this imply that verdicts by research participants or real jurors are generally unbiased. One of the most enduring biases in the United States criminal system is racial discrimination, which exerts its impact through police work, prosecution decisions, pretrial detention, guilty pleas, testimony, all the way to verdicts and sentencing (Kansal, Reference Kansal2005; Mitchell et al., Reference Mitchell, Haw, Pfeifer and Meissner2005; Sutton, Reference Sutton2013). Research is also growing to document similar biases toward other marginalized groups (e.g., Mirabito & Lecci, Reference Mirabito and Lecci2021). Such discrimination replays the hierarchical dominance patterns that have tilted punishment systems since early human settlement.
15.3.6 Understanding Blame
We have seen that people rarely mete out punishment in everyday life, and punishment recommendations occur more often in research studies than in the courtroom. By contrast, blame judgments occur frequently and naturally in everyday life. Interestingly, however, blame does not have the best of reputations in psychological research. Acts of blaming are often portrayed as distorted and dysfunctional. Some philosophical work, by contrast, has characterized blaming as more constructive moral criticism that calls out unacceptable behavior, affirms norms and values, and demands justice (Bell, Reference Bell, Coates and Tognazzini2012; Ciurria, Reference Ciurria2019). Research on “complaints” in sociology and pragmatics further shows the important function that criticism plays in regulating interpersonal behavior and the detailed work that goes into staging and warranting this kind of criticism (Morris, Reference Morris1988). Finally, recent research highlights the sophisticated information processing that seems to underlie blame judgments (Cushman, Reference Cushman2008; Monroe & Malle, Reference Monroe and Malle2019), at least when the community enforces standards of evidence (Malle et al., Reference Malle, Guglielmo, Voiklis and Monroe2022).
I now take a closer look at these opposing facets of blame, dividing the review into cognitive blame (judgments in the head) and social blame (acts of moral criticism).
15.3.6.1 Blame as Judgment
An Information Processing Model of Blame. A substantial literature shows that blame assigned to a person varies with a number of factors: the specific norm that was violated, the causal contributions the person makes, whether the person acted intentionally, what mental states or reasons the person had and how justified they were, and whether the person could have and should have prevented the norm violation (Cushman, Reference Cushman2008; Guglielmo & Malle, Reference Guglielmo and Malle2017; Monroe & Malle, Reference Monroe and Malle2019; Nadler & McDonnell, Reference Nadler and McDonnell2012).
My colleagues and I have aimed to integrate these findings in the Path Model of Blame (Malle et al., Reference Malle, Guglielmo and Monroe2014), depicted in Figure 15.1. The model a) specifies the various pieces of information that people normally process en route to blame; b) identifies a systematic order of processing – such that, for example, a question of intentionality comes up only when a causal contribution has been perceived (Guglielmo & Malle, Reference Guglielmo and Malle2017); and c) predicts patterns of blame updates when new information enters the processing stream (Monroe & Malle, Reference Monroe and Malle2019). A unique feature of the model is that intentionality does not merely influence blame judgments, like in other models (Alicke, Reference Alicke2000; Cushman, Reference Cushman2008), but it bifurcates the processing into distinct search paths (Monroe & Malle, Reference Monroe and Malle2017): either for the transgressor’s reasons that motivated and may justify an intentional violation; or for evidence that the transgressor could have prevented an unintentional violation.

Figure 15.1 The Path Model of Blame, adapted from Figure 2 of Malle et al. (Reference Malle, Guglielmo and Monroe2014). People form blame judgments by processing the depicted information components, typically in the four phases indicated. The fourth phase differs depending on whether the event is considered intentional (4i) or unintentional (4u).
The Path Model provides an orienting framework that allows us to examine a number of important questions about blame: to what extent it is motivationally biased, whether it is “intuitive,” and how it relates to emotions.
How Much Does Blame Suffer from Motivated Bias? All human information processing, moral judgment included, is imperfect: People have attentional lapses, forget information, or fail to integrate information under time pressure. Of greater concern in the moral psychology literature have been questions of motivated bias (Alicke, Reference Alicke2000; Ditto, Reference Ditto, Bayne and Fernández2009; Nadler & McDonnell, Reference Nadler and McDonnell2012).
Motivated moral judgment is usually described as a “desire to blame” (Alicke, Reference Alicke2000; Ames & Fiske, Reference Ames and Fiske2013), followed by information processing that strives to confirm the initial blame. However, when appropriate control conditions are implemented, people appear equally responsive to exacerbating information as to mitigating information (e.g., Monroe & Malle, Reference Monroe and Malle2019; Nadler & McDonnell, Reference Nadler and McDonnell2012, Experiment 2). Even without a general desire to blame, people might still show confirmation bias. Yet making an initial blame judgment and later revising it does not come with a confirmation bias (Monroe & Malle, Reference Monroe and Malle2019). Such a bias typically requires a more powerful source, such as the blamer’s relationship with the transgressor (Forbes & Stellar, Reference Forbes and Stellar2022), the blamer’s ideology (Niemi & Young, Reference Niemi and Young2016), or salient “extra-evidential” information (Alicke, Reference Alicke2000) – that is, information that one should not take into account when forming blame judgments.
The normative standards that determine what is extra-evidential, however, have been debated. Some scholars proposed that, even for everyday moral judgments, the standards should be set by “philosophers, legal theorists and psychologists” (Alicke, Reference Alicke2008, p. 179). Unfortunately, these experts do not necessarily agree on the pertinent normative standards, and such standards vary across cultures and historical times. So perhaps it is best to sidestep disputes over the normative standards themselves and ask, descriptively, which factors people take into account when forming blame judgments and how those factors influence blame. I will call these factors “exogenous” because they go beyond the canonical “endogenous” information components of blame, as assembled in the Path Model of Blame.
The current body of research appears consistent with the interpretation that, even when making blame judgments that some have called biased, people follow the information processing path depicted in Figure 15.1, though exogenous factors can alter steps in this path (Guglielmo, Reference Guglielmo2015). I briefly review research on two groups of exogenous factors.
As a first group, norms and values provide the standard against which people evaluate the initial violation (component 1 in Figure 15.1), which shapes the degree of downstream blame. In Sood and Darley (Reference Sood and Darley2012, Study 3), participants disapproved of a man who went to the supermarket in the nude, and they did even more so when he handed out flyers promoting a position on abortion opposite to their own. Participants’ blame (and recommended punishment) was greater presumably because they saw more or stronger norms violated. Likewise, people seem to apply different norms to intimate others than to strangers, finding strangers’ transgressions less acceptable than intimates’ (Weiss & Burgmer, Reference Weiss and Burgmer2021).
Norms also influence what people accept as justifications for a given intentional norm violation (component 4i in Figure 15.1). For example, the Southern and Northern US tend to differ in how justified they find violent acts in service of defending one’s honor after an insult (Cohen & Nisbett, Reference Cohen and Nisbett1994); hence these groups will differ considerably in how much they blame such violent acts.
Finally, people’s norms guide the obligations people impose on a person to prevent certain unintentional violations. For example, people with hierarchy-legitimizing ideologies believe that someone with implicit racial bias is not obligated to prevent their own discriminatory acts that follow from this bias; as a result of this normative belief about prevention, they blame a person less who unintentionally committed such discrimination (Redford & Ratliff, Reference Redford and Ratliff2016). Conversely, in liberal communities, people are obligated to avert microaggressions (Princing, Reference Princing2019) whereas others criticize this obligation as “ridiculous” (Dickey, Reference Dickey2019). In all these cases, blame varies considerably across groups of people, not because of motivated bias but because they interpret the same events in light of different norms.
Beliefs and attitudes toward the transgressor constitute the second group of exogenous factors. Several studies have shown that judgments of causality or preventability can be affected by negative character information (Alicke & Zell, Reference Alicke and Zell2009) or the perceiver’s attitude (Niemi & Young, Reference Niemi and Young2016). Moreover, character, past behavior, and stereotypes can affect people’s inferences of a transgressor’s intentionality or reasons for acting (Sood, Reference Sood2019). So while there is good evidence that exogenous beliefs and attitudes can insert themselves into the blame processing path, some authors have proposed that these exogenous factors directly influence blame (Alicke & Zell, Reference Alicke and Zell2009; Ciurria, Reference Ciurria2019, chapter 4), effectively bypassing the endogenous blame processing. The current evidence, however, does not provide convincing support for such direct effects (Guglielmo, Reference Guglielmo2015). Exogenous and endogenous factors would have to be properly compared in one and the same study, but they are rarely jointly measured (Dukes & Gaither, Reference Dukes and Gaither2017; Nadler & McDonnell, Reference Nadler and McDonnell2012); or, if they are measured, not all are included in model tests (e.g., Alicke & Zell, Reference Alicke and Zell2009; Nadler, Reference Nadler2012); or, if included, they do not yield the predicted direct effects (Mazzocco et al., Reference Mazzocco, Alicke and Davis2004).
There is no doubt that people sometimes blame unfairly, to protect their self-esteem, enact their prejudice, or protect their group. Extant evidence, however, suggests that even in these cases people rely on the same information components – causality, intentionality, reasons, and so on – on which they rely when blaming fairly.
Is Blame Intuitive? When scholars characterize moral judgments as intuitions, they rarely specify the type of moral judgment they are referring to. Some describe moral intuitions as feelings of “approval or disapproval” (Haidt, Reference Haidt2001, p. 818) or as some action being “bad” (Clark & Winegard, Reference Clark and Winegard2019, p. 14), which we may call evaluations (Malle, Reference Malle2021). To be moral evaluations, these intuitions rely on the relevant moral norms being activated in the observer, who then judges the relevant behavior as violating the norm. Evaluations incorporate explanations for simple causality (De Freitas & Alvarez, Reference De Freitas and Alvarez2018) and potentially about the intentionality at least of visibly performed behaviors (Decety & Cacioppo, Reference Decety and Cacioppo2012). By contrast, blame judgments also incorporate information about the intentionality of unobserved actions, the agent’s specific, potentially justifying reasons, and the complex assessment of preventability obligations and capacities (Monroe & Malle, Reference Monroe and Malle2019) – the kind of processing that typically does not fall under “intuitions.”
Suggestions that moral judgments are intuitions typically go hand in hand with the claim that people are “dumbfounded” when trying to justify their moral judgments (Haidt & Bjorklund, Reference Haidt and Bjorklund2008, p. 197). This claim faces a number of challenges. First, researchers routinely ask participants for simple wrongness judgments, which rely primarily on citing the relevant violated norm (Malle, Reference Malle2021). Proponents of the dumbfounding hypothesis treat a norm statement (e.g., “because it’s incest”) as a dumbfounding response, whereas others insist that a norm statement constitutes an actual justifying reason (Stanley et al., Reference Stanley, Yin and Sinnott-Armstrong2019). Second, proponents of the dumbfounding hypothesis instruct their experimenters to explicitly “undermine whatever reason the participant put forth in support of his/her judgment or action” (Haidt et al., Reference Haidt, Bjorklund and Murphy2000, p. 7), which arguably biases results in favor of the hypothesis (Gray et al., Reference Gray, Schein and Ward2014; Royzman et al., Reference Royzman, Kim and Leeman2015). Despite these favorable conditions, surprisingly few people give dumbfounding responses (e.g., 32 percent in McHugh et al., Reference McHugh, McGann, Igou and Kinsella2017; see Malle, Reference Malle2021, p. 309); and in less coerced situations, dumbfounding drops below 20 percent (McHugh et al., Reference McHugh, McGann, Igou and Kinsella2020). These weak dumbfounding rates can hardly support the claim that moral judgments are by nature “intuitive.”
Is there any dumbfounding evidence for blame judgments? Quite the contrary. Bucciarelli et al. (Reference Bucciarelli, Khemlani and Johnson-Laird2008, Study 3) showed that people had no trouble explicating their blame judgments in a think-aloud protocol, and Voiklis et al. (Reference Voiklis, Kim, Cusimano and Malle2016) documented that people provide rich and systematic explanations of their blame judgments. Importantly, these explanations referred to just the types of canonical information known to cause variations in blame judgments (e.g., the seriousness of the norm violation, intentionality, justified reasons).
How Does Blame Relate to Emotion? Determining the role of emotion in blame requires heeding distinctions between affect, evaluations, and emotions. Affect is often understood as a nonrepresentational valenced feeling state (Neumann et al., Reference Neumann, Seibt and Strack2001); evaluations are rapid valenced appraisals of some object (“this is bad”); and emotions are a large class of states (e.g., anger, resentment, or disgust) that are differentiated by appraisals – the cognitive processing of rich arrays of information (Scherer, Reference Scherer2013). Further, we must separate the different roles that these three phenomena can play in blame: They could fully constitute blame judgments, cause them, accompany (and possibly amplify) them, or be caused by them (Monin et al., Reference Monin, Pizarro and Beer2007; Strohminger, Reference Strohminger, Price and Walle2017).
Affect cannot constitute or cause blame because it lacks an object; but it can accompany blame. Evaluations of norm violating events lie at the beginning of any processing stream toward blame, so they make a causal contribution, but they do not process much of the information that blame judgments are normally based on and are therefore insufficient to fully cause, let alone constitute blame. Emotions such as anger rely on a number of appraisals (Ellsworth & Scherer, Reference Ellsworth, Scherer, Davidson, Scherer and Goldsmith2003) of just the kind of information that blame judgments respond to (e.g., causality, intentionality), so the two may co-emerge from this information processing. However, one can arrive at blame judgments by processing the relevant information without feeling angry, whereas it would be difficult to become angry before processing this kind of information (e.g., causality, intentionality). Emotions thus do not seem to constitute moral judgments but may accompany and amplify them.
Researchers have attempted to induce anger to test whether it indeed amplifies blame or related moral judgments. Although some studies showed such impact (Ask & Pina, Reference Ask and Pina2011; Lerner et al., Reference Lerner, Goldberg and Tetlock1998; Seidel & Prinz, Reference Seidel and Prinz2013), others did not (Gamez-Djokic & Molden, Reference Gamez-Djokic and Molden2016; Gawronski et al., Reference Gawronski, Conway, Armstrong, Friesdorf and Hütter2018). Conversely, judgments of blame and responsibility can alter emotions or mediate between norm violations and emotions (Quigley & Tedeschi, Reference Quigley and Tedeschi1996; Zajenkowska et al., Reference Zajenkowska, Prusik, Jasielska and Szulawski2021). Studies also found expressions of anger and moral outrage to arise after judgments of moral violations (Sasse et al., Reference Sasse, Halmburger and Baumert2020), and response times of expressing anger seem to be no faster or even slower than response times to blame judgments (Cusimano et al., Reference Cusimano, Thapa, Malle, Gunzelmann, Howes, Tenbrink and Davelaar2017).
We can conclude that when people encounter a norm violation, they are likely to morally evaluate it, they may or may not experience feelings along with it, and they routinely process a canonical set of information. This information processing guides blame judgments and may generate emotions, but which specific emotions arise (e.g., anger, resentment, indignation) will depend on the specific processed information. These emotions then help scale the intensity of a socially communicated blame judgment (Drew, Reference Drew1998). And this brings us to the phenomenon of blame as communicated moral criticism.
15.3.6.2 Blame as Moral Criticism
Social acts of blame, reproach, and rebuke serve several goals. They draw attention to a transgression (McGeer, Reference McGeer, Coates and Tognazzini2012), signal the blamer’s commitment to a norm system (Shoemaker & Vargas, Reference Shoemaker and Vargas2021) and try to enforce it (Bell, Reference Bell, Coates and Tognazzini2012); they try to change the transgressor’s ongoing or future behavior (Miller, Reference Miller2003; Przepiorka & Berger, Reference Przepiorka and Berger2016), but they also simply express moral judgments and moral emotions (Sorial, Reference Sorial, Abell, Smith, Abell and Smith2016). Acts of blaming are costly, however, for all involved parties: the transgressor, the moral critic, and the community (Malle et al., Reference Malle, Guglielmo, Voiklis and Monroe2022). The transgressor suffers a loss of public standing, damaged relationships, and just plain bad feelings. The moral critic faces potential retaliation from the transgressor (Balafoutas & Nikiforakis, Reference Balafoutas and Nikiforakis2012), damaged relationships, and possibly lack of support from the community. And the community itself carries the costs of escalating community strife (Allen, Reference Allen2002).
To keep these costs in check, most social communities impose norms on moral criticism (Coates & Tognazzini, Reference Coates, Tognazzini, Coates and Tognazzini2012; Eriksson et al., Reference Eriksson, Andersson and Strimling2017; Voiklis & Malle, Reference Voiklis, Malle, Gray and Graham2018), which regulate the standards of evidence and discourse of blame (Friedman, Reference Friedman2013; Malle et al., Reference Malle, Guglielmo, Voiklis and Monroe2022). When moral critics comply with these norms, they are also more likely to achieve the goals of their criticism.
Complying with norms of blaming means that moral critics will process the available evidence and assign blame proportionally to the observed transgression. If they ignore evidence, they are bound to blame out of proportion – they will either overblame or underblame. Overblaming is itself a norm violation and will provoke rejection or retaliation from the transgressor and may cause long-term damage to a relationship (Fincham et al., Reference Fincham, Beach and Nelson1987). Underblaming may downplay the violated norm and thus fail to change the offender’s behavior. However, mild, subtle forms of criticism for everyday transgressions (e.g., blatant littering; an able-bodied person using a disability parking spot) can be effective ways of changing behavior. In most social communities, even a cold stare or a snide remark may successfully communicate moral criticism (Miller, Reference Miller2003; Molho et al., Reference Molho, Tybur, Van Lange and Balliet2020). The critic thus affirms the offender as a respectable member of the moral community who is nonetheless worth criticizing (Bennett, Reference Bennett2002).
The transgressor can of course influence the success of moral criticism. Blame demands a response (Drew, Reference Drew1998; McGeer, Reference McGeer, Coates and Tognazzini2012; Shoemaker, Reference Shoemaker, Coates and Tognazzini2012), and transgressors have many options – on the one hand, they can deny, dismiss, or retaliate (Dersley & Wootton, Reference Dersley and Wootton2000); on the other hand, they can explain, apologize, and offer repair (Walker, Reference Walker2006; Watanabe & Laurent, 2020). Without such reconciling responses, the regulation of social relationships is bound to fail (Laforest, Reference Laforest2002).
The community, finally, upholds the norms of blaming by demanding warrant from those who blame, especially when they launch premature, biased, or inaccurate criticism (Voiklis & Malle, Reference Voiklis, Malle, Gray and Graham2018). By demanding such warrant, the community protects their members from unfair criticism. But it can also go too far in putting pressure on moral critics – for example, by suppressing moral criticism of those in power (Ciurria, Reference Ciurria2019, chapter 9).
Forces That Alter Costs and Success Conditions. The costs of moral criticism vary with a number of factors that have implications for the norms and success conditions of the criticism (Malle et al., Reference Malle, Guglielmo, Voiklis and Monroe2022). For example, costs vary with the addressee of the criticism. Second-person blame (expressed directly to the transgressor) comes with the risk of retaliation and fails more easily when overblaming occurs, but when communicated appropriately it can be immediately effective in regulating behavior. Third-person blaming (expressed to other parties) is safer and effective in affirming the community’s norm system, but any change in the transgressor’s behavior can be achieved only indirectly and with likely delay.
Another influential variable that alters costs is the in-group or out-group status of the transgressor (Malle et al., Reference Malle, Guglielmo, Voiklis and Monroe2022). When blaming in-group members, moral critics face pressure to heed the community norms that stem against disproportionate blaming and strife. But when the transgressor is outside the community, these norms lose their force. Critics may get away with sloppy information processing and unfair accusations, because the costs have been shifted almost entirely from critic and community to the transgressor. In the punishment literature, when an out-group member transgresses against an in-group member, sanctions are greatest – from mild monetary penalties (Bernhard et al., Reference Bernhard, Fischbacher and Fehr2006) to death by lynching (Equal Justice Initiative, 2017) or state execution (Eberhardt et al., Reference Eberhardt, Davies, Purdie-Vaughns and Johnson2006). And initial results suggest that blame, too, is less regulated when applied to out-group members (Monroe & Malle, Reference Monroe and Malle2019).
A recently emerging and powerful factor that alters costs is online discourse, especially on social media, which can deliver public shaming or social exclusion in just a few words (Klonick, Reference Klonick2015). Online moral critics bear fewer costs (Crockett, Reference Crockett2017), most obviously when they are anonymous. Even when identifiable, they launch their blame from the safe distance of a keyboard, which restricts the other’s retaliation and poses few risks to valued relationships; after all, those who are blamed, reviled, and excluded are typically strangers and out-group members. Online blaming also imposes fewer costs on the community because that community is indeterminate. Social bonds that normally tie community members to each other are far weaker online, so people are less committed to norms (including norms of blaming) and put fewer constraints on derogations (Márquez-Reiter & Haugh, Reference Márquez-Reiter and Haugh2019).
The looser restrictions on online moral criticism, however, can empower critics who have previously been silenced. The MeToo movement, for example, enabled acts of calling out celebrities or companies who transgressed, or documenting microaggressions and outright discrimination. Ideological camps differ in their reception of such amplified criticism – which some call “cancel culture” (Republican National Committee, 2020). Criticism of hurtful microaggressions is seen as a warranted rebuke in one community but as a ridiculous demand of political correctness in another community (Dickey, Reference Dickey2019). Such polarized reception stems from community differences in the norms for the original actions (e.g., political correctness, microaggression) as well as in the norms that govern criticism itself – including who is granted credibility and standing to criticize and how. Ultimately, online communities will have to calibrate what is a proportionate response to violations. We see the difficulty of such calibration in the following case: A critic used Twitter to blame two offenders for using the word dongle in what was interpreted to be a sexually offensive way, which then caused one of the offenders to lose his job after the publicity, but then the Twitter-posting critic was subsequently vilified and threatened by numerous people and fired from her own job for the act of public blaming (Brown, Reference Brown2013).
15.4 Conclusions
I have tried to make plausible the hypothesis that the distinct cultural histories of blame and punishment have shaped their distinct psychological underpinnings and social expressions. Today’s blame arose long ago as the primary sanctioning behavior in small, tight-knit, egalitarian communities, where moral criticism is often mild but effective, built on the power of reputation and the need to belong. Because of the costs of damaged relationships and the potential for retaliation, such blame is regulated by a set of norms that demand evidence and fair treatment – at least for members of one’s own community. Moral critics are expected to communicate their disapproval and to invite a response from the transgressor, with the goal of repairing and continuing the relationship and averting a repeat transgression. Such blame can be contested and mitigated by the transgressor’s response, and the critic may even admit a mistake and take back the criticism. This ideal can be missed in numerous ways, social media being a salient contemporary case.
Punishment arose in larger, more distant, hierarchical communities. Acts of punishment would normally themselves be serious norm violations, were it not for the legitimacy granted to, or taken by, the punisher. Frequently from high up in the hierarchy, the punisher – a person or an institution – coercively imposes costs on the transgressor that cannot be taken back. Some forms of punishment step in where moral criticism has failed, but punishment still often stops communication rather than reinstating it or damages relationships rather than repairing them.
One major shortcoming of current institutional punishment is its ineffectiveness in deterring crime and reforming individuals who committed crimes (Yukhnenko et al., Reference Yukhnenko, Sridhar and Fazel2020). As one alternative to the standard punitive system, restorative justice is an attempt to reconcile perpetrator and crime victim through a mediated conversation in which all thoughts and feelings are expressed, often restitution is agreed on, and the matter laid to rest (Braithwaite, Reference Braithwaite1999). Evidence shows that recidivism decreases considerably when restorative justice replaces legal punishment (Kennedy et al., Reference Kennedy, Tuliao, Flower, Tibbs and McChargue2019; Kuo et al., Reference Kuo, Longmire and Cuvelier2010). Restorative justice procedures are noticeably close to ancient and modern processes of moral criticism and reconciliation. A hope for the future is therefore that institutions more widely adopt constructive moral criticism as the powerful form of norm enforcement that it once was.
The title of this chapter contains two substantial and contestable terms. Each highlights a central feature of human psychology and behavior: first, that we understand and operate in accord with a set of norms that are distinctively moralized; and, second, that we communicate with one another – using language, of course, but also nonlinguistically. But how do these features of human psychology and behavior interact? Our aim in this chapter is to explore whether there is anything distinctive about moral communication over and above the fact that such communication is specifically focused on moralized norms and behavior.
The chapter is divided into two parts. In Section 16.1, we sketch the wider background to our topic, focusing in Section 16.1.1 on people’s proclivity for engaging in norm-governed behavior of various different kinds, and in Section 16.1.2 on the unique features of human communication that exploit and reinforce people’s capacity for norm-governed behavior. In Section 16.1.3, we consider what distinguishes moralized norms from other kinds of norms, both conventional and nonconventional (e.g., norms of rationality). We suggest that a satisfying answer to this question must emphasize the distinctive set of blaming emotions or “reactive attitudes” that people experience toward transgressors of moralized norms, inclining them to punish (rather than simply correct) transgressors (Strawson, Reference Strawson1962). We conclude the first part of the chapter with the suggestion that punishment in this context is best understood as a specialized form of moral communication, with various potential messages directed toward transgressors themselves but also toward others in the community.
In Section 16.2, we explore the advantages of regarding human punishment as a communicative reaction to violated moralized norms. This perspective is not mainstream in psychological research on punishment mechanisms, which is more traditionally focused on the motivations of retribution and deterrence. In 16.2.1, we review the history of this research, explaining why many theorists are drawn to the view that the psychology of punishment is deeply retributive with moral payback as its raison d’être. Against this conclusion, we argue in Sections 16.2.2 and 16.2.3 that empirical findings are more deeply supportive of the communicative view. In Section 16.2.4, we consider a range of empirical questions that the communicative view opens up. We conclude with some general theoretical reflections on the interdisciplinary advantages of regarding punishment as a distinctive and potentially modifiable form of moral communication.
16.1 Setting Moral Communication in Context: Psychological Building Blocks
16.1.1 A Norm-Governed Form of Life
Human beings are distinctively normative creatures, investing in attitudes and behavior that are regulated, and so shaped, by a wide variety of socially developed and endorsed norms. By “norms,” we mean rules, principles, and standards (whether formal or informal) that govern what counts as appropriate or proper behavior in the various activities in which humans engage. This covers a lot of ground, much of it outside the moral domain. As many cognitive theorists are now beginning to stress, humans are a thoroughly enculturated species, developing many unique and sophisticated cognitive/behavioral capacities via the internalization of norm-governed social practices that are developed and passed down through successive generations. (For recent representative discussion, see Hutchins, Reference Hutchins2014; McGeer, Reference McGeer2020; Sterelny, Reference Sterelny2012; see also Chapter 4, this volume.)
To give just a taste of such normative enculturation: we learn to speak, read, and write a public language; we learn to enumerate and calculate an endless variety of things using a standardized number system; we learn to regulate our beliefs and desires in accord with communally endorsed evidential, practical, and inferential “rules of reason”; we learn to organize our days in accord with time- and calendar-keeping practices; we learn to recognize and manipulate a broad range of task-specific tools; we learn to greet each other, stand at an appropriate social distance from one another, manage turn-taking with one another in a huge variety of contexts, and so on in staggering detail. In short, we learn to converse together, work together, play together, eat together, and in general navigate complex built environments in relatively smooth, expectable, and cooperative ways by internalizing a remarkable variety of explicit and implicit norms and meta-norms (i.e., norms for resolving normative disputes, or norms for setting new norms, etc.).
Such norms may be shared across cultures or be deeply culturally inflected; they may be specific to particular tasks and/or types of agents occupying variously defined social roles; they may be formally codified or informally and implicitly adopted; they may be policed with rigor or with relative nonchalance, inviting greater plasticity and individual maneuverability; they may be central to our sense of social identity or relatively peripheral to that identity. But regardless of their variation along these different dimensions, norms are, for human beings, as ubiquitous and necessary as the air we breathe. We only become cognitively sophisticated, mutually recognizable and reliable agents, able to engage in complex and coordinated activities with one another, because of our unique capacity to generate and internalize the rich and intricate set of norm-governed practices that pervade our existence.
This richly elaborated norm-governed form of life implicates a distinctive kind of psychology. Human beings are not only sensitive to the presence of norms, acquiring them readily via environmental cues and feedback from others, they are motivated to comply with them and police transgressions against them in a variety of ways (e.g., by way of simply pointing out “that’s not how it’s done”).
A key indication of this sensitivity to norms can be found in the special qualities of human imitation. For instance, unlike other social animals (including especially other primates), children from the age of 24 months begin to engage in a phenomenon called “over-imitation”: faithfully copying others’ behavior even when that behavior is instrumentally unnecessary for achieving a goal. For instance, children continue a learned routine of tapping on the top of a transparent box before pulling a lever that releases a treat, even though it is visually manifest that pulling the lever is sufficient for getting the treat (Horner & Whiten, Reference Horner and Whiten2005).
While there is more flexibility in this early emerging behavior than originally supposed, research now shows that so-called over-imitation is significantly driven by imitators interpreting the modeled action sequence in explicitly normative terms: That is, imitators see the model as demonstrating the “right” or “proper” way to engage in the activity, regardless of the instrumental value of the actions undertaken (Hoehl et al., Reference Hoehl, Keupp, Schleihauf, McGuigan, Buttelmann and Whiten2019; Keupp et al., Reference Keupp, Behne and Rakoczy2013).Footnote 1 Indeed, children seem biased toward a normative interpretation of others’ behavior (at least in many contexts, especially involving adult–child interactions), indicated not only by their own proclivity for over-imitation but also, more significantly, by the spontaneous protests and criticisms they make against those who engage in an activity without faithfully replicating what the model has done (“not like that!”) and by their own instructive modeling of the activity to others (Kenward, Reference Kenward2012).
Somewhat older children (4–5 years of age) will take account of moral costs in their imitative behavior. For instance, in contexts that would normally elicit normatively driven “over-imitation,” they are reluctant to reproduce a model’s actions, and object to others who do, if such actions entail the destruction of a valuable object belonging to the experimenter (Keupp et al., Reference Keupp, Bancken, Schillmöller, Rakoczy and Behne2016). These children show by their spontaneous protests and expression of moral concern (“she will be very sad if you do that”) that not all normative considerations are treated equally. This jibes with findings by Turiel and others that preschoolers are sensitive to a “moral/conventional” distinction, with this sensitivity emerging from about the age of 3 (Nucci, Reference Nucci2001; Smetana, Reference Smetana and Bennett1993; Turiel, Reference Turiel1983). Since children begin to engage in normatively driven “over-imitation” somewhat earlier (by the age of 2), this may indicate a more general sense of “normativity” that gets refined with experience and/or the development of linguistic competence. Findings suggest that language comprehension is related to an earlier emergence of this distinction (Smetana & Braeges, Reference Smetana and Braeges1990), possibly indicating that sensitivity to different kinds of norms is communicatively bootstrapped. Other lines of research highlight the importance of communication in supporting children’s development of myriad cognitive, social, and emotion-regulation capacities (Király et al, Reference Király, Csibra and Gergely2013).
16.1.2 Communicative Interaction of a Distinctively Human Kind Facilitates Normative Enculturation
Our readiness to detect norm-governed behavior in those around us, to engage in such behavior ourselves, and to police others when they fail to do likewise is a deep-seated feature of human psychology. But such normative proclivities are not an inflexible response to the world; rather, they constitute a predisposition or bias that can be triggered or inhibited by a variety of contextual cues, some of which may be discerned simply by observing the activities of others (Hoehl et al., Reference Hoehl, Keupp, Schleihauf, McGuigan, Buttelmann and Whiten2019). However, people generally do not rely on subtle observational cues to discern whether others’ behavior is norm-governed or not. Real-world interactions are typically rife with communicative cues, both linguistic and nonlinguistic, that explicitly mark how we are supposed to interpret what others are doing. And children are sensitive to such cues even from the age of 14 months, understanding the pedagogic intent behind a model’s behavior and revealing in their imitative responses a readiness to learn just what the model communicatively intends that they should (Gergely & Király, Reference Gergely, Király, Dukes and Clément2019; Király et al., Reference Király, Csibra and Gergely2013).
This turns our attention to the special qualities of human communication that make it particularly apt for facilitating normative enculturation. While the use of language itself is a uniquely human phenomenon, theorists have lately drawn attention to more fundamental features of human communication that distinguish it from the communicative practices of other animals. These are not obvious. For instance, in addition to vocalizing and displaying a range of relatively inflexible signaling behaviors in response to environmental cues, chimpanzees regularly gesture to one another to solicit nursing, food, or grooming in ways that are learned, flexible, and even idiosyncratic (Goodall, Reference Goodall1986). Such gestures are intentional acts of communication (Grice, Reference Grice1969; Moore, Reference Moore2016): They are directed toward particular others when, and only when, those others are attending and are appropriate targets of a particular solicitation; a given gesture may also be modified to suit the situation (Kaminski et al., Reference Kaminski, Call and Tomasello2004; Tomasello & Call, Reference Tomasello and Call2019; Yamamoto et al., Reference Yamamoto, Humle and Tanaka2012).
Still, while these are sophisticated communicative acts that depend on certain mentalizing skills (for instance, monitoring attention and likely even understanding what others can and cannot see), chimpanzees and other great apes will only use gestures in a proto-imperative way: to get others to do specific concrete things as a way of realizing their own goals.Footnote 2 They do not use gestures – in particular, the pointing gesture – in a proto-declarative way: to show others something or simply to share attention for the pleasure of it (Tomasello & Call, Reference Tomasello and Call2019).
This makes for a puzzling difference in the communicative behavior of great apes in contrast with humans, even at the nonlinguistic gestural level. In human interactions, the pointing gesture is naturally used and generally understood despite serving a variety of functions – proto-imperative (requesting something), proto-declarative (showing something), proto-informative (e.g., indicating the location of a hidden object), and proto-interrogative (e.g., soliciting more information about an indicated object). Such multifunctional use and understanding emerges in children as young as 12–14 months (Southgate et al., Reference Southgate, Van Maanen and Csibra2007; Tomasello et al., Reference Tomasello, Carpenter and Liszkowski2007) – and this despite the fact that the (prelinguistic) mentalizing skills of children at this stage of development are, in many respects, not hugely more sophisticated than those of chimpanzees and bonobos. So why this difference?
Drawing now on decades of comparative primate research, Tomasello and colleagues argue that a key determinant of this difference lies in one “small” cognitive and motivational factor that ramifies in powerful ways (Tomasello et al., Reference Tomasello, Carpenter, Call, Behne and Moll2005). Humans are distinctive insofar as their social activities are permeated by an irreducibly collective (shared, or “we”) intentionality (Gilbert, Reference Gilbert1994; Rakoczy & Tomasello, Reference Rakoczy, Tomasello and Tsohatzidis2007). When people interact with one another, they are not simply solo operators, aiming to engineer another’s behavior simply to forward their singular aims. Rather, humans often take themselves to be doing things together, naturally forming shared intentions to act in coordinated and mutually supportive ways in the context of joint activities: we are playing a game together; we are finding a hidden object together; we are looking at something together. And from here it seems a short step to structuring and understanding such activities in explicitly normative terms. Hence, the capacity for shared or collective intentionality, emerging early in development and potentially unique to the human species, undoubtedly plays a foundational role in explaining people’s distinctive proclivity to adopt norm-governed attitudes and behavior.
The capacity for shared intentionality thus heralds a fundamental change in the nature of human communication. The transformation is from an activity that is essentially behavior-focused and unidirectional (seen, for instance, in great apes) to an activity that is essentially attitude-focused and dialogical. Whether people communicate in language or in a variety of nonlinguistic ways (using ostensive eye contact, emotional expressions, prosody, bodily gestures and/or other symbolic acts), they invariably regard one another as potential candidates for shared intentionality, presupposing and reinforcing the joint activities and/or shared frames of reference that give meaning to their particular communicative acts (see, e.g., Tomasello et al., Reference Tomasello, Carpenter, Call, Behne and Moll2005). Hence, in communicating with others, people are not just aiming to shape or regulate others’ behavioral responses as a way of realizing their individual goals. People are aiming to make others understand the point of their communicative behavior – whether it be asking for something, sharing attention, sharing pertinent information, committing to some course of action, and so on in wide variety. Hence, in communicating with others, people invariably solicit a suitable communicative response from their interlocutors in turn – a reaction that makes sense in light of the shared understanding and collaborative support people aim to achieve. This is the sense in which human communication is essentially dialogical.
In the context of shared norm-governed activities, we see a special instance of this insofar as people’s communicative efforts are naturally bent toward signifying, in both word and deed, that there are proper ways of behaving that they expect others to understand and internalize. This message may be conveyed ex ante by instruction and modeling or ex post by protest and correction. But either way, people’s aim is not simply to elicit a suitable behavior response from the other. Qua communicators, people are dialogically focused on soliciting an appropriate attitudinal response – namely, others’ appreciation and acceptance that “this is the way we do things around here.”
16.1.3 Moralized Norms and Communication
Against this rich background of communicatively supported normative enculturation, we now pose the following two questions: What is distinctive about the norms people identify as “moral” – and/or attitudes and behavior governed by such norms? And is there anything distinctive about communication regarding such norms?
As a preliminary matter, we note there is a persisting tendency within the moral-psychological literature to treat norms as essentially dividing into two types, “moral” and “conventional” (Nucci & Nucci, Reference Nucci and Nucci1982; Smetana & Braeges, Reference Smetana and Braeges1990; Turiel, Reference Turiel1983). But, as the foregoing discussion makes clear, people are variously invested in a rich variety of norms as a matter of social and practical identity, and many of these norms defy easy binary categorization: Either they have elements of both the “moral” and the “conventional” (such as norms of “fair play”), or they belong to a category of their own (such as “norms of rationality”). This naturally complicates the search for criteria by which moralized norms can generally be distinguished from the rest. But, as our discussion in this section will show, it also places new emphasis on the nature of our communicative interactions in tracking this distinction.
In their quest to elucidate the distinctive nature of moralized norms, theorists have generally pursued two main lines of inquiry. The first considers the substantive content of the norms themselves. The second considers the psychological attitudes people take toward such norms, signaling a distinctive kind of normative commitment taken toward the content in question, whatever it may be. We discuss each of these approaches in turn, noting that a satisfying answer to the question may well appeal in some measure to both content and attitude.
With regard to the substantive content of moralized norms, theorists generally agree that this is something of a moving target. In any given society, ordinary people may make an intuitive distinction between moral norms and other kinds of rules, principles, conventions, or standards; but there are invariably borderline areas of disagreement. This problem is only compounded across different cultures; some communities operate with a relatively expanded conception of moral norms, others with a relatively contracted conception (e.g., as regards matters of food, clothing, or sexual preference). Yet even amid this variability, certain “themes” are discernible in the norms generally identified as moral. For instance, as Sripada and Stich (Reference Sripada, Stich, Carruthers, Laurence and Stich2006, p. 283) observe, most societies have “rules that prohibit killing, physical assault, and incest”; “rules promoting sharing, reciprocating, and helping”; “rules regulating sexual behavior among various members of society”; and “at least some rules that promote egalitarianism and social equality.”
Is there any underlying unity discernible in this motley collection of moralized norms? This is a matter of ongoing debate. At one extreme, reductive monists aim to show that all moralized norms are grounded in a dominant substantive concern. Accounts differ as to the nature of this concern: For instance, some argue that moralized norms are all concerned with prohibiting “harm” (Gray et al., Reference Gray, Waytz and Young2012); others, that they are concerned with prohibiting “unfairness/injustice” (Baumard et al., Reference Baumard, André and Sperber2013; Piazza et al., Reference Piazza, Sousa, Rottman and Syropoulos2019). But this has been a challenging position to defend, especially in a cross-cultural context (see for instance Berniūnas et al., Reference Berniūnas, Dranseika and Sousa2016).
A less extreme form of monism discerns some unity among moralized norms at a higher level of generality – for example, they are specifically concerned with placing a check on selfish goals or interests where such interests may threaten those of the community (Fehr & Fischbacher, Reference Fehr and Fischbacher2004). Undoubtedly this is true of many such norms; but, again, others seem to escape this characterization – at least in any direct sense. For instance, it is hard to see how moralized norms regarding treatment of the dead or eating certain kinds of food are specifically geared toward checking selfish goals or interests, though they may serve certain communal goals or interests (such as stabilizing a sense of group identity). Yet not all moralized norms are apt for serving communal goals or interests either, at least on a broad scale: For instance, many societies endorse “agent-relative” norms that permit (or even obligate) individuals to prioritize the needs of close relations over the community itself or other members within it, at least to some degree.
In light of these difficulties, other theorists take an avowedly pluralist approach to the content of moralized norms, arguing in addition that there is significant cross-cultural variation in the kind of “disinterested” concerns individuals are required to prioritize in their attitudes and behavior. Such moralized norms might prioritize a range of concerns for particular others (families, friends, strangers in need…), a range of concerns for maintaining the community as a whole (including its structure and operations), and even a range of concerns that relate to respecting or implementing some transcendentally determined cosmic order (such as “God’s will” or “Dhamma”) (McGeer, Reference McGeer and Sinnott-Armstrong2008; Shweder et al., Reference Shweder, Much, Mahapatra, Park, Brandt and Rozin1997). This pluralistic approach leads naturally to a different sort of question – namely, why human beings should have such a range of disinterested concerns, however much these require development or stabilization via communally endorsed norms. This is a challenging explanatory project in itself, perhaps explaining the continuing allure of monism for many theorists. It may also recommend taking a different approach to characterizing the distinctiveness of moral norms altogether – one that focuses less on their content and more on the psychological attitudes people take toward them (Nucci, Reference Nucci2001; Skitka et al., Reference Skitka, Hanson, Morgan and Wisneski2021).
In keeping with this second approach, there are a range of attitudes that theorists have identified as seemingly peculiar to moralized norms. Following Sripada and Stich (Reference Sripada, Stich, Carruthers, Laurence and Stich2006), we group these attitudes under three headings: motivation for compliance, the metaphysical status attributed to the norms themselves, and reaction to transgressions (cf. Skitka et al., Reference Skitka, Hanson, Morgan and Wisneski2021). To give Sripada and Stich’s summary characterization of the attitudes under each heading: 1) people find moralized norms “intrinsically motivating”; 2) they view them as having “independent normativity”; and 3) they react to their violation in a distinctively punitive way, characteristically involving a range of equally distinctive “reactive” emotions (condemnatory anger, resentment, indignation, guilt, shame, and remorse – depending, of course, on whether the transgressor is other or self) (Strawson, Reference Strawson1962). We discuss each of these features in turn.
To say that people find moralized norms (1) “intrinsically motivating” is to say they experience norm compliance as an end in itself. They do not comply with such norms simply for instrumental reasons – because of other benefits they may gain through norm compliance (e.g., reputational benefits, reciprocity benefits, realization of cooperation-dependent goals) – or, indeed, to avoid certain costs noncompliance may bring (e.g., retaliation, cooperation-dependent goal frustration). As Sripada and Stich (Reference Sripada, Stich, Carruthers, Laurence and Stich2006) emphasize, compliance is its own reward, other potential benefits notwithstanding.
The problem is that this psychological feature does not seem distinctively related to moralized norms. As our earlier discussion of normative enculturation suggests, humans are heavily dependent on developing normatively shaped attitudes and behavior of a more general kind, enabling a huge and sophisticated array of distinctively human capacities (including, in particular, linguistic capacities, but many others as well). Such deep enculturation would surely be impossible without some psychological machinery innately geared toward norm detection and compliance, where people regulate their behavior, not just in imitation of what others do, but in light of their explicit pedagogical and corrective feedback and where discovering the “right way” to do something is rewarding in itself. This, we think, is amply demonstrated in humans’ species-specific proclivity for over-imitation, emerging at least as early, if not prior to, a sensitivity to specifically moralized norms.
Next we consider (2) “independent normativity,” the idea that people view moralized norms as specifying what they ought to do “independently of any legal or social institution or authority.” Such norms are viewed as having the metaphysical status of “objective” truths, implying generality of scope across different times and places (Kelly & Stich, Reference Kelly, Stich, Carruthers, Laurence and Stich2007; Nucci & Nucci, Reference Nucci and Nucci1982; Skitka et al., Reference Skitka, Hanson, Morgan and Wisneski2021; Smetana, Reference Smetana and Bennett1993; Turiel, Reference Turiel1983).Footnote 3 It has been suggested that people are intrinsically motivated to comply with moralized norms because they view them as having independent normativity.Footnote 4 Be that as it may, we think these two phenomena are psychologically separable: People can be intrinsically motivated to comply with norms they do not regard as having independent normativity. Still, the question remains: Is independent normativity/objective validity really a distinctive feature of moralized norms? We think not. Rational norms – for instance, valid rules of inference – are generally held to have “independent normativity,” but it would be strange to classify them as moral norms (see Southwood, Reference Southwood2011 for a similar observation).
Finally, we come to the third psychological feature associated with moralized norms: the very strong disposition (3) to react to norm violations of this kind in a distinctively punitive way. Here there is some indication of a genuine difference in people’s attitudes to different sorts of norms. For while all norm violations attract notice, how individuals respond to such violations can vary widely depending on a range of factors, including how they are related to the transgressor, the nature of the norm (moral or nonmoral), the transgressor’s degree of normative understanding, the transgressor’s intention in violating the norm, the consequences of violating the norm, and so on. Such factors may influence whether any norm-enforcing action is called for at all (some norm violations simply call for understanding, or even appreciation, of the novelty so displayed). But even when the transgression calls for some norm-enforcing response, this too can take a variety of material and/or psychological forms, ranging from gently corrective to strongly sanctioning – what some might call “punitive.” However, this use of the term neglects an important dimension of norm-enforcement that is only incidentally related to the severity of sanction imposed – namely, what the sanction in question is meant to express.
We suggest that what makes the punitive reaction distinctive is a blaming attitude, or set of attitudes, that underlies it and that the punitive reaction is characteristically taken to express (resentment, indignation, personal offence, and condemnation) (Strawson, Reference Strawson1962). In the context of state-authorized sanctions, Feinberg (Reference Feinberg1965) argues that we make an intuitive distinction between “punishments” and “penalties” on this ground: Punishment is a legal sanction that expresses the state’s condemnation of both norm-violator and norm-violation, whereas “penalty” is a legal sanction that expresses no such thing; it is meant to operate on (potential) norm-violators as a purely instrumental deterrent.Footnote 5 Compare, for instance, the way we might respond to someone who violates a rational norm versus someone who violates (what we take to be) a moral norm. Even supposing the transgressor is at fault in both cases (i.e., we take them to be normatively competent in the relevant domain), the faulty reasoner may attract (at worst) derision or contempt, whereas the faulty moral agent attracts some degree of reprobative blame – that is, blame that characteristically expresses resentment or indignation. Why do these norm violations attract such different responses?
We think a satisfying answer to this question must finally appeal to the substantive nature of the norms themselves: in particular, to the idea that moralized norms generally require us to give due weight to a range of disinterested, mainly other-regarding concerns versus our own idiosyncratic interests or desires (whether “selfish” or not) (Voiklis & Malle, Reference Voiklis, Malle, Gray and Graham2017). To violate these norms is thus to prioritize our own interests and desires in a way that shows disregard or disrespect for a communally endorsed conception of the moral order. And though other kinds of norms are communally endorsed, violating them is not seen to be essentially self-prioritizing in the manner of a moral norm violation. In short, as moral philosophers are wont to emphasize, people view (normatively competent) violators of moral norms as expressing a distinctive attitude in their transgressive behavior – an attitude of disrespect, disregard, or even “ill will” toward others in the moral community, or to the community at large (Strawson, Reference Strawson1962). And it is this expressed attitude that calls for a distinctive response: one that communicates condemnation of the transgressor’s attitude, in addition to whatever (psychological or material) cost is imposed for the transgressive act itself.
But is the mere expression of condemnation sufficient to make punishment a genuinely communicative act? As we have argued, human communication is essentially dialogical in nature. Hence, if punitive responses to moral transgressions are genuinely communicative, they should be psychologically and/or normatively linked with an expectation of (perhaps even demand for) a suitable response from the target audience – for instance, from transgressors themselves, who as communicative targets of punishment are thereby called upon to regret and renounce their immoral acts and attitudes. A communicative view of punishment thus entails that punishment has a certain forward-looking drive or rationale, one that ultimately seeks the ratification of shared moral norms from its target audience. As we have noted, the target audience may – and perhaps should – include transgressors themselves; but as a communal or even state-authorized act, it is likely to include others as well, in whose name the punishment is implicitly or explicitly imposed (Feinberg, Reference Feinberg1965).
A communicative view of punishment has been explored and defended in the philosophical literature, most notably by Anthony Duff (Reference Duff2001) (for a sample of views, see also Bennett, Reference Bennett2008; Braithwaite & Pettit, Reference Braithwaite and Pettit1990; Feinberg, Reference Feinberg1965; Lacey, Reference Lacey1988; von Hirsch, Reference von Hirsch1993).Footnote 6 While this work focuses on seeking an adequate justification for the practice of punishment, our concern in the remainder of this chapter takes a more empirical direction. In what follows, we focus specifically on people’s actual motivations for punishment and how a communicative framework can forward psychological research on this topic.
16.2 Communication and the Psychology of Punishment
16.2.1 A Brief History of Punishment Research
In the psychological literature, punishment has been somewhat broadly defined as “intentionally providing another person with negative or unwanted outcomes, as a motivated response to the perception that this person has violated widely shared norms, values, or rules” (van Prooijen, Reference van Prooijen2018, p. 10). (For a narrower treatment of punishment, see Chapter 15, this volume.) Decades of research into people’s punishment behavior indicate that it is motivated by several factors (for recent reviews, see Raihani & Bshary, Reference Raihani and Bshary2019; van Prooijen, Reference van Prooijen2018; Wenzel & Okimoto, Reference Wenzel, Okimoto, Sabbagh and Schmitt2016). Broadly, this research can be grouped into two different kinds. One broad kind comes from behavioral economics and evolutionary psychology and primarily focuses on the ultimate functions of punishment and its adaptive value in the context of evolution. Traditionally, this area of research has operationalized punishment as payoff reductions in the context of incentivized economic games. The most influential explanation for punishment as an adaptive mechanism over the past two decades is that it stimulates cooperation within groups by way of deterring wrongdoers, as well as communicating and reaffirming the group’s norms and values (Fehr & Fischbacher, Reference Fehr and Fischbacher2004). Other functional explanations highlight punishment’s competitive functions and beneficial effects for punishers, for instance, reestablishing their status and relative level of resources (Krasnow et al., Reference Krasnow, Cosmides, Pedersen and Tooby2012; Raihani & Bshary, Reference Raihani and Bshary2019), as well as signaling their trustworthiness to others (Jordan et al., Reference Jordan, Hoffman, Bloom and Rand2016).
The second broad kind of research on punishment comes from social psychology and focuses on the proximate mechanisms of punishment, examining why people punish in a given situation. In line with findings from economics and evolutionary psychology, social psychological research has identified the importance of rebalancing status and power and restoring value consensus (e.g., justice restoration theory, Okimoto & Wenzel, Reference Okimoto and Wenzel2008; Shnabel & Nadler, Reference Shnabel and Nadler2008). Still, over the last few decades this research tradition has been largely shaped by traditional philosophical theories regarding the appropriate justification for punishment. In this normative tradition, deontological justifications, historically associated with Kant, are purely retributive and backward-looking: The offender simply deserves to be punished as a reaction to their wrongdoing. By contrast, a consequentialist approach, historically associated with Bentham, justifies punishment in terms of specific forward-oriented consequences: For instance, it serves as a general deterrent by imposing a cost for criminal behavior and/or serves as a specific deterrent by incapacitating offenders; it may also prevent future crime by changing their behavior (rehabilitation). Numerous psychological studies have investigated to what extent retribution (as the deontological factor) versus deterrence (as the main consequentialist factor) drives people’s punishment-related reactions (for reviews, see van Prooijen, Reference van Prooijen2018; Wenzel & Okimoto, Reference Wenzel, Okimoto, Sabbagh and Schmitt2016), whereas other philosophical justifications (especially focusing on the expressive and communicative dimensions of punishment) have received comparatively little attention.
Results from these studies indicate more support for retribution than deterrence as punishment motive. For instance, in line with retributive theories, studies show that people punish proportionally to the seriousness of a transgression and that people’s punishment recommendations are less affected by factors central to deterrence theories, such as likelihood of detection or transgression frequency (e.g., Carlsmith et al., Reference Carlsmith, Darley and Robinson2002; Carlsmith et al., Reference Carlsmith, Wilson and Gilbert2008). In addition, if people can select which transgression-relevant information they would like to know before they recommend punishment, they select information on transgression seriousness first versus information related to deterrence, such as transgression frequency (Carlsmith, Reference Carlsmith2006; Keller et al., Reference Keller, Oswald, Stucki and Gollwitzer2010). A further set of studies has aimed to isolate retributive motives by studying people’s punitive responses toward animals that have attacked human beings (Goodwin & Benforado, Reference Goodwin and Benforado2015).Footnote 7 In line with the retributive view, the researchers found that people rate animals as “deserving” to be killed for their violent attacks; furthermore, people’s support for inflicting pain on animals during their execution increases with the perceived severity of the attacks. In sum, this body of social psychological research has been taken to support the view that people are first and foremost retributive in their punishment behavior, as they mainly care about giving transgressors what they “deserve.”
It is important to stress that most of these findings have been interpreted within a binary choice theoretical framework – retribution versus deterrence, thereby sidelining the potential role of communication as a key motivating factor. Nevertheless, researchers have recognized that communication may still play an important role in people’s punishment behavior. Subsequent work has thus aimed to control for this using carefully designed behavioral experiments. For instance, in so-called hidden punishment experiments, people are purportedly barred from using punishment in an immediately communicative way because the targets of their punishment will not be informed of the punishment they receive (Crockett et al., Reference Crockett, Özdemir and Fehr2014; Nadelhoffer et al., Reference Nadelhoffer, Heshmati, Kaplan and Nichols2013). Yet a small percentage of people still punish under these conditions. Similarly, studies have shown that people punish in repeated interactions, even if their targets will not be informed about whether or how much they have been punished until the very end of the experiment, apparently removing its immediate communicative rationale (Fudenberg & Pathak, Reference Fudenberg and Pathak2010). Many researchers conclude that this argues in favor of a purely, or at least predominantly, retributive punitive psychology. (For a critique of this conclusion, see Chapter 15, this volume, Section 15.3.3.)
We think the cited results do not make a strong case against the idea that punishment has a communicative dimension. In the first place, the putative isolation of retributive motives in controlled laboratory studies does not entail the stronger conclusion that people’s punishment motives are purely retributive. As an analogy, imagine a laboratory study that examines why people wear clothes. Assume that researchers want to study the effect of fashion preferences. If the results showed that people put on clothes of certain colors even if these clothes are hard to wear, it would be valid to conclude that fashion preferences affect people’s choice of clothing. Yet, it would not be a valid conclusion that comfort (or other factors such as warmth) play no role in people’s clothing decisions in everyday life. Indeed, one or more of these factors may remain an important, even dominating motivation in naturalistic conditions. Returning now to the punishment case, even if “hidden punishment” occurs, other factors may continue to play an important, perhaps dominating role, in more naturalistic conditions. Hence, it does not rule out communication as a proximate mechanism for punishment.Footnote 8
Second, if communication is an inherent feature of ordinary punishment behavior, studies on hidden punishment may simply demonstrate that people will engage in communicative behaviors even if they know their putative audience cannot hear them. Such behavior is a common occurrence in everyday life. People may show their communicative intent (verbally or otherwise) even when they know their target audience is not “listening” in any demonstrable way. They may do so for many reasons: to express who they are, to get things off their chest, to rehearse what to say if a suitable occasion arises, to solicit (imagined) advice, to (imaginatively) resolve disputes, or simply to seek comfort in the absent person’s (imagined) presence. Of course, “communicative behavior” could be defined in such a way as to require successful uptake in a target audience. But since our goal is to understand the psychological mechanisms motivating specific behaviors, we here take communicative intent to be independent of uptake. Building on our earlier discussion of human communication, the point we make here is that performing actions with communicative intent is likely to be a deeply rooted psychological proclivity in beings that are naturally oriented toward securing shared intentionality with others – a proclivity that is not easily “shut off” or subject to rational control. Thus, even in the context of anonymous single-shot punishment, it is certainly possible that punishers are likewise engaging in communicative behavior, even if potential targets of their communicative acts are not physically present.Footnote 9
These considerations point toward the same conclusion: Even if retributive aspects can be isolated experimentally, such findings do not establish that punishment is primarily retributive. Indeed, several research findings suggest that people’s punishment goals are context-dependent, as context changes the salience of punishment-relevant aspects of a given situation (e.g., salience of the crime itself, the offender, or the community; see Gromet & Darley, Reference Gromet and Darley2009; Twardawski et al., Reference Twardawski, Tang and Hilbig2020). For instance, people tend to put more emphasis on the need for deterrence and rehabilitation if the longer-term consequences of punishment are made salient. This already undermines the rationale for a binary choice between “the two” punishment motives (retribution versus deterrence). We think psychological studies would benefit from adopting a research framework that remains open to a potential range of several punishment motives (e.g., Fitness & Peterson, Reference Fitness, Peterson, Forgas and Fitness2008), while perhaps also discerning a unifying thread that may be present in most or all of them.
To this end, we think it helpful to regard punishment as a form of moral communication, capable of conveying a range of potential messages. As we have noted elsewhere (Funk et al., Reference Funk, McGeer and Gollwitzer2014; McGeer & Funk, Reference McGeer and Funk2017), a first advantage to this shift in framework is that it resolves a major ongoing debate between retributive and consequentialist theories of punishment: whether people’s punishment motives are, or should be, primarily backward-looking or primarily forward-looking. Under a communicative paradigm, punishment can be both retributive and instrumental (see also Duff, Reference Duff2001).
A second advantage of the communicative approach is that it frames punishment as a specifically moral response to wrongdoing – in that way, it satisfies retributivists’ normative ambitions; but it does so without relying on the strange moral alchemy that lies at the root of traditional retributivist views, that returning suffering for suffering somehow rectifies a moral wrong. The point of punishment, on the communicative view, is to convey a suitably targeted, expressively powerful, morally loaded message; it is not to inflict suffering as an end in itself (cf. Feinberg, Reference Feinberg1965).
Given these normative advantages, it is not surprising that the communicative view of punishment has received considerable attention and defense in the philosophical community. But, as a descriptive theory regarding people’s actual punishment motives, psychologists have given it surprisingly little attention in the design of their empirical studies over the past two decades. Luckily, times are changing.
16.2.2 Punishment as Communication: An Emerging Trend in Psychological Research
Recently, psychologists have started to study the communicative aspects of punishment experimentally and to differentiate them from retributive components. A first line of findings comes from research on the hedonic effects of punishment. From a retributive perspective, one could expect punishment in itself to be satisfying and restore people’s sense of justice. Yet this does not seem to be the case. In general, findings from social psychological studies on the hedonic consequences of punishment show that there are positive as well as negative effects of punishing (Eadeh et al., Reference Eadeh, Peak and Lambert2017). In particular, punishers are very sensitive to whether punishment sends a message and how that message is received. If punishers do not hear back from wrongdoers after punishment, they ruminate and are dissatisfied (Carlsmith et al., Reference Carlsmith, Wilson and Gilbert2008). Yet when punished wrongdoers indicate that they understand that they are being punished (Gollwitzer et al., Reference Gollwitzer, Meder and Schmitt2011) and especially if wrongdoers indicate a change in their behavior and attitude (Fischer et al., Reference Fischer, Twardawski, Strelan and Gollwitzer2022; Funk et al., Reference Funk, McGeer and Gollwitzer2014), punishers’ sense of justice is restored. Numerous studies have shown that people have less desire to punish offenders who are remorseful (e.g., Corwin et al., Reference Corwin, Cramer, Griffin and Brodsky2012). These findings fit well with the view that punishment is ideally a form of bidirectional (i.e., dialogical) moral communication: To wit, punishers intentionally express something via punishment that punished individuals are meant to understand – punishment carries symbolic meaning; and, likewise, punishers adjust their punishment behavior in reaction to intentional messages they receive back from the wrongdoers (for similar thoughts, see for instance Boon & Yoshimura, Reference Boon and Yoshimura2020; Nahmias & Aharoni, Reference Nahmias, Aharoni and Surprenant2018).
A second line of research explores the necessity of punishment’s materially punitive edge. In support of the communicative view, studies show that people expect harmless but symbolic punishment to be as effective as harmful punishment if its message is clear (Sarin et al., Reference Sarin, Ho, Martin and Cushman2021). Similarly, in real interactions, people accept trade-offs between punishment and other forms of moral communication. For instance, they reduce their punishment if they can also send along a note of disapproval (Xiao & Houser, Reference Xiao and Houser2005), and they opt for the less severe punishment if they can let the wrongdoer know why they are being punished (Molnar et al., Reference Molnar, Chaudhry and Loewenstein2023). Once again, these findings highlight that punishment behavior does not seem to be solely about punitive paybacks and proportionality as a purely retributive account of punishment would suggest.
16.2.3 Toward a Systematic Approach to Moral Communication: Content, Targets, and Means
We turn now to a consideration of how conceptualizing punishment as communication can provide researchers with a scaffold to systematically study and understand how people react to wrongdoing. If the underlying psychological mechanism of the urge to punish is communicative in nature, then three interrelated questions emerge as important to explore in depth: What is being communicated? Who is the target of communication? And how is this communicative content conveyed? Ultimately, a better understanding of these dimensions of punishment will allow theorists to question the efficacy of punishment in achieving its communicative goals and to invite comparison with other potentially more effective forms of moral communication.
The Content of Communication. What people aim to communicate via punishment is closely related to how people have construed the wrongdoing and what it meant to them psychologically. Did people construe it as a violation of their own personal status, for instance, or did they mainly see it as a violation of group norms? Punishment may communicate that someone has been wronged, or it may send a message reaffirming the victim’s standing, their values, or their moral identity (see, e.g., Okimoto & Wenzel, Reference Okimoto and Wenzel2008; Shnabel & Nadler, Reference Shnabel and Nadler2008). It may effectively “say” to others: “I won’t allow anybody to walk all over me” (Crombag et al., Reference Crombag, Rassin and Horselenberg2003), or “I’m a trustworthy person because I condemn such wrongdoing” (Jordan & Rand, Reference Jordan and Rand2020). It may communicate “I’m a member of a particular moral tribe”; “I care about a set of norms and their breaches, and I’m disposed to police the norms in question” (Shoemaker & Vargas, Reference Shoemaker and Vargas2019). It may communicate a desire regarding the future – for example, that the wrongdoer feels remorse and commits to change (Funk et al., Reference Funk, McGeer and Gollwitzer2014); it may even communicate a desire that wrongdoers suffer in light of their wrongdoing. And such desires might be essentially self-regarding (e.g., related to a victim’s need to restore their own standing) or they might be essentially other-regarding (e.g., related to a victim’s desire to prevent the victimization of others). There is a range of possibilities, with varying degrees of moral content. For researchers, it would be interesting to investigate factors that modulate expressed moral content; for instance, which message gets priority in what contexts – and why.
Importantly, once punishment is conceptualized as a nonlinguistic form of moral communication, it has one salient advantage over explicitly linguistic forms of communication: It may carry many symbolic messages simultaneously, either to the same target or to several targets, and potentially each with a different grade of urgency or importance to the punisher.
The Target of Communication. An interesting feature of punishment as communication is that the message may or may not be targeting the recipient of punishment. Obviously, punishment may communicate something to the wrongdoer, as previous research on the hedonic effects of punishment suggests (e.g., Funk et al., Reference Funk, McGeer and Gollwitzer2014; Gollwitzer et al., Reference Gollwitzer, Meder and Schmitt2011). But, as we noted earlier, the message may also be intended for others – for example, for particularly vulnerable members of the community (“You will be supported and protected”) or for the community as a whole (“Let’s all agree that this is a serious moral wrong that should be generally condemned”) (see also Feinberg, Reference Feinberg1965). Notably, the two consequentialist punishment motives, specific prevention and general prevention, can be accommodated, albeit reconfigured, within this communicative framework. While in the traditional picture, punishment is viewed as a relatively crude behavioral-conditioning device, it is here explicitly conceptualized as sending a message to the original wrongdoer or to the public at large that such wrongdoings will not be tolerated. A specific example on using punishment to communicate with various possible targets can be found within the legal system. While offenders are clearly the target of institutionalized punishment, the criminal justice system explicitly aims to communicate the validity of norms to several other targets, including the victims and their families, future potential wrongdoers, as well as potential voters and concerned citizens who are afraid of crime and who are supposed to get the message that wrong deeds are taken seriously (Feinberg, Reference Feinberg1965; Sunstein, Reference Sunstein1996).
The Means of Communication. There are many different forms of punishment and many ways in which punishment can be administered. Sometimes people react to wrongdoings with punishment that is proportional to the severity of the transgression (Carlsmith et al., Reference Carlsmith, Darley and Robinson2002) and may use retaliation as a quick default mode of communication. Within the communicative framework, this core retributive principle can be expressively reconceptualized. “Proportional” punishment may communicate to the community “this is how wrong it was,” or to the offender “this is how much you need to change” (Duff, Reference Duff2001). Still, the communicative framework can also explain why, at other times, a disproportional, more lenient reaction may be observed. In such cases, particular contextual factors may ensure that a more lenient punishment (or other symbolic gesture) is sufficient to convey a particular message effectively (see also Sarin et al., Reference Sarin, Ho, Martin and Cushman2021). For instance, community service is often sufficient in restorative justice settings to assure all the stakeholders that the crime has been taken seriously and suitable reparative measures have been undertaken (Johnstone, Reference Johnstone2013; Rossner & Bruce, Reference Rossner and Bruce2016; Shapland et al., Reference Shapland, Atkinson, Atkinson, Colledge, Dignan, Howes, Johnstone, Robinson and Sorsby2006). Or, to go to another example, a skeptically raised eyebrow may be enough to publicly lower a perpetrator’s status and reputation in the eyes of others. At the other extreme, punishers may decide not to interact again with the wrongdoer at all in order to communicate condemnation of the wrongdoer’s actions and attitudes (to the wrongdoer and/or to the group as a whole), to reaffirm solidarity with whatever group norms the wrongdoer has violated, and, finally, to endorse by way of enacting group norms for dealing with wrongdoers.
16.2.4 Putting the Communicative Framework to Work in Current and Future Research
Application to Current Findings. We here suggest that some otherwise puzzling results in punishment research may be profitably addressed by adopting the communicative framework we recommend – for instance, the mixed profile of findings related to harsher and/or more lenient punishment in intra- versus intergroup settings. On the one hand, studies show that deviant in-group transgressors are evaluated more negatively than deviant out-group transgressors (the so-called black-sheep effect; Marques et al., Reference Marques, Yzerbyt and Leyens1988). In addition, there is an independent effect of out-group leniency where punishers adjust their punishment downward if the transgressor belongs to a group that is of lower status than their own group (Braun & Gollwitzer, Reference Braun and Gollwitzer2012). On the other hand, there are also instances of in-group leniency (“in-group favoritism”), that is, in-group transgressors are punished less harshly than out-group transgressors (see, e.g., Sommers & Ellsworth, Reference Sommers and Ellsworth2000, or, for research findings from developmental psychology with children as participants, see Jordan et al., Reference Jordan, McAuliffe and Warneken2014). These apparently contradictory results may possibly be disentangled by way of identifying who the primary communicative target of punishment is meant to be (along with identifying the exact message it is meant to convey). For instance, is the primary communicative target the actual wrongdoer, in which case in-group leniency might be preferred if the message is regarded as effectively conveyed in a nonpunitive way (see, e.g., Sarin et al., Reference Sarin, Ho, Martin and Cushman2021)? Or are other in-group members the primary target, so that punishment would validate the group’s morals under attack, potentially resulting in harsher treatment of the black sheep? Or is punishment about confirming the victim’s membership status in the group? Or, finally, are out-group members the primary communicative target, which may result in either harshness or leniency, depending on what exactly the intended message is meant to be? Again, these examples highlight that punishment may be a means of communication, and the content depends on how the transgression has been construed.
In emphasizing that the punished entity is not necessarily the target of communication, research findings on punishing animals (Goodwin & Benforado, Reference Goodwin and Benforado2015) need not be interpreted in line with a purely retributive perspective. Instead, animal-punishers might be sending a message to those humans who observe the punishment, to the group as a whole, or even to the experimenter who will see participants’ response choices. Animals may also serve as a symbolic proxy for human beings, so that punishing them sends a message about what kinds of behavior in human beings would be considered wrong and subject to group condemnation. Similarly, identifying the target of communication helps in understanding the psychology of displaced punishment (i.e., when punishers do not punish the actual transgressor but someone else). Research on displaced revenge has illustrated that punishment can be satisfying if it does not punish the original wrongdoer but another member of the target group (Sjöström & Gollwitzer, Reference Sjöström and Gollwitzer2015) or even a person who is symbolically similar to the original transgressor (Washburn & Skitka, Reference Washburn and Skitka2015). These findings are hard to square with a purely retributive psychology of punishment but make sense on the communicative view.
Scaffolding Future Research. A communicative view of punishment invites researchers to further explore the content, target, and means of this robust human phenomenon, leading to a broader understanding and integration of the various findings uncovered thus far as well as inspiring work that still needs to be done. For instance, investigating punishment as communication generates interesting new research questions regarding the difference as well as the commonalities between second- and third-party punishment (i.e., punishment carried out by victims or observers). Is it the content, the target(s), and/or the means of communication that differ between them? Do second- versus third-party punishers, for instance, care differently about reaching various targets of communication (such as the transgressor and/or the community), potentially because they react more automatically versus deliberately in a given situation?
Such future research should also look into the ontogeny of punishment (and its alternatives) as communication. Adding to the many interesting studies about children as punishers (e.g., McAuliffe et al., Reference McAuliffe, Jordan and Warneken2015; Riedl et al., Reference Riedl, Jensen, Call and Tomasello2015), it would be worth studying at what age children start using punishment in a dialogical, communicative way. In addition, while gaining a better understanding about how socialization and cultural learning affect children’s displayed punishment behavior (see, e.g., Salali et al., Reference Salali, Juda and Henrich2015; Wu & Gao, Reference Wu and Gao2018) and how punishment behavior differs between children and adults (e.g., related to second- versus third-party punishment or in regards to being sensitive toward intent and outcome; see Bernhard et al., Reference Bernhard, Martin and Warneken2020), it would be worth studying how content, target, and/or means of punishment as communication may change over the course of children’s development.
Nonpunitive Forms of Moral Communication. Importantly, people can react to wrongdoing with communicative acts that do not involve retributive or punitive elements. Indeed, the prevalence of punishment behavior in economic games varies between subjects from different societies (e.g., Henrich et al., Reference Henrich, McElreath, Barr, Ensminger, Barrett, Botyanatz, Cardenas, Gurven, Gwako, Henrich, Lesorogol, Marlowe, Tracer and Ziker2006). In addition, if nonpunitive options are included in research paradigms, victims often prefer compensation over punishment (Heffner & FeldmanHall, Reference Heffner and FeldmanHall2019; for similar findings with children as participants see for instance Yang et al., Reference Yang, Wu and Dunham2021). Moreover, third-party observers often choose to help the victim instead of punishing the perpetrator (Chavez & Bicchieri, Reference Chavez and Bicchieri2013; for children as participants see Lee & Warneken, Reference Lee and Warneken2020), prioritizing communicative acts that acknowledge the victim’s value as a means of restoring justice. All of these reactions to injustice carry meaning and can be considered vehicles of communications.
Hence, a final advantage of the communicative approach we advocate here is that it has resources for explaining not just why people punish but why they sometimes refrain from punishing. Ideally, a theory of punishment will explain both. Notably, a purely retributive framework is not up to this task and cannot make any predictions. By its logic, so long as there are wrongdoers, they simply deserve to be punished. By contrast, punishment from a framework of moral communication may help predict when and why people might forego punishment. What was the wrong? What did the wrong communicate to begin with? How does it affect me as a victim or as an observer? What do I think about how others think it affects me? The construal of wrongs determines how people react to them. Future research would benefit from systematically using the framework of moral communication to identify the aspects that make punishment (versus compensation, helping, etc.) the most likely reaction to injustice, thereby more fully explaining why people do in fact punish.
Identifying the situations in which people’s underlying motives can also be addressed without punishment has important normative implications. If people’s need for punishment changes depending on the availability of other communicative options that are perceived to be suitable in a given setting, it would be desirable to educate people about these effects in order to de-escalate conflicts and to avoid a ratcheting-up effect of blame (McGeer, Reference McGeer, Coates and Tognazzini2013; McGeer & Funk, Reference McGeer and Funk2017). It would then directly benefit victims, transgressors, and observers of injustice to make such options available in the broader interests of “restoring justice,” as advocates of this explicitly named movement strongly recommend (Braithwaite, Reference Braithwaite2002; Braithwaite & Pettit, Reference Braithwaite and Pettit1990; Johnstone, Reference Johnstone2013; Strang, Reference Strang2002).
16.3 Conclusion
In this chapter, we have examined our human concern with, and commitment to, practices of moral communication – communicating about people’s attitudes and behavior regarding a distinctive range of moralized norms. In the first part of the chapter, our aim was to situate people’s concern with moralized norms in the context of their more general commitment to a norm-governed form of life. This enabled us to zero in on what is distinctive about moralized norms from a more encompassing perspective, leading us to place special emphasis on people’s characteristically punitive response to their violation, a response that invariably expresses moral disapproval or condemnation of the transgressors’ acts and attitudes. This led to our primary hypothesis that punishment (versus other kinds of merely instrumental negative sanction, e.g., fines) is fruitfully understood as a distinctive form of (essentially dialogical) moral communication. Our aim in the second part of the chapter was to put this hypothesis to work in an empirical context, showing how it profitably reframes many of the debates over existing data on the psychology of punishment, as well as opening up a number of interesting avenues for further research.
We close by reinforcing three critical points made salient by this communicative approach: 1) as a nonlinguistic form of communication, punishment is particularly apt for conveying a number of distinct messages to different target audiences at once (i.e., it can carry multiple symbolic meanings); 2) as a materially and/or psychologically costly form of communication, punishment has a distinctive power, conveying a message that is difficult to ignore (for instance, that transgressions of these sorts of norms will simply not be tolerated); and, finally, 3) as a form of human communication, punishment is essentially attitude-focused (i.e., concerned about other people’s mental states) as well as bidirectional or dialogical, aiming to elicit an appropriate response from its audience given the message conveyed. This last point is particularly worth emphasizing, as it opens the door to a range of questions concerning how people’s taste for punishment can be modulated (in terms of severity and/or means) once its communicative raison d’être has been sufficiently brought to light.

