Hostname: page-component-76fb5796d-5g6vh Total loading time: 0 Render date: 2024-04-27T07:08:16.119Z Has data issue: false hasContentIssue false

Skepticism Motivated: On the Skeptical Import of Motivated Reasoning

Published online by Cambridge University Press:  12 May 2020

J. Adam Carter*
Affiliation:
Philosophy, University of Glasgow, Glasgow, Scotland
Robin McKenna
Affiliation:
Philosophy, University of Liverpool, Liverpool, England
*
*Corresponding author: Email: adam.carter@glasgow.ac.uk
Get access
Rights & Permissions [Opens in a new window]

Abstract

Empirical work on motivated reasoning suggests that our judgments are influenced to a surprising extent by our wants, desires, and preferences (Kahan 2016; Lord, Ross, and Lepper 1979; Molden and Higgins 2012; Taber and Lodge 2006). How should we evaluate the epistemic status of beliefs formed through motivated reasoning? For example, are such beliefs epistemically justified? Are they candidates for knowledge? In liberal democracies, these questions are increasingly controversial as well as politically timely (Beebe et al. 2018; Lynch Forthcoming, 2018; Slothuus and de Vreese 2010). And yet, the epistemological significance of motivated reasoning has been almost entirely ignored by those working in mainstream epistemology. We aim to rectify this oversight. Using politically motivated reasoning as a case study, we show how motivated reasoning gives rise to three distinct kinds of skeptical challenges. We conclude by showing how the skeptical import of motivated reasoning has some important ramifications for how we should think about the demands of intellectual humility.

Type
Article
Copyright
© The Author(s) 2020. Published by Canadian Journal of Philosophy

Access options

Get access to the full version of this content by using one of the access options below. (Log in options will check for institutional or personal access. Content may require purchase if you do not have access.)

1.

Empirical work on motivated reasoning suggests that our judgments are influenced to a surprising extent by our wants, desires, and preferences (Kahan Reference Kahan2016; Lord, Ross, and Lepper Reference Lord, Ross and Lepper1979; Molden and Higgins Reference Molden, Higgins, Holyoak and Morrison2012; Taber and Lodge Reference Taber and Lodge2006). In this article, we address the import of motivated reasoning for the epistemic statuses of beliefs that are impacted by it. Our central questions are: Are such beliefs epistemically justified? Are they candidates for knowledge? In liberal democracies, these questions are increasingly controversial as well as politically timely (Beebe et al. Reference Beebe, Baghramian, Drury and Dellsén2018; Lynch Reference LynchForthcoming, 2018; Slothuus and de Vreese Reference Slothuus and de Vreese2010). And yet, the epistemological significance of motivated reasoning has been almost entirely ignored by those working in mainstream epistemology.Footnote 1 We aim to rectify this oversight.

Our general aim is to provide some reasons for thinking that politically motivated reasoning has negative import for the epistemic statuses of beliefs formed in part through such reasoning. We submit that there is some initial pressure to think that beliefs formed through politically motivated reasoning are (epistemically) unjustified. If our assessments of arguments and evidence is partly dictated by our political convictions, then how could these assessments be epistemically justified? Our aim in this paper is to make this initial thought more precise, and to argue that it is more than just initially plausible. In fact, as we’ll show, the widespread presence of motivated reasoning in the political domain should lead us to be much less confident in our political beliefs, and for reasons that have not been adequately appreciated by epistemologists.

Of course, one very natural strategy would be to look at whether politically motivated reasoning is a reliable way of going about forming beliefs. There is empirical evidence which suggests that it isn’t, in that it leads many of us to form beliefs about scientific topics that conflict with the scientific consensus (Kahan, Jenkins-Smith, and Braman Reference Kahan, Jenkins-Smith and Braman2011). If this is right, then empirical work on politically motivated reasoning poses a direct skeptical challenge: knowledge requires reliable belief formation, so if beliefs formed via politically motivated reasoning are unreliable, they can’t be known.Footnote 2

While this line of argument represents one straightforward way in which politically motivated reasoning might give us pause, we want to explore another way in which politically motivated reasoning should give us pause even if it doesn’t lead us astray, but rather fortuitously “guides” us toward the truth. Maybe some beliefs formed through politically motivated reasoning are unjustified even if they generally tend to be true. We are going to argue that we should take this possibility seriously.Footnote 3

We will start with an overview of work on motivated reasoning. We will then develop three skeptical challenges that this work poses for certain of our beliefs. Finally, we will finish by taxonomizing these skeptical challenges according to their “level” of skeptical power, and by showing how motivated reasoning has some important ramifications for the way we think about the demands of intellectual humility.

2.

Manifestations of motivated reasoning can be distinguished in terms of the motivations they serve (Molden and Higgins Reference Molden, Higgins, Holyoak and Morrison2012). There is a large body of work that documents the ways in which our desire to maintain a positive self-conception leads us to give more credence to information that confirms our perception of ourselves as kind, competent, and healthy than to information that challenges these perceptions. For example, Kunda (Reference Kunda1987) reports a study in which subjects read an article which stated that caffeine consumption causes serious health problems in women but not in men. Women who were heavy caffeine consumers found the article less convincing than women who were light caffeine consumers.Footnote 4

We are here particularly interested in the impact that our political beliefs and convictions have on our assessment of arguments that pertain to those beliefs and convictions. Call this politically motivated reasoning. For example, Lord, Ross, and Lepper (Reference Lord, Ross and Lepper1979) report a study in which subjects were given the same set of arguments for and against capital punishment, but their assessments of the strength of these arguments correlated with their existing views about the rights and wrongs of capital punishment. Put simply, subjects who were predisposed to object to capital punishment found anticapital punishment arguments more convincing, whereas those who were predisposed to accept it found procapital punishment arguments more convincing. Dan Kahan and his collaborators present a series of studies which show that our political beliefs also impact our assessment of arguments and evidence that pertain to scientific issues with clear political relevance, such as global warming and the safety of nuclear power (Kahan, Jenkins-Smith, and Braman Reference Kahan, Jenkins-Smith and Braman2011; Kahan Reference Kahan2016, Reference Kahan, Boykoff and Crow2014).Footnote 5

Three aspects of work on politically motivated reasoning are worth emphasizing from the outset. First, while we have been talking of politically motivated reasoning in terms of the influence of our background political beliefs on our cognitive processes, it is important to note that the operative notion of “political belief” is very broad indeed. This is perhaps clearest in Dan Kahan and his collaborators’ work on cultural cognition. Kahan calls attention to the fact that politically motivated reasoning is as much driven by our broader cultural backgrounds as our political beliefs, narrowly construed. The thought is that the motivation or goal that is served by politically motivated reasoning is, broadly speaking, the goal of identity protection— that is, the goal of forming beliefs that protect and maintain our status within a group that defines our identity and whose members are united by a shared set of values. As Kahan puts it:

When positions on some risk or other policy relevant fact have come to assume a widely recognised social meaning as a marker of membership within identity-defining groups, members of those groups can be expected to conform their assessment of all manner of information—from persuasive advocacy to reports of expert opinion; from empirical data to their own brute sense impressions—to the position associated with the respective groups. (Reference Kahan2016, 1)

While it will do no real harm to simplify things and talk in more narrowly political terms (“right wing” vs. “left wing”), it should be borne in mind that this is a simplification.

Second, the empirical evidence does not suggest that politically motivated reasoning is less prevalent in “politically sophisticated subjects.” In fact, it is more like the opposite. Taber and Lodge (Reference Taber and Lodge2006) found that the more politically sophisticated a subject was, the more likely they were to conform their assessment of the quality of the arguments and information they consider to their political beliefs.Footnote 6 Taber and Lodge hypothesize that this is because politically motivated reasoning involves two biases. We tend to spend more time and resources finding flaws with arguments that challenge deeply held political beliefs than we do on finding flaws with arguments that support these beliefs, and, when we are free to choose which information to expose ourselves to, we tend to seek out arguments that will confirm deeply held political beliefs rather than arguments that will challenge these convictions. Put simply, politically sophisticated subjects have more information at their disposal, and the more information you have at your disposal, the better you will be at finding flaws in arguments with conclusions you don’t like, and the better you will be at seeking out information that confirms rather than challenges your existing views.Footnote 7

Third, politically motivated reasoning is something that many of us engage in, at least some of the time.Footnote 8 This is particularly clear in Kahan and his collaborators’ work. They look at the impact of our political beliefs on the following topics, among others:

  • Attitudes about global warming

  • Attitudes about the safety of “burying” nuclear waste underground

  • Attitudes about the efficacy of “concealed-carry” laws

One would expect laypersons to look for the existence of scientific consensus on these sorts of topics. Kahan, Jenkins-Smith, and Braman (Reference Kahan, Jenkins-Smith and Braman2011) report the following findings:

  • Seventy-eight of those on the “political left” think that most scientists agree that global temperatures are rising, whereas only 19% of those on the “political right” think that most scientists agree. (Fifty-six percent of those on the right think that most scientists are divided and 25% think that most disagree.)

  • Sixty-eight percent of those on the left think that most scientists agree that humans are causing global warming, whereas only 12% of those on the right think that most scientists agree. (Fifty-five percent of those on the right think that most scientists are divided and 32% think that most disagree.)

  • Thirty-seven percent of those on the right think that most scientists agree that there are safe methods of geologically isolating (“burying”) nuclear waste, whereas only 20% of those on the left think that most scientists agree. (Thirty-five percent of those on the left think that most scientists disagree and 45% on both sides of the political spectrum think that scientists are divided.)

  • Forty-seven percent of those on the right think that most scientists agree that permitting adults without criminal records or histories of mental illness to carry concealed handguns in public decreases violent crime, whereas only 10% of those on the left think that most scientists disagree. (Forty-seven percent of those on the left think that most scientists disagree and much the same percentage on both sides of the political spectrum think that scientists are divided.)

What does this tell us? We want to highlight one crucial point. It is often claimed that there are important cognitive asymmetries between “liberals” (in the US sense) and “conservatives” (Hodson and Busseri Reference Hodson and Busseri2012; Kanazawa Reference Kanazawa2010). It may well be that there are some asymmetries. But these findings suggest that there are also some important symmetries. Footnote 9 Both those on the right and those on the left form beliefs about scientific topics in ways that accord with their political convictions. In doing so, they end up at odds with expert scientific opinion, at least as it is represented in “expert consensus reports” from the US National Academy of Sciences (NAS). The NAS has issued reports with the following conclusions:

  • There is scientific consensus that global warming is real, and human activity is a central cause of it (National Research Council Committee on Analysis of Global Change Assessments 2007).

  • There is scientific consensus that there are safe methods of geologically isolating (or “burying”) nuclear waste (National Research Council Board on Radioactive Waste Management 1990).

  • The available evidence does not permit forming a conclusion on the efficacy of concealed-carry laws (National Research Council Committee to Improve Research Information and Data on Firearms 2004).

This suggests that those on the right are out of step with the scientific consensus on global warming, whereas those on the left aren’t. But those on the left are out of step with the scientific consensus on the safety of nuclear waste disposal. And both sides overestimate the extent to which the evidence supports their preferred position on concealed-carry laws. Now, it may be claimed that these findings still support the conclusion that those on the right are more out of step with the scientific consensus than those on the left. That is, politically motivated reasoning is more pronounced (or at least more damaging) amongst those on the right than amongst those on the left. This might be true, but it hardly shows that those on the left don’t also engage in politically motivated reasoning, or that, in doing so, they don’t regularly end up at odds with expert scientific opinion.Footnote 10

This completes our brief overview of work on politically motivated reasoning. We will now turn to its epistemological implications and, in particular, its skeptical import.

3.

Politically motivated reasoning is reasoning that serves a motive other than a desire to arrive at true beliefs (Kahan Reference Kahan2016). If a subject engages in politically motivated reasoning when assessing some evidence or argument, their assessment of that evidence or argument is nontrivially influenced by their background political beliefs.Footnote 11 If the evidence or argument causes trouble for those beliefs, they try to reject it, explain it away, or minimize its importance; if the evidence or argument supports those beliefs, they enthusiastically endorse it, and exaggerate its importance. The evidence outlined above suggests that when subjects engage in politically motivated reasoning, they are often led astray: they end up forming false beliefs. Consider, for instance, the attitude of many people on the political right toward global warming. Sometimes, though, cultural cognition does not lead us astray, but rather fortuitously “guides” us toward the truth. In such cases, are our beliefs thereby in the clear, epistemically speaking? Perhaps not.Footnote 12

Consider the following pair of cases:

Naïve Student-1: Tim is an impressionable student, one who is inclined to blindly trust socially approved opinions with little to no critical scrutiny. Fortunately for Tim, he is raised in a household where exposure to (and social praise of) sound climate science is abundant, and climate change “denial” is discussed only in a negative light. Tim (like most around him) believes that global temperatures are rising, and that human activity is the cause.

Naïve Student-2: Tim* is an impressionable student, one who is inclined to blindly trust socially approved opinions with little to no critical scrutiny. Unfortunately for Tim*, he is raised in a household where exposure to (and social praise of) sound climate science is rare, and “climate change” is discussed only in a negative light. Tim* (like most around him) believes that it’s not the case that global temperatures are rising.

With the literature on politically motivated reasoning in hand, we can give a plausible psychological explanation of what is going in in both these cases. Both Tim and Tim* assess any information they receive pertaining to climate science in light of their background political beliefs and broader cultural backgrounds.Footnote 13 But, at least in part because these background beliefs are very different, these assessments lead them in opposite directions. Tim ends up believing that global temperatures are rising, and that human activity is the cause. Tim* ends up denying this. We would submit that neither Tim nor Tim* are particularly unusual. While our descriptions of both are undeniably simplistic, many (though of course not all) individuals are, in most essential respects, like Tim or Tim*.

Tim*’s belief in Naïve Student-2 is clearly epistemically defective.Footnote 14 But what about Tim’s belief? Interestingly, it might be argued that there is also something amiss with Tim’s belief, and this is so even though: (i) the target belief is true and (ii) Tim possesses excellent evidence for his belief (i.e., the evidence that we may suppose he is exposed to, and which is socially approved in his epistemically friendly environment).

We can start with the distinction, familiar in mainstream epistemology, between propositional and doxastic justification. Roughly, this is the distinction between having good reasons for one’s belief (propositional justification) and properly basing one’s belief on the good reasons one possesses (doxastic justification).Footnote 15 In Naïve-Student-1, Tim has good reasons for his belief, but it is not at all clear that he is properly basing his belief on the good reasons that he has. Rather, it seems, to the extent that he is basing his beliefs on the good reasons he has, he is doing this only because these beliefs and reasons have the epistemically irrelevant property of being socially approved. And if that’s right, then—as this line of thought goes—Tim, despite possessing propositional justification for his belief that global temperatures are rising, lacks doxastic justification (and ipso facto propositional knowledge) for believing that global temperatures are rising.Footnote 16

If the above line of thought is correct, then an interesting and rather troubling result is that politically motivated reasoning can have epistemically deleterious effects even when it guides one toward the truth, and thus, regardless of whether politically motivated reasoning is reliable.Footnote 17 But is it correct? In the next two sections we consider two objections to it.

4.

We have argued that Tim’s belief that global temperatures are rising is (doxastically) unjustified because, to the extent that he bases this belief on the good reasons he has, he does this because those reasons have the epistemically irrelevant property of being socially approved. That is, he believes on the basis of these reasons because those around him regard these reasons as being good reasons. One might object that this property is not epistemically irrelevant at all. Tim is utilizing a generally effective strategy for deciding what to believe (and why to believe it): defer to socially recognized authorities. Deferring to recognized authorities is not a bad epistemic strategy. It isn’t going to work in all environments, of course—recognized authorities can be wrong (see Tim*)—but this is just an instance of the more general point that good inferential strategies won’t work well in all possible environments (see Gigerenzer Reference Gigerenzer2000).

We agree that there need not be anything wrong with deferring to recognized authorities, and that good inferential strategies need not work well in all possible environments, but we don’t think this spells trouble for our argument. Let’s start with the first concession (that there need not be anything wrong with deferring to recognized authorities). The crucial question is: What is this deference based on? It is one thing to defer to recognized authorities because you have reason to think they are likely to have the right answers (or, at least, more likely than you). It is another to defer to them because you recognise that you share certain core values or belong to the same social group. We would submit that while the former makes perfect sense from the epistemic point of view, the latter is problematic. The mere fact that you share certain core values with someone surely can’t be a good reason to defer to them, at least on factual matters like global warming.Footnote 18

The literature on politically motivated reasoning suggests that, when we engage in it, our patterns of deference are best explained in terms of a preference for listening to those with whom we share certain core values. Kahan, Jenkins-Smith, and Braman (Reference Kahan, Jenkins-Smith and Braman2011) look at what drives subjects’ assessments of the expertise of (fictional) putative authorities on issues like global warming. They find that these assessments are driven by fit between the views of a putative authority and one’s ideological predispositions. Thus, many conservatives will regard scientists who accept global warming as less authoritative than scientists who deny it, and liberals will regard scientists who accept that there are safe nuclear waste disposal methods as less authoritative than scientists who deny this.

What about the second concession—that good inferential strategies need not work well in all environments? The problem with the strategy of deferring to those with whom you share certain core values isn’t that there are possible environments in which it isn’t very effective in producing true beliefs. The problem is rather that, at least for most of us, it isn’t very effective in the world in which we find ourselves. There may, of course, be people for whom it is effective. But work on politically motivated reasoning suggests that there are a lot of people for whom it isn’t. Take again Dan Kahan’s work on cultural cognition (discussed above). He finds that while liberals tend to have views about global warming that align with the scientific consensus, they also tend to have views about the safety of nuclear power that don’t. This, we think, demonstrates the shortcomings of the strategy of deferring to those with whom you share core values.

5.

The second objection is as follows: In Naïve-Student-1, Tim does properly (enough) base his belief on the reasons he has. To see why, compare Tim with an individual who clearly does not properly base his belief on the reasons he has, Jennifer Lackey’s “racist juror”:

Racist Juror: Martin is a racist juror. Martin receives, over the course of the trial, compelling testimony that the defendant is guilty. Martin however bases his belief that the accused is guilty not on the good reasons he has for thinking so, but on his racist belief that individuals of the ethnicity of the defendant are likely to commit criminal acts.Footnote 19

Both Tim and Martin are propositionally justified in their respective beliefs: there are good reasons for them to believe the relevant propositions. But, so far as doxastic justification is concerned, there seems to be a crucial difference between Tim and Martin. First, while Tim bases his belief that scientists agree that global temperatures are rising on the good reasons that he has, Martin bases his belief that the defendant is guilty on bad, racist reasons. Granted, Tim bases his belief on these good reasons because they are socially approved in his environment. But, nonetheless, he still based his belief on these reasons. So, a fortiori, we should maintain that unlike Martin in Racist Juror, Tim in Naïve Student-1 is not only propositionally justified but also doxastically justified.

We think this objection fails for two principal reasons. The first is a general reason about the relationship, thus far not made explicit, between basing and doxastic justification. If basing suffices for doxastic justification in the sense that, if a subject S bases her belief that proposition p on a reason, R, and R is a good reason for believing p, then S is doxastically justified in believing that p, then we concede that it follows that Tim’s situation in Naïve Student-1 wouldn’t be defective in the way we’ve indicated.

But, crucially, basing is not sufficient for doxastic justification in this way, even if basing is necessary for doxastic justification. This point does not rely on a commitment to any particular substantive view of basing.

For example, on a doxastic account of the basing relation (e.g., Audi Reference Audi1982; Ginet Reference Ginet1985), basing a belief on a reason is a matter of possessing a meta-belief to the effect that that reason is a good reason to hold that belief. If that meta-belief is itself a matter of, say, wishful thinking, the target belief is surely not doxastically justified even if (on the doxastic theory) it counts as being based on a good reason; otherwise, wishful thinking could convert an unjustified belief into a justified belief.

The same point can be made on an inferential account of basing (e.g., Bondy and Carter Reference Bondy, Carter and Carter2020) according to which inference from believed premises to a believed conclusion is sufficient for basing belief in the conclusion on belief in the premises. Ordinarily, on an inferential account of basing, basing will issue in doxastically justified belief. However, inference (like any mental activity) itself can (as Ryle [Reference Ryle1945] famously noted) be competent or incompetent. Incompetent inference (e.g., one that transitions through a bad step) can result in basing, on an inferential account of basing, even if it does not result in a doxastically justified belief.

Finally, consider a causal account of basing. On such a view, basing is a matter of causation in the sense that a subject S’s belief that p is based on R iff S’s belief that p is causally sustained by R. Footnote 20 Even if we assume R is a good reason for P, it is a mistake to think that the mere presence of such a causation relation will suffice for S’s belief that p to be doxastically justified. Here it is important to distinguish causation simpliciter from competent causation, where the latter requires not just that R cause p nondeviantly (see, e.g., Plantinga Reference Plantinga1993, 69n 8) but also that R cause p via the exercise of a reliably truth-conducive disposition of the subject.

In short, the idea is that doxastic justification requires not just basing as such, but good or proper basing—where proper basing requires having a good reason as one’s basis, but is not secured by simply having a good reason as one’s basis.

Here is not the place to offer a full account of proper basing (though offering such an account remains a live project in contemporary epistemology).Footnote 21 Rather, we want to emphasize that cases like that of Naïve-Student-1 aren’t going to be insulated from the kind of criticism we raise simply because basing on a good reason is present in a way that it is not in cases like, e.g., Racist Juror.

Of course—and this bring us to the second line of resistance to the anticipated reply—it doesn’t follow from the fact that basing on a good reason doesn’t suffice for doxastic justification that Tim’s belief is not in fact both based on a good reason and, additionally, doxastically justified. Why, exactly, should we think that features of Tim’s situation should lead us to deny that his belief is doxastically justified even though it is ex hypothesi based on a good reason?

The answer, in short, has to do with an epistemically problematic feature of the way in which the target belief is based. To sharpen what’s concerning about the way Tim bases his belief in Naïve-Student-1, consider now the following case:

Francophile Cartographer: Rae (irrationally) believes that French cartographers are the only reliable sources of cartographical information. Rae’s Francophilia so strongly influences her assessment of the reliability of maps that she distrusts all information written by non-French authors. Rae stumbles upon several pieces of evidence (E1, E2, and E3) for believing cartographical claim X. But, simply because E1, E2, and E3 were written by Italian cartographer Giacomo Gastaldi, Rae disregards this evidence and so does not come to believe claim X on the basis of it. Later that day, Rae encounters the very same pieces of evidence (E1, E2, and E3) for cartographical claim X, but this time they were written by French cartographer Pierre Desceliers. Because—and only because—Desceliers is French, she accepts this evidence, and comes to believe X on the basis of it.

In Francophile Cartographer, although Rae bases her belief on what we may assume are the good reasons she has for believing claim X—viz, E1, E2, and E3—she does not properly base her belief on these reasons, viz, in a way that would plausibly be required for doxastic justification. When Rae encounters these pieces of evidence in the work of Gastaldi, she disregards them; when she encounters them in the work of Desceliers, she accepts them. Rae’s belief isn’t properly based on these pieces of evidence if whether she believes on the basis of them depends primarily on the nationality of the putative expert (an epistemically irrelevant feature). She is basing her belief on good reasons not because they are good reasons but, rather, primarily because of something epistemically irrelevant. And a result is that the process by which she is basing her belief on a good reason is problematic as a candidate for issuing doxastic justification on any of the three accounts of basing considered.

With reference to the causal and inferential accounts, it’s worth noting that Rae’s basing process is an incompetent one, even when it reliably issues in a good reason as a basis.Footnote 22 (After all, the disposition manifested here is a generally unreliable one, even if all the French communicators Rae is in contact with happen to speak truthfully.)Footnote 23 Because incompetent basing is justification-undermining on a causal account as well as on an inferential account, Rae’s belief, based on a good reason, falls short of doxastic justification despite being based on a good reason with reference to either of these accounts.

Likewise, with reference to the doxastic account, it should be noted that, while Rae we may grant has a meta-belief to the effect that E1, E2 and E3 are good reasons for believing X (and so bases her belief that X on these good reasons), this meta-belief is itself unjustified--viz., it is a result of irrational prejudice. Given that irrational prejudice surely can’t convert an otherwise doxastically unjustified belief into a doxastically justified belief, it follows that Rae’s belief, based on a good reason, falls short of doxastic justification on a doxastic account of basing despite being based on a good reason.

But notice that once we accept this line of thinking in Francophile Cartographer, then, by parity of reasoning, we should also accept that Tim in Naïve Student-1 isn’t doxastically justified in believing that global temperatures are rising. Both cases are structurally on a par; in both cases, the subject bases her belief on good reasons not because they are good reasons, but rather, primarily because of something epistemically irrelevant. And, a result in both cases is that the process by which the subject is basing her belief on a good reason is problematic as a candidate for issuing doxastic justification regardless of what view of basing one subscribes to.

So our argument is as follows. There are (at least) two ways in which one’s belief can fail to be doxastically justified. One may be like Martin in Racist Juror, believing some proposition for which there are good reasons, but not on the basis of those good reasons. Or one may be like Tim in Naïve Student-1 or Rae in Francophile Cartographer. That is, one might believe some proposition for which there are good reasons on the basis of those good reasons not because they are good reasons, but for some other epistemically irrelevant (or plain bad) reason, so as to render the basing epistemically incompetent or otherwise defective. If this is right, then our skeptical worry remains: politically motivated reasoning can have epistemically deleterious effects even when it guides one toward the truth.

As we have already indicated, we think that Tim is not particularly unusual, so what we say about Tim’s belief that global temperatures are rising is going to hold for many people who believe that global temperatures are rising. In what follows we will do two things. First, as promised (footnote 9), we will indicate how our argument relates to the literature on irrelevant influences on belief. Second, we will explore a further skeptical worry that we think might arise in even the “best” sort of case.Footnote 24

6.

We have been looking at one way in which our beliefs can be influenced by factors that have nothing to do with the truth of their contents: they can be formed via politically motivated reasoning. Such factors are called “epistemically irrelevant” because they are irrelevant to the truth of the beliefs in question. The literature on epistemically irrelevant influences on belief looks at the general question of what to say about beliefs that are influenced by epistemically irrelevant factors.Footnote 25 While we are primarily concerned with politically motivated reasoning in particular, rather than irrelevant factors in general, the reader might want to know how our argument relates to this literature.

There are two (related) respects in which the argument above differs from much of the literature on epistemically irrelevant influences. First, there is a divide between those who think that irrelevant factors pose a problem for (some) of our beliefs (e.g., Avnur and Scott‐Kakures Reference Avnur and Scott‐Kakures2015; Vavova Reference Vavova2018), and those who are more skeptical (e.g., Mogensen Reference Mogensen2016; Srinivasan Reference Srinivasan2015; White Reference White2010). Those who think that irrelevant factors can pose a problem tend to focus on the idea that learning your belief(s) have been influenced by irrelevant factors gives you a defeater for your belief(s). For instance, Katia Vavova defends this principle:

Good Independent Reason Principle: To the extent that you have good independent reason to think that you are mistaken with respect to p, you must revise your confidence in p accordingly—insofar as you can. (Reference Vavova2018, 145)

Imagine that my political beliefs have been shaped by my upbringing, and I come to recognise this. Vavova’s thought is that this would occasion a skeptical worry about my political beliefs if it gave me good reason to think my political beliefs might be mistaken.

Setting aside the implications of Vavova’s proposal, the important point for our purposes is that, for Vavova, irrelevant factors are important insofar as being aware of them may gave us reasons to think certain beliefs are mistaken.Footnote 26 The idea is that recognizing the influence of irrelevant factors on our beliefs may give us a defeater for these beliefs. Our argument above was that politically motivated reasoning constitutes a problem for some of our beliefs because beliefs formed through politically motivated reasoning are not held on the right sort of basis (or not held on the right basis in the right way). Our argument therefore differs from Vavova’s in that we are interested in the consequences of politically motivated reasoning for the grounds on which our beliefs are held, rather than in whether the recognition that our beliefs have been influenced by irrelevant factors can yield a defeater for our beliefs.

This difference is the consequence of a second and more fundamental difference. Vavova, like many of the authors who’ve written on irrelevant influences, is primarily interested in the consequences of the recognition that our beliefs have been influenced by irrelevant factors for the epistemic statuses of our beliefs. Her question is: Under what conditions should this recognition lead us to revise our confidence? But our argument has not focused on the consequences of the recognition that some of our beliefs have been influenced by politically motivated reasoning. Rather, we have focused on the epistemic status of beliefs formed through politically motivated reasoning whether the subjects are aware of the origins of their beliefs or not. In our discussions of Tim (and Rae) above, we didn’t specify whether Tim (or Rae) are aware of how they formed their respective beliefs. This doesn’t preclude us from thinking of the issue in terms of defeaters. One might think there are defeaters for one’s beliefs of which one is not aware, and that the existence of such defeaters has consequences for the epistemic statuses of one’s beliefs. So we are interested in a different issue than Vavova and others, though there may be some interesting connections between our projects.

7.

We have shown that our argument is importantly distinct from Vavova’s (and much of the literature on irrelevant influences on belief) because it doesn’t focus on the consequences of recognizing that our beliefs have (or might have been) influenced by irrelevant factors. That said, we want to finish by describing a case in which a subject does suspect that their beliefs have been influenced by politically motivated reasoning:

Enlightened Student (Part I): Diane is like Tim in the following respects: (i) Diane is raised in a household where exposure to (and praise of) sound climate science is abundant, and climate change “denial” is discussed only in a negative light; (ii) Diane (like most around her) believes that scientists agree that global temperatures are rising. Diane is unlike Tim in the following respects: (iii) Diane is not influenced or impressed by (as Tim is) which opinions are the socially approved ones; the matter of whether a given opinion is socially approved (or disapproved) has a negligible effect on whether Diane accepts it, even though Diane ends up accepting many claims (including the claim that <scientists agree that global temperatures are rising>) that are the accepted ones in her epistemically friendly environment.

It seems clear that Diane in Enlightened Student (Part 1) is not just propositionally justified but also doxastically justified in believing that global temperatures are rising, and that human activity is the cause. Unlike Tim, Diane seems to properly base this belief on the good reasons that she has for holding it. She doesn’t believe on the basis of these reasons because they are socially approved but, rather, because they strike her as good reasons. We take this to show that there is nothing wrong with Diane so far as doxastic justification and the basis of her beliefs is concerned.

But even Diane may face some trouble. Imagine Enlightened Student continues:

Enlightened Student (Part II): Diane starts to get interested in empirical research on the social causes of belief formation. She reads the literature cited in this paper, paying special attention to Kahan’s work on politically motivated reasoning. She realizes that politically motivated reasoning is ubiquitous, and even sophisticated subjects like her are susceptible to it. Now, when she thinks about the reasons for which she holds her beliefs about things like climate change, it doesn’t seem to her like she believes on the basis of these reasons because they happen to vindicate her broader political and cultural beliefs. But she has also read the literature on the (un)reliability of introspectionFootnote 27 and “hindsight bias”:Footnote 28 she recognizes that she is not necessarily a good judge of why she thinks as she does, and that, like all of us, she tends to engage in a lot of ex post facto rationalization. She therefore regards it as very possible—if not likely—that she believes on the basis of the reasons she does not because they strike her as good reasons, but because they fit with her broader political and cultural beliefs.

What are we to say about Diane in Enlightened Student (Part II)? There is an error possibility that she regards as “live,” and it is such that, were it to obtain, she would (we have argued in sections 3 and 4) not be doxastically justified in her belief that global temperatures are rising. We can view Diane’s unease as (albeit unconsciously) a response to this situation.

But is there any problem with Diane’s epistemic situation? While we wouldn’t want to say for sure that there is, it is worth noting that there are some influential views in contemporary epistemology on which Diane’s belief may well be unjustified. Take, for instance, Michael Bergmann’s (Reference Bergmann2006) influential account of epistemic defeat, on which (put roughly) it is necessary and sufficient for some condition to defeat S’s belief that S take it that the condition defeat their belief (e.g., by counting against the truth of the belief or against the reliability of its formation). On this account, if Diane takes it that her inability to rule out this error possibility constitutes a defeater for her belief that global temperatures are rising, then her belief is defeated, and she is no longer justified in believing that global temperatures are rising. So if you are attracted to a Bergmann-style account of epistemic defeat, there is room for a worry about whether Diane’s belief is justified.

A similar point can made with reference to Lewis’s (Reference Lewis1996) account of relevant alternatives and their relationship to knowledge attributions. For Lewis, whether it is true to say that some S knows that some p is true depends on whether S can rule out all relevant alternatives, and various things can lead an alternative to be relevant in some conversational context. One thing that is sufficient (but not necessary) for an alternative to be relevant in a context is that it be taken seriously in the context. So, for instance, if a Cartesian skeptical scenario is taken seriously in some context, then it is relevant in that context. A Lewisian contextualist may hold that as Diane moves from Enlightened Student (Part I) to Enlightened Student (Part II) the possibility that she has been engaging in motivated reasoning becomes relevant, and because she can’t rule it out, she can no longer truly say that she knows that global temperatures are rising. It is, of course, a further step to hold that she can no longer truly say that she is justified in believing that global temperatures are rising. But perhaps this further step can be made and, in any case, it would be unsettling enough if Diane couldn’t truly say that she knows global temperatures are rising.Footnote 29

8.

So what is the skeptical import of motivated reasoning? We want to finish by clarifying what we have been trying to do in this paper and by commenting on what all this might teach us about intellectual virtue.

With reference to the preceding discussion, three ascending levels of skeptical import can be identified:

Level 3 is serious indeed. As the Enlightened Student pair of cases indicates, the kind of skeptical threat generated by the prevalence of motivated reasoning might plausibly imperil the epistemic status of our beliefs even when these beliefs are not formed through such reasoning at all. And in this respect, even what look like perfectly “good cases” might be ones where we turn out to be much worse off epistemically than it would initially seem. This should certainly give the epistemologist cause for concern.

To be clear, the above table distinguishes levels of skeptical import—not skeptical conclusions. Compare: it is uncontentious that the skeptical import of Descartes’ arguments in Meditation 1 included at least all empirical propositions. And yet, almost no one accepts the corresponding conclusions of these arguments. Of course, any skeptical argument—no matter the scope of skepticism threatened—might be defused by a suitable anti-skeptical response. In the present case, there is, as we see it, no suitable response to the kind of (minimal) skeptical import raised at Level 1; after all, these are beliefs that we would normally take to be unjustified.

What about Levels 2 and 3? There are two forms of antiskeptical response one might try. The first form would be to try and find a weakness in the details of the skeptical arguments. So far as the Level 2 argument is concerned, one could challenge the empirical data from section 2, or the epistemological assumptions about propositional and doxastic justification relied on in sections 3–4. As for the Level 3 argument, one could challenge either of these things, or the Bergmann-style account of defeat (or the Lewisian semantics for knowledge attributions) from section 7.

The second form would be to target our overall strategy in this paper.Footnote 30 We have argued for skeptical conclusions about a class of our beliefs (beliefs in scientific issues that have become politically contentious, like global warming) on the basis of highly simplified toy cases (Naïve Student and Enlightened Student). We supposed that the causes of Tim’s and Diane’s beliefs can be captured in very simple terms (they defer to people recognized as experts in their community), and we have said very little about the reasons underlying these patterns of deference. One might argue that the causes and reasons underlying the beliefs that actual people have about issues like global warming are far more complex than this and, due to these complexities, the skeptical conclusions we have derived about Tim and Diane may not transfer over to our own beliefs about these issues. We grant this point, but our aim has not been to argue conclusively for either a Level 2 or 3 skeptical conclusion. More modestly, we hope to have shown that the phenomenon of motivated reasoning raises insidious skeptical problems—and accordingly that the epistemological ramifications of motivated reasoning are much more serious than one may initially think (see, e.g., Kelly Reference Kelly2008).

We want to conclude by noting how the above ascending cascade of skeptical threats posed by motivated reasoning might teach us something instructive about intellectual virtue.

One of the most popular mantras in contemporary social epistemology is that intellectual humility is valuable to its possessors as well as valuable more broadly in liberal democracies predicated upon a toleration of differing viewpoints.Footnote 31 One prominent account of the nature of intellectual humility, defended by Dennis Whitcomb, Heather Battaly, Jason Baehr, and Daniel Howard-Snyder (Reference Whitcomb, Battaly, Baehr and Howard-Snyder2017) maintains that the virtue centrally involves an owning of our epistemic limitations. Footnote 32 Intellectual humility, so understood, is characterized by a dispositional profile that includes cognitive, behavioral, motivational, and affective responses to the various ways in which individuals are epistemically limited. But what are these limitations exactly? As Whitcomb et al. write:

We’re all familiar with what we gather under the rubric of “intellectual limitations”: gaps in knowledge (e.g. ignorance of current affairs), cognitive mistakes (e.g. forgetting an appointment), unreliable processes (e.g. bad vision or memory), deficits in learnable skills (e.g. being bad at math), intellectual character flaws (e.g. a tendency to draw hasty inferences) … (Reference Whitcomb, Battaly, Baehr and Howard-Snyder2017, 516).Footnote 33

One feature that unifies the examples that Whitcomb et al. point to is that they are all, in an important respect, individual-level limitations. Individual-level limitations, however, should only be part of the story. Some of our intellectual limitations arise simply in virtue of our social situatedness—viz, our membership in epistemic communities that antecedently value the things that they do. This at any rate seems to be an important moral to draw from sections 2–7, and this is especially so for those who join us in taking seriously the idea that the skeptical import of the kind of motivated reasoning that features in cultural cognition rises to (at least) Level 2.

One lesson to draw from canvassing the skeptical import of motivated reasoning is that there is a distinctively social way that the virtuously intellectually humble person may own her epistemic limitations, apart from the more traditional ways of owning individual-level limitations (e.g., owning knowledge gaps, skills deficits, etc.)Footnote 34 That is, she may own the kinds of epistemic limitations one inherits specifically by embedding oneself in (and reaping other epistemic benefits from) a social community, limitations that—if our argument is sound—stand to threaten us epistemically on potentially three levels of ascending epistemic concern.

J. Adam Carter is a lecturer in philosophy at the University of Glasgow, where he directs the COGITO Epistemology Research Centre. He works mostly in epistemology, and his recent projects have included work on know-how, virtue epistemology, extended epistemology, social epistemology, and relativism.

Robin McKenna is a lecturer in philosophy at the University of Liverpool. Most of his work is in epistemology, but he is also interested in philosophy of language and philosophy of science and ethics. Current topics of interest include the epistemology of persuasion, the epistemology of climate change denial, epistemic injustice, and social constructivism.

Footnotes

1 Exceptions include Avnur and Scott-Kakures (Reference Avnur and Scott‐Kakures2015), Kelly (Reference Kelly2008), Huemer (Reference Huemer, Brennan, Munger and Sayre-McCord2016), and Kornblith (Reference Kornblith1999). Interestingly, in Huemer’s case, many cases of politically motivated reasoning can be assimilated to political irrationality. Huemer reaches this conclusion in the service of attempting to explain widespread, strong, and persistent disagreements in politics. Huemer’s position is that irrationality better explains such disagreements than other candidate explanations, such as miscalculation, ignorance, and divergent values.

2 For a distinct though related challenge, see Carter and Pritchard (Reference Carter and Pritchard2016).

3 This kind of possibility space has been explored in the context of the metaethical debate between realists and antirealists about value by Street (Reference Street2006) (specifically as the second horn of her “Darwinian dilemma”). Thanks to an anonymous referee at the Canadian Journal of Philosophy for drawing this paper to our attention.

4 For similar results, see Bradley (Reference Bradley1978), Ditto et al. (Reference Ditto, Scepansky, Munro, Apanovitch and Lockhart1998), and Kunda (Reference Kunda1990).

5 For similar results see Bullock, Gerber, and Hill (Reference Bullock, Gerber and Hill2015); Jost, Hennes, and Lavine (Reference Jost, Hennes, Lavine and Carlston2013); Redlawsk (Reference Redlawsk2002). For related discussion, see D. Greco (Reference GrecoForthcoming).

6 Note that by political sophistication Taber and Lodge mean “knowledgeableness,” where this is, for the purposes of their study, a function of the proportion of correct answers provided on a survey test (2006, 760).

7 See also Burnstein and Vinokur (Reference Burnstein and Vinokur1977) and Brown (Reference Brown1986) for related discussions on how exposure to more arguments that support an antecedently held position can contribute to polarization within a group by causing individuals to embrace more extreme versions of positions previously held. Related points have been registered by Huemer (Reference Huemer, Brennan, Munger and Sayre-McCord2016) and also Kahan (Reference Kahan2013). In Kahan’s study, it was shown that polarization increased as people got more “numerate.”

8 That said, there is some evidence that financial incentives can minimize the impact of motivated reasoning (Bullock, Gerber, and Hill Reference Bullock, Gerber and Hill2015; Prior, Sood, and Khanna Reference Prior, Sood and Khanna2015), and that it doesn’t impact on experts’ assessments of arguments and evidence that pertain directly to their fields of expertise (Kahan et al. Reference Kahan, Hoffman, Evans, Devins, Lucci and Cheng2016).

9 One interesting such symmetry is registered in a recent study by Gampa et al. (Reference Gampa, Wojcik, Motyl, Nosek and DittoForthcoming). This study found that ideological beliefs (on both sides) impaired people’s ability to do two things that are central to basic political reasoning: (i) recognize logical validity in arguments that oppose their own political beliefs, and (ii) recognize the lack of logical validity in arguments that support their own political beliefs.

10 For relevant discussion, see Ditto et al. (Reference Ditto, Liu, Clark, Wojcik, Chen, Grady, Celniker and Zinger2019) and Baron and Jost (Reference Baron and Jost2019).

11 Note that, at this juncture, we are merely describing politically motivated reasoning.

12 The reader might already be wondering how our argument will relate to the literature on epistemically irrelevant influences on belief (Avnur and Scott‐Kakures Reference Avnur and Scott‐Kakures2015; Cohen Reference Cohen2001; DiPaolo and Simpson Reference DiPaolo and Simpson2016; Mogensen Reference Mogensen2016; Schoenfield Reference Schoenfield2013; Sher Reference Sher2001; Srinivasan Reference Srinivasan2015; Vavova Reference Vavova2018; White Reference White2010). We address this in section 6.

13 Granted, a full specification of their psychological profiles would include not just beliefs, but also desires (e.g., to fit in or to be seen in a certain way). We abstract away from these complications here.

14 For one thing, these beliefs are false. For another, they are plausibly unjustified—or, at least, as unjustified as the beliefs of Goldman’s (Reference Goldman1988) “benighted cognizer.” For discussion, see Simion (Reference Simion2019).

15 For representative statements of this relationship between propositional and doxastic justification, see Alston (Reference Alston1985); Korcz (Reference Korcz2000); Kvanvig (Reference Kvanvig2003); Pollock and Cruz (Reference Pollock and Cruz1999, 35–36); Turri (Reference Turri2010).

16 See Davis (Reference Davis2009) for related discussion.

17 The above diagnosis of Naïve Student-1 also makes mischief for those (like Kallestrup and Pritchard Reference Kallestrup and Pritchard2014, 2012; Lackey Reference Lackey2009, Reference Lackey2007b) who hold that in “friendly” epistemic environments, knowledge acquisition may require very little by way of cognitive effort and skill on the part of the knowledge recipient, and far less than it would involve in less friendly environments.

18 It is worth noting that there is some dissent on this in the literature (e.g., Rini Reference Rini2017). But we are inclined to agree with Worsnip (Reference Worsnip, Fox and Saunders2018), who argues that the strategy of deferring to those in one’s social group may be rational from a prudential point of view, but isn’t from an epistemic point of view.

19 This is a variation on Lackey’s (Reference Lackey2007a, 598) original case.

20 For some notable defences of this kind of position, see, for example, Harman (Reference Harman1970), Swain (Reference Swain1981), Moser (Reference Moser1989), Turri (Reference Turri2011), and McCain (Reference McCain2012).

21 See, for example, essays in Carter and Bondy (Reference Carter and Bondy2020).

22 By way of comparison, consider an analogy with testimony more generally. The process of trusting whatever any speaker says with no sensitivity to potential defeaters, is not a process that issues in doxastic justification via testimony. And this is the case even if one happens to be in a friendly environment. This is a point that both reductionists and nonreductionists about testimony can agree on. For discussion, see, e.g., the debate between Lackey (Reference Lackey2009, 2007b) and Greco (Reference Greco2010) regarding Lackey’s Chicago Visitor case.

23 This raises a well-known data point in debates about the generality problem for reliabilism. Fine-grained reliability lacks epistemic relevance. (Example: guessing is reliable whenever the environment is structured so that the guess is correct.) The kind of reliability that is of importance epistemically is general, or normal, reliability—viz, where the relevant reference class includes normal environments. For different ways of unpacking the kind of reference class that is pertinent to reliability assessments, see Graham (Reference Graham2012) for an etiological functionalist account and Sosa (Reference Sosa2015) for a virtue-theoretic account. See also Lyons (Reference LyonsForthcoming) for a recent proposed solution to the generality problem.

24 Thanks to a referee at the Canadian Journal of Philosophy for suggesting many helpful changes to this section.

25 Henceforth we drop the qualifier “epistemically.”

26 Much the same applies to Avnur and Scott-Kakures (Reference Avnur and Scott‐Kakures2015), who focus on what they call “debunkers.”

27 For references and philosophical discussion, see Schwitzgebel (Reference Schwitzgebel2011).

28 The classic study is Fischhoff and Beyth (Reference Fischhoff and Beyth1975).

29 Another way of approaching this case would be via considerations about higher-order evidence. We could say that, due to her research about the prevalence of bias, Diane ends up with misleading evidence that there is something wrong with the evidential basis of her first-order belief. Whether this has any implications for the justificatory status of Diane’s belief depends on one’s views about higher-order evidence, and we lack the space to address this issue here. But we think this would be an interesting topic for future work. Thanks to a referee for suggesting this angle.

30 We would like to thank a referee for pushing us to consider this response.

31 For an overview of this literature, see Gunn, Johnson, Lynch and Sheff (Reference Gunn, Johnson, Lynch, Sheff and Pritchard2017). For recent work on the importance of intellectual humility in education, see the articles in Carter et al. (Reference Carter, Kallestrup and Pritchard2016). Cf. Gordon (Reference Gordon2018) for discussion of the epistemic value of humility in counselling.

32 For some additional accounts of intellectual humility, see Church and Samuelson (Reference Church and Samuelson2017), Church (Reference Church2017), Kidd (Reference Kidd2016), Kallestrup and Pritchard (Reference Kallestrup and Pritchard2016), and Tanesini (Reference Tanesini2018).

33 In offering these examples, Whitcomb et al. are to some extent borrowing from a passage they cite earlier in the paper from June Tangney (Reference Tangney, Lopez and Snyder2009).

34 It’s worth noting that this line of thinking gains support from views that emphasise the social dimension of reasoning; see, e.g., Mercier and Sperber (Reference Mercier and Sperber2017).

References

Alston, William. 1985. “Concepts of Epistemic Justification.” The Monist 68 (1): 5789.CrossRefGoogle Scholar
Audi, Robert. 1982. “Believing and Affirming.” Mind 91 (361): 115–20.CrossRefGoogle Scholar
Avnur, Yuval, and Scott‐Kakures, Dion. 2015. “How Irrelevant Influences Bias Belief.” Philosophical Perspectives 29 (1): 739.10.1111/phpe.12060CrossRefGoogle Scholar
Baron, Jonathan, and Jost, John T.. 2019. “False Equivalence: Are Liberals and Conservatives in the United States Equally Biased?” Perspectives on Psychological Science 14 (2): 292303.CrossRefGoogle ScholarPubMed
Beebe, James R., Baghramian, Maria, Drury, Luke, and Dellsén, Finnur. 2018. “Divergent Perspectives on Expert Disagreement: Preliminary Evidence from Climate Science, Climate Policy.” https://arxiv.org/ftp/arxiv/papers/1802/1802.01889.pdf.Google Scholar
Bergmann, Michael. 2006. Justification without Awareness. New York: Oxford University Press.CrossRefGoogle Scholar
Bondy, Patrick, and Carter, J. Adam. (2020). "The Superstitious Lawyer's Inference." In Well-Founded Belief: New Essays on the Epistemic Basing Relation, edited by Patrick Bondy and Carter, J. Adam. New York: Routledge.Google Scholar
Bradley, Gifford W. 1978. “Self-Serving Biases in the Attribution Process: A Reexamination of the Fact or Fiction Question.” Journal of Personality and Social Psychology 36 (1): 5671.CrossRefGoogle Scholar
Brown, Roger. 1986. Social Psychology. 2nd ed. New York: Free Press.Google Scholar
Bullock, John G., Gerber, Alan S., and Hill, Seth J.. 2015. “Partisan Bias in Factual Beliefs about Politics.” Quarterly Journal of Political Science 10: 519–78.CrossRefGoogle Scholar
Burnstein, Eugene, and Vinokur, Amiram. 1977. “Persuasive Argumentation and Social Comparison as Determinants of Attitude Polarization.” Journal of Experimental Social Psychology 13 (4): 315–32.CrossRefGoogle Scholar
Carter, J. Adam, and Bondy, Patrick, eds. 2020. Well-Founded Belief: New Essays on the Epistemic Basing Relation. New York: Routledge.Google Scholar
Carter, J. Adam, and Pritchard, Duncan. "Cognitive Bias, Scepticism and Understanding." Explaining Understanding: New Perspectives from Epistemology and Philosophy of Science (2016): 272–91.Google Scholar
Carter, J. Adam, Kallestrup, Jesper, and Pritchard, Duncan. 2016. “Intellectual Humility.” Logos and Episteme 7 (7): 409–11.CrossRefGoogle Scholar
Church, Ian M. 2017. “The Limitations of the Limitations-Owning Account of Intellectual Humility.” Philosophia 45 (3): 1077–84.CrossRefGoogle Scholar
Church, Ian M., and Samuelson, Peter L.. 2017. Intellectual Humility: An Introduction to the Philosophy and Science. London: Bloomsbury Academic.Google Scholar
Cohen, G. A. 2001. If You’re an Egalitarian, How Come You’re So Rich ? Cambridge, MA: Harvard University Press.Google Scholar
Davis, John K. 2009. "Subjectivity, Judgment, and the Basing Relationship." Pacific Philosophical Quarterly 90 (1): 2140.CrossRefGoogle Scholar
DiPaolo, Joshua, and Simpson, Robert Mark. 2016. “Indoctrination Anxiety and the Etiology of Belief.” Synthese 193 (10): 3079–98.CrossRefGoogle Scholar
Ditto, Peter H., Scepansky, James A., Munro, Geoffrey D., Apanovitch, Anne Marie, and Lockhart, Lisa K.. 1998. “Motivated Sensitivity to Preference-Inconsistent Information.” Journal of Personality and Social Psychology 75 (1): 5369.CrossRefGoogle Scholar
Ditto, Peter H., Liu, Brittany S., Clark, Cory J., Wojcik, Sean P., Chen, Eric E., Grady, Rebecca H., Celniker, Jared B., and Zinger, Joanne F.. 2019. “At Least Bias Is Bipartisan: A Meta-Analytic Comparison of Partisan Bias in Liberals and Conservatives.” Perspectives on Psychological Science 14 (2): 273–91.CrossRefGoogle ScholarPubMed
Fischhoff, Baruch, and Beyth, Ruth. 1975. “‘I Knew It Would Happen.’” Organizational Behavior and Human Performance 13 (1): 116.CrossRefGoogle Scholar
Gampa, Anup, Wojcik, Sean, Motyl, Matt, Nosek, Brian, and Ditto, Peter. Forthcoming. “(Ideo)Logical Reasoning: Ideology Impairs Sound Reasoning.” Social Psychological and Personality Science. https://doi.org/10.1177/1948550619829059.CrossRefGoogle Scholar
Gigerenzer, Gerd. 2000. Simple Heuristics That Make Us Smart. Oxford: Oxford University Press.Google Scholar
Ginet, Carl. 1985. “Contra Reliabilism.” The Monist 68 (2): 175–87.CrossRefGoogle Scholar
Goldman, Alvin. 1988. "Strong and Weak justification." Philosophical Perspectives, 2: 5169.CrossRefGoogle Scholar
Gordon, Emma C. 2018. “Intellectual Humility, Spirituality and Counselling.” Journal of Psychology and Theology 49 (4). https://doi.org/10.1177/0091647118807185.Google Scholar
Graham, Peter J. 2012. “Epistemic Entitlement.” Noûs 46 (3): 449–82.CrossRefGoogle Scholar
Greco, Dan. Forthcoming. “Climate Change and Cultural Cognition.Philosophy and Climate Change.Google Scholar
Greco, John. 2010. Achieving Knowledge: A Virtue-Theoretic Account of Epistemic Normativity. Cambridge: Cambridge University Press.CrossRefGoogle Scholar
Gunn, Hanna, Johnson, Casey Rebecca, Lynch, Michael P., and Sheff, Nathan. 2017. “Intellectual Humility.” In Oxford Bibliographies in Philosophy, edited by Pritchard, Duncan. New York: Oxford University Press.Google Scholar
Harman, Gilbert H. 1970. “Knowledge, Reasons, and Causes.” Journal of Philosophy 67 (21): 841–55.CrossRefGoogle Scholar
Hodson, Gordon, and Busseri, Michael. 2012. “Bright Minds and Dark Attitudes: Lower Cognitive Ability Predicts Greater Prejudice through Right-Wing Ideology and Low Intergroup Contact.” Psychological Science 23 (2): 187–95.10.1177/0956797611421206CrossRefGoogle ScholarPubMed
Huemer, Michael. 2016. “Why People Are Irrational about Politic.” In Philosophy, Politics and Economics: An Anthology, edited by Brennan, Geoffrey, Munger, Michael C., and Sayre-McCord, Geoffrey, 456–67. New York: Oxford University Press.Google Scholar
Jost, John T., Hennes, Erin P., and Lavine, Howard. 2013. “‘Hot’ Political Cognition: Its Self-, Group-, and System-Serving Purposes.” In The Oxford Handbook of Social Cognition, edited by Carlston, Dona, 851–75. New York: Oxford University Press.Google Scholar
Kahan, Dan. 2013. “Ideology, Motivated Reasoning, and Cognitive Reflection.” Judgment and Decision Making 8 (4): 407–24.Google Scholar
Kahan, Dan. 2014. “Making Climate-Science Communication Evidence-Based—All the Way Down.” In Culture, Politics and Climate Change, edited by Boykoff, M. and Crow, D., 203–20. New York: Routledge.Google Scholar
Kahan, Dan. 2016. “The Politically Motivated Reasoning Paradigm, Part 1: What Politically Motivated Reasoning Is and How to Measure It.” In Emerging Trends in the Social and Behavioral Sciences, 1–16. https://doi.org/10.1002/9781118900772.etrds0417.CrossRefGoogle Scholar
Kahan, Dan, Hoffman, David, Evans, Danieli, Devins, Neal, Lucci, Eugene, and Cheng, Katherine. 2016. “‘Ideology’ or ‘Situation Sense’? An Experimental Investigation of Motivated Reasoning and Professional Judgment.” University of Pennsylvania Law Review 164 (349): 349438.Google Scholar
Kahan, Dan, Jenkins-Smith, Hank, and Braman, Donald. 2011. “Cultural Cognition of Scientific Consensus.” Journal of Risk Research 14 (2): 147–74.CrossRefGoogle Scholar
Kallestrup, Jesper, and Pritchard, Duncan. 2012. “Robust Virtue Epistemology and Epistemic Anti-Individualism.” Pacific Philosophical Quarterly 93 (1): 84103.CrossRefGoogle Scholar
Kallestrup, Jesper, and Pritchard, Duncan. 2014. “Virtue Epistemology and Epistemic Twin Earth.” European Journal of Philosophy 22 (3): 335–57.10.1111/j.1468-0378.2011.00495.xCrossRefGoogle Scholar
Kallestrup, Jesper, and Pritchard, Duncan. 2016. “From Epistemic Anti-Individualism to Intellectual Humility.” Res Philosophica 93 (3): 533–52.CrossRefGoogle Scholar
Kanazawa, Satoshi. 2010. “Why Liberals and Atheists Are More Intelligent.” Social Psychology Quarterly 73 (1): 3357.CrossRefGoogle Scholar
Kelly, Thomas. 2008. “Disagreement, Dogmatism, and Belief Polarization.” Journal of Philosophy 105 (10): 611–33.CrossRefGoogle Scholar
Kidd, Ian James. 2016. “Intellectual Humility, Confidence, and Argumentation.” Topoi 35 (2): 395402.CrossRefGoogle Scholar
Korcz, Keith. 2000. “The Causal-Doxastic Theory of the Basing Relation.” Canadian Journal of Philosophy 30 (4): 525–50.CrossRefGoogle Scholar
Kornblith, Hilary, 1999. "Distrusting Reason." Midwest Studies in Philosophy, 23 (1): 181–96.CrossRefGoogle Scholar
Kunda, Ziva. 1987. “Motivated Inference: Self-Serving Generation and Evaluation of Causal Theories.” Journal of Personality and Social Psychology 53 (4): 636–47.CrossRefGoogle Scholar
Kunda, Ziva. 1990. “The Case for Motivated Reasoning.” Psychological Bulletin 108 (3): 480.CrossRefGoogle ScholarPubMed
Kvanvig, Jonathan. 2003. “Propositionalism and the Perspectival Character of Justification.” American Philosophical Quarterly 40 (1): 317.Google Scholar
Lackey, Jennifer. 2007a. “Norms of Assertion.” Noûs 41 (4): 594626.CrossRefGoogle Scholar
Lackey, Jennifer. 2007b. “Why We Don’t Deserve Credit for Everything We Know.” Synthese 158 (3): 345–61.CrossRefGoogle Scholar
Lackey, Jennifer. 2009. “Knowledge and Credit.” Philosophical Studies 142 (1): 2742.CrossRefGoogle Scholar
Lewis, David. 1996. “Elusive Knowledge.” Australasian Journal of Philosophy 74 (4): 549–67.CrossRefGoogle Scholar
Lord, Charles G., Ross, Lee, and Lepper, Mark R.. 1979. “Biased Assimiliation and Attitude Polarization: The Effects of Prior Theories on Subsequently Considered Evidence.” Journal of Personality and Social Psychology 37 (11): 2098–109.CrossRefGoogle Scholar
Lynch, Michael P. Forthcoming. “Epistemic Humility, Epistemic Arrogance and Political Conviction.” Episteme.Google Scholar
Lynch, Michael P. 2018. “Epistemic Arrogance and Political Dissent.” In Voicing Dissent: The Ethics and Epistemology of Making Disagreement Public, edited by Johnson, Casey Rebecca, 129–39. London: Routledge.CrossRefGoogle Scholar
Lyons, Jack. Forthcoming. “Algorithm and Parameters: Solving the Generality Problem for Reliabilism.” Philosophical Review.Google Scholar
McCain, Kevin. 2012. “The Interventionist Account of Causation and the Basing Relation.” Philosophical Studies 159 (3): 357–82.CrossRefGoogle Scholar
Mercier, Hugo, and Sperber, Dan. 2017. The Enigma of Reason. Cambridge, MA: Harvard University Press.CrossRefGoogle Scholar
Mogensen, Andreas L. 2016. “Contingency Anxiety and the Epistemology of Disagreement.” Pacific Philosophical Quarterly 97 (4): 590611.10.1111/papq.12099CrossRefGoogle Scholar
Molden, Daniel C., and Higgins, E. Tory. 2012. “Motivated Thinking.” In The Oxford Handbook of Thinking and Reasoning, edited by Holyoak, Keith and Morrison, Robert, 390409.Google Scholar
Moser, Paul. 1989. Knowledge and Evidence. Cambridge: Cambridge University Press.Google Scholar
National Research Council Board on Radioactive Waste Management. 1990. “Rethinking High-Level Radioactive Waste Disposal: A Position Statement of the Board on Radioactive Waste Management.” Washington, DC: National Academy Press.Google Scholar
National Research Council Committee on Analysis of Global Change Assessments. 2007. “Analysis of Global Change Assessments: Lessons Learned.” Washington, DC: National Academies Press.Google Scholar
National Research Council Committee to Improve Research Information and Data on Firearms. 2004. “Firearms and Violence: A Critical Review.” Washington, DC: National Academies Press.Google Scholar
Plantinga, Alvin. 1993. Warrant: The Current Debate. Oxford: Oxford University Press.CrossRefGoogle Scholar
Pollock, John L., and Cruz, Joseph. 1999. Contemporary Theories of Knowledge. Lanham, MA: Rowman & Littlefield.Google Scholar
Prior, Markus, Sood, Gaurav, and Khanna, Kabir. 2015. “You Cannot Be Serious: The Impact of Accuracy Incentives on Partisan Bias in Reports of Economic Perceptions.” Quarterly Journal of Political Science 10: 489518.CrossRefGoogle Scholar
Redlawsk, David P. 2002. “Hot Cognition or Cool Consideration? Testing the Effects of Motivated Reasoning on Political Decision Making.” The Journal of Politics 64 (4): 1021–44.CrossRefGoogle Scholar
Rini, Regina. 2017. “Fake News and Partisan Epistemology.” Kennedy Institute of Ethics Journal 27 (S2): 4364.CrossRefGoogle Scholar
Ryle, Gilbert. 1945. “Knowing How and Knowing That.” Proceedings of the Aristotelian Society 46: 116.CrossRefGoogle Scholar
Schoenfield, Miriam. 2013. “Permission to Believe: Why Permissivism Is True and What It Tells Us About Irrelevant Influences on Belief.” Noûs 47 (1): 193218.Google Scholar
Schwitzgebel, Eric. 2011. Perplexities of Consciousness. Cambridge, MA: MIT Press.CrossRefGoogle Scholar
Sher, George. 2001. “But I Could Be Wrong.” Social Philosophy and Policy 18 (2): 64.CrossRefGoogle Scholar
Slothuus, Rune, and de Vreese, Claes H.. 2010. “Political Parties, Motivated Reasoning, and Issue Framing Effects.” The Journal of Politics 72 (3): 630–45.CrossRefGoogle Scholar
Simion, Mona. 2019. "Knowledge‐First Functionalism." Philosophical Issues 29 (1): 254–67.CrossRefGoogle Scholar
Sosa, Ernest. 2015. Judgment and Agency. Oxford: Oxford University Press.CrossRefGoogle Scholar
Srinivasan, Amia. 2015. “The Archimedean Urge.” Philosophical Perspectives 29 (1): 325–62.CrossRefGoogle Scholar
Street, Sharon. 2006. “A Darwinian Dilemma for Realist Theories of Value.” Philosophical Studies 127 (1): 109–66.CrossRefGoogle Scholar
Swain, Marshall. 1981. Reasons and Knowledge. Ithaca, NY: Cornell University Press.Google Scholar
Taber, Charles S., and Lodge, Milton. 2006. “Motivated Skepticism in the Evaluation of Political Beliefs.” American Journal of Political Science 50 (3): 755–69.CrossRefGoogle Scholar
Tanesini, Alessandra. 2018. “Intellectual Humility as Attitude.” Philosophy and Phenomenological Research 96 (2): 399420.CrossRefGoogle Scholar
Tangney, June P. 2009. “Humility.” In Oxford Handbook of Positive Psychology, edited by Lopez, Shane P. and Snyder, C. R., 483–90. New York: Oxford University Press.Google Scholar
Turri, John. 2010. “On the Relationship between Propositional and Doxastic Justification.” Philosophy and Phenomenological Research 80 (2): 312–26.CrossRefGoogle Scholar
Turri, John. 2011. “Believing For a Reason.” Erkenntnis 74 (3): 383–97.CrossRefGoogle Scholar
Vavova, Katia. 2018. “Irrelevant Influences.” Philosophy and Phenomenological Research 96 (1): 134–52.CrossRefGoogle Scholar
Whitcomb, Dennis, Battaly, Heather, Baehr, Jason, and Howard-Snyder, Daniel. 2017. “Intellectual Humility: Owning Our Limitations.” Philosophy and Phenomenological Research 94 (3): 509–39.10.1111/phpr.12228CrossRefGoogle Scholar
White, Roger. 2010. “You Just Believe That Because …” Philosophical Perspectives 24 (1): 573615.CrossRefGoogle Scholar
Worsnip, Alex. 2018. “The Obligation to Diversify One’s Sources: Against Epistemic Partisanship in the Consumption of News Media.” In Media Ethics: Free Speech and the Requirements of Democracy, edited by Fox, Carl and Saunders, Joe. Abingdon, UK: Routledge.Google Scholar