Chapters 11 and 12 canvassed the main twentieth-century alternatives to the method a priori. None shows how economists can rationally commit themselves to a highly inexact science such as economics. Each of the alternatives runs into internal philosophical difficulties, and (except Koopmans) each implies drastic changes in methodological practice.
Perhaps methodological practice in economics was due for a major overhaul. Many writers on methodology, including myself in the first edition, called for a greater engagement with the details of actual economic relations among individuals and firms. Let us look again at Mill’s deductive method and the criticisms it faced to see what is defensible and what is mistaken.
As I said before, Mill’s inexact deductive method has been subject to logical, methodological, and practical criticisms:
1. The logical criticism, which one finds in Hutchison and to some extent in Samuelson, maintains that inexact (ceteris paribus) laws are scientifically illegitimate, because they are meaningless or unfalsifiable. But the arguments in Chapters 9 and 10 show that qualified claims are not meaningless or untestable and, as argued in Section 12.2, no interesting scientific claims are logically falsifiable.
2. The methodological criticism of Mill’s inexact deductive method, which one finds in Popperians such as Mark Blaug, is that it is too dogmatic: it rules out the possibility of disconfirming the basic “laws.” Adhering to Mill’s inexact deductive method thus, it is alleged, impedes the progress of economics and leads to ad hoc responses to apparent disconfirmation characteristic of a degenerating research program. I accept this criticism of the method, but not generally of economists, who, despite appearances, rarely adhere to it.Footnote 1
3. Furthermore, methodological vice is alleged to lead to practical impotence by authors such as Hutchison and, from a different perspective, Friedman. Even if the inexact laws and the other statements needed to deduce a prediction are true, the unspecified ceteris paribus clauses mean that the prediction follows only if there are no interferences. Since these ceteris paribus qualifications are vague, it is hard to know when they are satisfied. If economists do not know when they are satisfied, then economic theory is of little use in practice. For practical purposes, economists need to know what will happen if a policy is instituted, not what would happen if there were no disturbing causes. Even if Milton Friedman’s views are mistaken, at least he is concerned about when theories actually work. Does not the reliance on the deductive method render economics useless?
Although the methodological rules of Mill’s inexact method a priori, as summarized at the end of Chapter 10, cannot be defended as stated, I shall nevertheless defend (for the most part) the existing practices of theory assessment among economists. These practices appear to follow Mill’s inexact deductive method, but they are, I contend, also consistent with standard methods of theory appraisal in the special circumstances with which economists have to cope. Although apparent Millians in practice, economists can be good Bayesians or hypothetico-deductivists in principle.Footnote 2 After demonstrating this possibility in Sections 13.1 and 13.2, I sketch the method of theory appraisal economists typically employ, and I discuss the large and legitimate role of pragmatic factors in economic theory choice. Only then does this chapter discuss the practical objection. The chapter concludes by pointing to the real source of dogmatism in economics.
13.1 Apparent Dogmatism and the Weak-Link Principle
To maintain that economists should never attribute apparent disconfirmations to shortcomings in their theories is unjustifiable. To this extent, the critics of the deductive method are correct. To follow such a rule would preclude theoretical and empirical progress. Such a rule is objectionably dogmatic.
It looks as if economists are adhering to this rule, but, given the tasks and difficulties economists face, widely accepted theories of confirmation, such as the Bayesian and hypothetico-deductivist views, recommend confirmational practice that is largely indistinguishable from what Mill’s inexact method a priori recommends. The methods of theory appraisal economists employ may be defensible, even though the method a priori is indefensible, and economists appear to conform to it.
It is not unacceptably dogmatic to refuse to find disconfirmation of economic “laws” in typical failures of their market predictions. When the anomalies are those cast up by largely uncontrolled observation of complicated market phenomena, it may be more rational to pin the blame on some of the many disturbing causes, which are always present. Since the confidence of economists in the simplifications and ceteris paribus assumptions necessary to apply economic theory to actual market phenomena will generally be much lower than their confidence in the basic laws, the more likely explanation for the apparent disconfirmation will usually be a failure of the simplifications and ceteris paribus assumptions. In consequence, little can be learned about the purported laws from such observations, but the failure will lie in the difficulties of the task, not in methodological mistake. The possibility of discovering errors in the “laws” of equilibrium theory may be foreclosed by the inadequacies in the data and limitations in economic knowledge, not by unjustifiable methodological fiat.
In responding this way to apparent disconfirmation, economists are implicitly relying on what I call “the weak-link principle”:Footnote 3
The weak-link principle: when a false conclusion depends on multiple premises, attribute the mistake to the most questionable of the premises.
This is but one of many principles that economists might use in modifying a model so that it conforms to some observation and serves as a reliable predictive tool. Scientists and nonscientists alike use the weak-link principle, and (pace Popper) it is rationally justifiable to do so. If either
or
is false and the subjective probability of
is less than that of
, then (other things being equal) it is more likely that
is false than that
is. However, the weak-link principle is neither inviolable nor always appropriate. Details concerning the model’s failures might point to a different premise as the culprit.
Since the simplifications and ceteris paribus clauses needed to derive predictions concerning uncontrolled market phenomena from equilibrium theory are the weak links, mistaken predictions rarely disconfirm the theory. Hence, one can see why Mill’s views seemed so plausible and were so easily refuted, and yet why methodological practice apparently continues to conform to them.
Given their subject matter, economists are bound to look like followers of Mill’s inexact deductive method. Powerful tests require either experimentation, with its possibilities of intervention and control, a great deal of knowledge, or good fortune in finding natural experiments, and without such tests (or superior alternatives) it would be irrational to react to apparent disconfirmations by surrendering credible hypotheses with great pragmatic attractions. If economists could do experiments easily, then they could control for disturbances and avoid the complexity of the phenomena with which they are presented nonexperimentally. If they knew enough, they could exert much the same control even if experiments are not possible. If they were blessed with a comparatively simple set of phenomena such as those of celestial motion, then neither the inability to experiment nor the paucity of their knowledge would be crippling. But the combination of these handicaps makes knowledge of economic phenomena hard to garner.
Limitations in the ability to test can make the basic “laws” of economics de facto unfalsifiable, even if economists were explicitly employing a Bayesian account of confirmation (§§10.1 and A.7). Let
be either a “law” of equilibrium theory or a conjunction of such laws and
be the conjunction of all the other statements needed to derive a prediction,
from
. The prior probability of
,
is much larger than the prior probability of
,
. In the case of uncontrolled market predictions,
will be tiny. For each of the simplifications and ceteris paribus qualifications is improbable, and the probability of the conjunction will be much smaller than that of the separate conjuncts. If
is deducible from
, then
, but that does not imply that
is large. Given how weakly evidence bears on
, the credible “laws” with which economists begin will be de facto nonfalsifiable.
13.2 Are Economists Too Dogmatic?
The deficiencies of market data coupled with the weak-link principle will mimic Mill’s inexact method a priori only if economists judge the “laws” of equilibrium theory to be much more probable than the simplifications and ceteris paribus claims that are needed to test them. Given the empirical problems with those “laws,” can such a judgment be defended? And, if it cannot be defended, then are not economists as unjustifiably dogmatic as critics of the method a priori have alleged?
Few economists regard the “laws” of equilibrium theory as proven or obvious truths, although they may take them for granted when trying to predict the outcome of some proposed policy. Only some fancy philosophical footwork permits one to regard these “laws” as true (see Chapter 9), and it is questionable whether they can be regarded as well established (see Chapter 10). Why then do economists takes these behavioral postulates for granted? Why do they make use of them in the face of their apparent disconfirmation? Is it introspection, as Mill maintains, or everyday experience, as in Robbins’ view, or are these assumptions implicit in the very concept of action, as has been maintained by Austrian theorists such as von Mises (Reference Mises1978, p. 8)?
A full answer would take us back to the discussion of the theoretical strategy of economics in Chapter 7 or forward to the conclusion of this chapter. Everyday experience and introspection are sufficient to establish that some of these laws, such as diminishing marginal rates of substitution and diminishing returns, are reasonable approximations. Without qualifications and a margin of error, they are false, but with these they are unlikely to lead one astray; and economists have good reason to be committed to them.
Furthermore, each of the laws of equilibrium theory possesses pragmatic virtues, for each plays an important role in making the theory mathematically tractable, consistent, and determinate. Indeed, this is about the only virtue of the postulate of constant returns to scale, to which economists are not nearly as committed. Constant returns to scale figures in many economic theories for essentially mathematical reasons, and because it is hoped that its falsity does not do much harm.
Claims such as acquisitiveness and profit maximization are not such good approximations to the truth as are diminishing returns or diminishing marginal rates of substitution, but neither are they as poorly established as constant returns to scale. And what is the alternative? There is a great deal of truth to them, and their virtues in permitting determinate mathematical formulations are considerable. Firms pursue all sorts of objectives besides profits, and a usable theory that heeds these facts should be more accurate. But the accuracy would be purchased at the cost of complexity, and complications could destroy the normative force equilibrium theory has when it is coupled with minimal benevolence. In such circumstances, pragmatic factors may justifiably be more than empirical tie-breakers. If the empirical benefits of a theory change are small – that is, if (1) a slightly more accurate theory does not serve the often practical purposes of economists appreciably better and (2) economic theorists do not believe that an appreciably more exact or useful economic theory is feasible – then the pragmatic virtues of current theory may be decisive. It may be more sensible to treat other objectives of managers and other behavioral generalizations concerning individuals as disturbing causes that may usually be ignored, even if they are important in particular contexts.Footnote 4
Thus, one finds a combined empirical and pragmatic basis for refusing to regard the basic propositions of equilibrium theory as disconfirmed. Although not necessarily unjustifiably dogmatic, there is a serious risk that economists become so entranced by their models that they overlook anomalies and do not consider alternatives. As Mill so presciently remarked (though only when criticizing his father, not his own work):
We either ought not to pretend to scientific forms, or we ought to study all the determining agencies equally, and endeavour, so far as it can be done, to include all of them within the pale of the science; else we shall infallibly bestow a disproportionate attention upon those which our theory takes into account, while we misestimate the rest, and probably underrate their importance.
In my view, the dogmatism of economists, such as it is, lies in an overly complacent commitment to equilibrium theory and to the theoretical strategy underlying it, not in a mistaken view of theory appraisal.
Because economists have reasonable grounds for judging their basic explanatory and predictive generalizations to be less open to revision than are the simplifications and ceteris paribus claims that are also needed to derive conclusions about market phenomena, they may behave in the way that Mill’s inexact deductive method recommends, without being committed to a dogmatic view of theory appraisal. The apparent dogmatism may be just the result of the good fortune of beginning with a set of plausible generalizations coupled with the bad luck of being unable to perform good tests. Consider these remarks of Robert Lucas (Reference Lucas1980, pp. 710–11):
How is confidence [in the components of models] of this sort earned? This is a question on the answer to which economists are fairly well agreed, yet I cannot recall where I have seen the nature of this agreement articulated. The central idea is that individual responses can be documented relatively cheaply, occasionally by direct experimentation, but more commonly by means of the vast number of well-documented instances of individual reactions to well-specified environmental changes made available “naturally” via censuses, panels, other surveys, and the (inappropriately maligned as “casual empiricism”) method of keeping one’s eyes open …
Notice that, having specified the rules by which interaction occurs in detail, and in a way that introduces no free parameters, the ability to predict individual behavior is nonexperimentally transformed into the ability to predict group behavior.
How can one tell whether economists are committed to Mill’s inexact method a priori or whether they are good Bayesians or hypothetico-deductivists doing the best they can in the face of poor data coupled with informal sources of knowledge? If the only data economists could gather were the results of uncontrolled observations of markets, then we might not be able to find out what the methodological commitments of economists were. But experimentation in economics has never been impossible, and in some experiments going back to the middle of the twentieth century, the auxiliary assumptions – the additional premises needed to derive predictions from equilibrium theory – have been sufficiently strong links that the experimental results could actually disconfirm the theory. By examining how economists have responded in such cases, one can determine whether they are proponents of a dogmatic theory of confirmation or whether their apparent dogmatism in nonexperimental circumstances is a rational response to weak evidence.
13.3 Expected Utility Theory and Its Anomalies
Although I cannot offer a comprehensive survey of how economists have reacted to well-established anomalies, I offer some examples here and in Chapters 14 and 15. Expected utility theory (§1.3) has many of the same axioms as equilibrium theory (including completeness, an axiom I focus on). It is, like parts of equilibrium theory, a theory of rationality, and it is accepted on the same sort of grounds as equilibrium theory. But unlike equilibrium theory it is readily testable. Consequently, it is easier to consider how apparent disconfirmations bear on expected utility theory.
The case for expected utility theory, as for equilibrium theory, seems to rest upon an application of Mill’s inexact method a priori. Consider the following remarks of Daniel Ellsberg:
However, this proposition [that individuals have cardinal expected utility functions], which we will call the Hypothesis on Moral Expectations, has little inherent plausibility. The major feat of von Neumann and Morgenstern is to show that the Hypothesis on Moral Expectations is logically equivalent to the hypothesis that the behavior of given individuals satisfies certain axiomatic restrictions. Since the axioms appear, at first glance, highly “reasonable,” the second hypothesis seems far more intuitively appealing than the equivalent Hypothesis on Moral Expectations. It is thus more likely to be accepted on the basis of casual observation and introspection, although the two hypotheses would both be contradicted by exactly the same observations.
Ellsberg is pointing out that economists sometimes accept theories, such as expected utility theory, because the axioms appear “reasonable.” The credibility of the axioms is largely prior to any testing of predictions derived from the theory.
One cannot regard the axioms of expected utility theory as proven scientific truths, though one may say on their behalf (1) that they are, as Ellsberg notes, “reasonable”; (2) that there is some experimental evidence that confirms them; and (3) that, if people do not conform to the axioms of expected utility theory, then, contrary to observation, they may make fools of themselves.Footnote 5 These grounds provide the axioms with some credibility, and (via the weak-link principle) they provide the theory with an ability to withstand casual falsifications.
Unlike equilibrium theory, expected utility theory is readily testable, and the assumptions necessary to derive predictions from the theory need not always be weak links. Psychologists and decision theorists have shown that human behavior sometimes does not conform to the “laws” of expected utility theory. Some of these anomalies can be explained as the consequence of nonrational disturbing causes – such as the result of minor peculiarities in how people process information or of people’s failure to take small differences in probabilities seriously.Footnote 6 Others are more troubling. Let us see how economists and decision theorists do and should deal with some of these.
13.3.1 The Allais Problem
In the early 1950s, Maurice Allais formulated the problem shown in Table 13.1.Footnote 7 A ball is drawn from an urn containing one red ball, eighty-nine white balls, and ten blue balls. So the probabilities are known. Depending on the color and the choice of
or
in problem I or of
or
in problem II, one receives one of the prizes in the table.
Table 13.1 The Allais problem
Problems | Choices | Pay-offs | ||
|---|---|---|---|---|
Red (1) | White (89) | Blue (10) | ||
I | A | $1 million | $1 million | $1 million |
B | $0 | $1 million | $5 million | |
II | C | $1 million | $0 | $1 million |
D | $0 | $0 | $5 million | |
Many people are inclined to prefer option
to option
in problem I and to prefer option
to option
in problem II. Even the Bayesian statistician, Leonard Savage, was at first so inclined (Savage Reference Savage1972, p. 103). If these choices reflect preferences, then they violate the independence principle, for the only difference between the choice pairs is in the magnitude of the pay-off if a white ball is drawn, which should be irrelevant to the choices because it does not depend on whether
or
in problem I or
or
in problem II is selected. Thus,
should be preferred to
if and only if
is preferred to
Footnote 8 Yet many individuals are unpersuaded. In one view:
In Situation
[problem I], I have a choice between $1,000,000 for certain and a gamble where I might end up with nothing. Why gamble? The small probability of missing the chance of a lifetime to become rich seems very unattractive to me.
In Situation
, there is a good chance that I will end up with nothing no matter what I do. The change [sic] of getting $5,000,000 is almost as good as getting $1,000,000 so I might as well go for the $5,000,000 and choose Gamble 4 [
] over Gamble 3 [
].
If expected utility theory is a correct normative theory of rationality, then this reasoning must be fallacious or irrational.Footnote 9
Allais devised the example as a criticism of the normative adequacy of expected utility theory, not as an empirical refutation. So, even if people do stubbornly choose
in problem I and
in problem II, one can still ask whether these choices are evidence against subjective expected utility theory or evidence of human irrationality. The latter view would be supported, if one could find some obvious sign of irrationality, but there is none. If many people are inclined to choose
over
and
over
, and a variety of thoughtful decision theorists are prepared to defend the rationality of choosing
and
, such as Allais himself, Levi (Reference Levi1986), or Sugden (Reference Sugden1986), then there are significant grounds for questioning the independence condition as a normative condition of rationality.
If one is concerned with independence as a generalization about how people actually choose, then it might seem that it does not matter why people make these choices. The fact that they do is sufficient to show that they do not act in accordance with expected utility theory. But predictions do not follow from expected utility theory all by itself, and the diagnosis of the reasoning responsible for the anomalous decisions may still be important. Since we believe that people’s choices are influenced by their reasoning and that fallacious reasoning is unstable, such a diagnosis remains of the utmost importance. For paradoxes such as Allais’ call for a fundamental modification of expected utility theory as an inexact positive theory of choice behavior only if the choices cannot reasonably be attributed to disturbing causes of secondary importance. There is another regard in which examples such as these may have less force in an empirical critique of expected utility theory than in a challenge to its normative adequacy, for it might be objected that choice problems such as these are unusual and unimportant.
Are people’s choices in the Allais problem evidence against subjective expected utility theory or do they show that there is some disturbing cause? There is little evidence of irrationality, but perhaps the interference might be some further rational factor. One might want to supplement subjective expected utility theory with some further rational tendency counteracting the independence principle. Expected utility theory is falsified only if such a hypothesis is inferior to one which denies rather than merely qualifies one or more of its “laws.” What makes this issue more tractable than those raised by the myriad apparent “falsifications” of equilibrium theory revealed by market data is the possibility of experimentation. Instead of an impenetrable mess in which one can do no better than to hold on to what is independently plausible, one has a partially penetrable mess from which one may learn how to correct or improve what one begins with.
13.3.2 Qualification versus Disconfirmation
One important effect of mitigating the empirical difficulties that stand in the way of testing inexact claims is to make the conceptual difficulties clearer. To patch up an apparently disconfirmed model by changing an auxiliary hypothesis or citing a disturbing cause (whether rational or nonrational) is to change one’s model or applied theory in response to disconfirming evidence. The new applied theory has different empirical consequences than the old. Hence it is wrong to say that those who cite interferences to explain away unfavorable evidence ignore disconfirmations. Perhaps they do not react correctly, but they do react.
Disturbing causes, like all causes, have their (inexact) laws, and to explain away a disconfirmation by citing an interference may not be purely ad hoc, at least in Lakatos’ first two senses (1970, p. 112n). The disturbing cause cited is to be expected in similar circumstances, and the modification has nonvacuous empirical content – although the complexity of the phenomena may make testing impossible. The more general the disturbing cause, the more contentful and less ad hoc is the hypothesis that cites it.
Once one has largely ruled out failures of rationality, the question “does the Allais paradox reveal mistakes in expected utility theory, or does it merely reveal a mistake in some simplification or the influence of some disturbing cause?” turns out to be less straightforward than it might appear. The right question in a well-controlled experimental context is not “is the theory disconfirmed or is there an interference?” but “what should one do about this disconfirmation? Should one add a qualification to the model or narrow the independent specification of its scope (which might in many contexts harmlessly be ignored), or should one revise the model in some more fundamental way?” One cannot draw any sharp line between qualifications and modifications, but one does not need to do so either. In both cases, empirical evidence exerts some control over theory change. The difference is pragmatic: qualifications can often be dropped while modifications leave a permanent mark. The significant question is whether theorists can, for particular purposes in a particular context, ignore the necessary changes and employ the original model.
Another way of grasping the issue would be to ask how economists are supposed to know, in Mill’s terminology, that equilibrium theory has captured the “greater” causes of economic phenomena.Footnote 10 Introspection provides evidence that acquisitiveness is a significant causal factor affecting economic phenomena, but introspection does not give economists good reason to believe that acquisitiveness is a more important cause of economic behavior than, for example, the attitudes toward risk that seem to influence choices in the Allais paradox.
How can economists decide whether a disturbing cause is “major” or “minor” and whether they may justifiably regard expected utility theory as capturing the “greater causes” of choice behavior? The quantitative statistical question “how much of the variation in some dependent variable is due (in the actual complicated circumstances) to a particular independent variable?” is subject to fallible statistical investigation. But Mill’s concern in distinguishing major and minor causes is not simply quantitative. “Major” causes are fundamental and have universal scope, while minor causes are more superficial and have narrower scope. The decision whether to deal with an empirical anomaly by changing one’s model or by citing a disturbing cause is tantamount to the decision of whether to treat the factors mentioned by one’s current model and only those factors as the “major” causes. If expected utility theory leaves out a major cause that is responsible for the Allais paradox, then a serious theory change is called for. If it encompasses all the major causes, then anomalies such as the Allais paradox only call for qualifications, which for many purposes can be ignored. The decision depends on both pragmatic and empirical factors.
In its pragmatic aspect this question demands that economists be clear about both practical and theoretical employments and aspirations for the model. What do they want the model for and what sort of theoretical grasp of the subject matter do they think is possible?Footnote 11 Although this way of thinking is most congenial to instrumentalists, it carries no instrumentalist commitments. Realists can also think about the cognitive jobs they want particular models to do and how well they think such jobs can be done. Some of the pragmatic virtues of the axioms of expected utility theory – of the “laws” of equilibrium theory – have already been mentioned: they lead to a mathematically tractable and determinate theory. But there are other pragmatic virtues, to which I return shortly, which are related to the fact that these are also theories of rationality.
The decision whether to qualify or to modify also hinges on the empirical scope, frequency, and distribution of the apparent disconfirmations experimenters have uncovered. If, for example, the disconfirmations are not very important in the domain that is of the greatest theoretical and practical importance, and economists do not believe that a much better theory is likely to be found (which obviously will depend on what alternatives have been suggested), then it would be reasonable to account for the disconfirmations in terms of “interferences.” If, on the other hand, the qualifications need to be invoked often and economists believe that considerably more exactness is possible, then it would be more reasonable to seek to modify the model fundamentally.
The presence and promise of alternatives also influence theory choices. It is fair to say (following to some extent Lakatos’ views) that what converts anomalies or difficulties such as Allais’ paradox into disconfirming evidence demanding fundamental theory modification is the formulation of alternatives, which accommodate such anomalies within a theory that can do the job done by expected utility theory.
At this point a distinctive element enters the picture about which I have more to say in Section 16.3. For one job that expected utility theory, like ordinal utility theory, does is provide a theory of rationality. Should the fact that utility theory is a theory of rationality affect its empirical appraisal?
The suggestion may seem ludicrous. To argue that utility theory is a good theory of how people actually behave because it is also a plausible approximate theory of how they ought to behave seems like the argument that the moral judgment that people ought not to cheat on their taxes implies that people do not in fact do so. The argument seems to presuppose what is in question, which is whether people behave as (according to this theory) they ought rationally to behave.
This response does not settle the matter. Irrationality can be costly, and the costs of irrational behavior may make it unstable. Although people’s behavior diverges from that predicted by expected utility theory, it may be that there can be no better general theory precisely because of the instability of these divergences. Furthermore, people’s behavior is influenced by its theoretical description. For example, it is plausible that the arguments of economists concerning the advantages of index funds over actively managed mutual funds is a large part of the explanation for why index funds have captured such a large portion of individual investment in equities. There is evidence that students who learn economics learn to conform more closely to utility theory.Footnote 12 The defense of utility theory as a first approximation may be self-supporting, while espousing nonrational theories of choice may be self-defeating. The fact that utility theory is a theory of rationality seems to provide some grounds to believe that it is a correct approximation to how people actually choose.
The fact that utility theory is a theory of rationality may provide pragmatic reasons not to give it up and to attribute anomalies to interferences. There are two pragmatic considerations here. First, the fact that utility theory is a theory of rationality permits explanations in economics to be reason-giving explanations in addition to causal explanations (see §16.3 and §A.9). Explanations in economics justify as well as explain choices, and they consequently depend on the factors that economic agents focus on and find of interest. A different sort of explanation might be more successful empirically, but the costs of severing the links between economics and the concerns of economic agents are significant, and they give economists reason to favor a positive theory of choice that is, like the current theory, also a theory of rational choice.
In addition, rather than changing their theories to conform to how people behave, perhaps economists should try to change the behavior. Those who are unclear on what rationality requires or who are lazy or ineffective in their efforts to conform need reeducation. This educative function of expected utility theory provides a pragmatic reason for accepting it, unless a competitor is much better supported by the evidence or better able to guide choice. I am not proposing that theorists pretend that people behave according to expected utility theory even when they do not do so. But the educative function of a theory of choice gives economists reason to describe the divergences as lapses or interferences and to retain expected utility theory (or some alternative with similar normative force) rather than opting for a nonnormative alternative. This reason may not be decisive, for the empirical advantages of a nonnormative alternative might be overwhelming. But such pragmatic grounds are neither trivial nor irrational.Footnote 13
The similarities between the complex mixture of empirical and pragmatic elements one finds in defenses of the use of equilibrium models and Mill’s inexact method a priori are superficial. What drives most economists to regard interferences as minor disturbing causes is not the manifest truth of the basic axioms, nor any methodological rule prohibiting revisions of them, but the nature of the disconfirmations coupled with the pragmatic attractions of accepted theory.
13.3.3 Incomplete Preferences: Levi’s Alternative
An argument for an alternative to expected utility theory illustrates how this complicated process of theory assessment might work. There are several alternatives to expected utility theory which purport to inherit much of its normative and predictive virtues and to accommodate anomalous examples, of which Allais’ problem is but one instance: regret theory, theories which surrender the independence principle, such as Edward McClennen’s (Reference McClennen, Stigum and Wenstop1983) and Mark Machina’s (Reference Machina1987), theories which surrender completeness, such as Isaac Levi’s (Reference Levi1980), theories which surrender independence and completeness, such as Edward McClennen’s (Reference McClennen1990), and theories that surrender context independence, like the prospect theory of Tversky and Kahneman (Reference Tversky and Kahneman1981), about which I say more later.
In this section, I discuss only one of these – Levi’s proposal – which brings out the methodological points clearly. Levi is a philosopher (one of my teachers), and his views, unlike Machina’s, for example, have little following within economics. However, his work illustrates the issues that arise concerning the structure of rational criticism and theory change. For a case study concerning how mainstream economists behave in the face of apparently disconfirming evidence, see Chapter 14.
As discussed in Section 1.1.2, one questionable axiom of both ordinal utility theory and expected utility theory is completeness or comparability – that among any two options
and
a rational agent will either prefer
to
or
to
or the agent will be indifferent. The axiom appears to be false: individuals are often unable to rank alternatives. The related claim that agents can and should form precise subjective probability judgments, which is required by completeness of preferences over gambles, is similarly dubious.
The standard defense of completeness assumes that choice demonstrates preference. What one chooses is what one prefers. But the standard defense gives rise to spurious intransitivities. The only remaining ground upon which to accept completeness is that it is a reasonable approximation or a harmless idealization that permits the development of a simple and systematic theory of rationality. Levi argues that paradoxes such as Allais’ – as well as a pragmatic perspective on inquiry – suggest that assuming completeness is not harmless.
In Levi’s view, people are often unable to rank options with respect to expected utility, owing to indeterminacies in their utility functions or in their probability judgments. Given this inability, Levi maintains that they ought to suspend judgment, as indeed people often do, rather than making arbitrary presumptions. After screening out those options that are unambiguously inferior with respect to expected utility for any admissible utility function or probability judgment, agents should choose on the basis of secondary criteria such as security. Consider the Allais problem again. Option
, $1 million for sure, obviously beats option
on this criterion of security. It is, however, less obvious that
is more secure than
, although Levi argues that it is.Footnote 14
Levi argues that his account of rational choice, which surrenders ordering (and hence completeness) and sharply distinguishes preference and choice, accounts for a wide range of choice behavior that conflicts with expected utility theory and in which (as in the Allais paradox) many subjects refuse to see the error of their ways. Permitting indeterminacies in probability judgments and utilities is not ad hoc, but is required by a pragmatic theory of inquiry that takes ignorance seriously, and the theory that results is neither normatively nor empirically empty (Seidenfeld et al. Reference Seidenfeld, Schervish and Kadane1987). If these claims are defensible, then Levi has presented a strong case disconfirming completeness.
Levi’s alternative cannot be classified unambiguously as a fundamental theory change, although this is clearly how Levi would classify it (compare Kaplan Reference Kaplan1989). Since expected utility theory is preserved as a special case within Levi’s theory, when one has precise preferences and precise probability judgments, one might plausibly argue that Levi is offering a theory that includes more causal factors and thus supplements rather than replaces expected utility theory. Yet Levi believes that circumstances in which agents can be treated as if they had precise preferences and precise probability judgments are exceptional and that failures of completeness should not be treated as unusual complications.Footnote 15
There was never any question of an empirical proof of completeness. At best it appeared to be a reasonable approximation. Levi argues that these appearances are misleading. The example shows that theories that are regarded as inexact and that are defended by means of what looks like Mill’s inexact deductive method are not immunized against refutation.
13.4 Behavioral Economics and Methodological Changes
One way to determine whether economists are committed to the dogmatic method a priori or whether they are instead handicapped by the complexity of the phenomena coupled with the inability to experiment is to see what happens when economists discover new possibilities for exerting or finding experimental controls. Do they continue to defend their theories from the possibility of disconfirmation, or do they modify them when their implications are not borne out by the data? Do they expand the range of the variables and relations discussed in their theories as they acquire the ability to make testable and significant claims about them?
Over the last half-century, slowly at first but with an increasing pace, economists have discovered new possibilities for both laboratory and field experiments, and they have uncovered ways to interpret changes in policies and other historical events as if they were experimental interventions. Although some economists have argued that much of this experimentation is irrelevant to economics (notably Gul and Pesandorfer 2008), experimentation and historical investigation that identify natural experiments have become standard tools of economics. Fifty years ago, the leading journals published only a tiny fraction of the empirical investigations to be found in the same journals today. Chapter 14 is devoted to an extended case study concerning the reception by economists of psychological investigations purporting to reveal the existence of preference reversals. In that history, one can trace the evolution of the responses of economists to this phenomenon, from attempting to explain it away, to attempting to relegate the problems it suggests to peripheral aspects of economic models, to uncomfortable acceptance of the phenomenon, and then finally to inquiry concerning its causes and the ramifications.
One serious issue raised by critics of behavioral and neuroeconomics,Footnote 16 such as Gul and Pesandorfer, concerns the relevance of behavioral and neurological findings to the questions that economists are concerned with, which focus on the aggregate consequences of individual choices rather than on the idiosyncratic factors that are responsible for preference rankings. The discovery that people are no less cooperative when hungry (Reynolds Reference Reynolds2019) may be very useful to an employee approaching her boss for a raise before lunch, but it doesn’t bear significantly on our understanding of economic fluctuations or the properties of markets.
Although a fair criticism of some work in behavioral and neuroeconomics, these qualms do not apply to all of it. Consider, for example, anchoring, loss aversion, framing, and the endowment effect. Many experiments show that individuals “anchor” their evaluation of alternatives to a reference state of affairs, typically the status quo, and they assess gains and losses from that reference point differently. People weigh losses more heavily than gains with both marginal losses and marginal gains of diminishing importance. People’s choices reflect a value function shaped like the one shown in Figure 13.1.

Figure 13.1 Anchoring and loss aversion.
Like a standard preference ranking, this value function is supposed to determine choices, but unlike a preference ranking, it depends on the reference point, and for that reason it is obviously context dependent. Contextual factors that affect people’s reference point, which are irrelevant to a preference ranking that satisfies the standard axioms, can dramatically shift the rankings of alternatives. Tversky and Kahneman (Reference Tversky and Kahneman1981) present the following striking example, which is by now very well known. They asked groups of subjects to assess alternative policies to treat a disease that threatens the lives of 600 people. When given the choice between programs A and B in Table 13.2, about three-quarters of experimental subjects prefer program A.
Table 13.2 Saving lives
Program A | Saves 200 people |
Program B | Saves 600 people with probability Saves no one with probability |
Other experimental subjects were asked to compare Programs C and D in Table 13.3. Faced with a choice between C and D, about three-quarters of experimental subjects prefer program D. Yet A and C are descriptions of the same state of affairs, as are B and D. Anchoring and loss aversion explain this framing effect. The same outcomes look different depending on whether the reference point is the death of 600 people (and one’s action is seen as saving 200 for sure or 600 with a probability of one-third) or whether the reference point has all 600 living with policies permitting the deaths of different numbers.
Table 13.3 Allowing deaths
Program C | 400 people die |
Program D | No one dies with probability 600 people die with probability |
Because people weigh gains and losses differently, their choices depend on the reference point. Rather than unsystematic failures of the axioms, these data identify an additional systematic influence on choice, which is relevant to the behavior of consumers and firms. As Kahneman et al. put it:
It is in the nature of economic anomalies that they violate standard theory. The next question is what to do about it. In many cases there is no obvious way to amend the theory to fit the facts, either because too little is known, or because the changes would greatly increase the complexity of the theory and reduce its predictive yield. The anomalies that we have described… may be an exceptional case, where the needed amendments in the theory are both obvious and tractable. The amendments are not trivial: the important notion of a stable preference order must be abandoned in favor of a preference order that depends on the current reference level. A revised version of preference theory would assign a special role to the status quo, giving up some standard assumptions of stability, symmetry and reversibility which the data have shown to be false. But the task is manageable.
The crucial points are, first, that these anomalies show that the influence of the diverse factors that motivate people cannot be summarized in a single ranking that is complete, transitive, context independent, and choice determining; and second, that some of the divergences are systematic and predictable. The preferences of consumers and the ranking of investments by firms are likely to depend significantly and regularly on the reference point from which individuals evaluate alternatives.
Loss aversion is also manifest in the so-called endowment effect (Kahneman et al. Reference Kahneman, Knetsch and Thaler1991) whereby individuals demand more to part with some commodity than they would have been willing to pay to acquire it. Although the effect was demonstrated with students and coffee mugs, the phenomenon is far from trivial. As Benjamin Friedman (Reference Friedman2005) argues, loss aversion and the endowment effect are among the factors that make redistributive policies so much more difficult to implement when there is little economic growth. When the economy is growing, redistribution will look to those who lose out to be a lesser improvement rather than a loss.
The newly discovered factors that affect choice behavior derive from a blossoming of empirical economics. This takes the forms of laboratory experimentation, field experimentation, and the exploitation of natural experiments, particularly with the assistance of instrumental variable techniques. Empirical investigation of economic phenomena has always faced serious problems. Economists can hardly cast some individuals into poverty to observe the effects of an unexpected loss of wealth on a random sample. It would be unethical and scarcely feasible. Giving some random sample of the population $1,000,000 to observe the effects on consumption might be ethical, but it would be prohibitively expensive. Controlled experiments on participants in realistic markets are difficult to carry out, given how many factors there are to control for. When controlled experiments are possible, economists face the huge (“external validity”) problem of determining whether the findings in the controlled environment of the experiment will hold true outside of the laboratory.
These problems are ineliminable, but economists have found ways of lessening them. Modeling interactions between firms or consumers as simple games has made it possible to simulate economic problems facing market participants by interactions among students and other experimental subjects communicating via computer terminals. Shifting from explicit controls to the control exercised by randomization has made it easier to investigate specific putative causal factors and has lessened (although certainly not eliminated) the problems of external validity.
Randomized control trials are especially useful in field experiments, where explicit controls are infeasible, but field experiments raise special problems of external validity, because they are necessarily carried out within a specific and typically not-well-understood social context. For example, the Tamil Nadu Integrated Nutrition Project (TINP) provided nutritional education to pregnant women in rural districts of the state of Tamil Nadu and apparently greatly lessened malnutrition among infants. Encouraged by these results, policy-makers instituted a similar program in Bangladesh, where it failed. The explanation lies in a difference in an additional causal factor: in Bangladesh, mothers-in-law distribute the family’s food, while in Tamil Nadu, mothers do (Cartwright and Hardie Reference Cartwright and Hardie2012).
Although the TINP was not a randomized control trial, it illustrates the difficulties of extrapolating a finding from one context to another. Moreover, since not all of the districts in Tamil Nadu received this assistance, one might conjecture that the receipt of the assistance was close enough to an experimental intervention that the results could be attributed to the intervention rather than to unknown confounding factors. Such conjectures can be more or less reasonable, and sometimes the implementation of policies can be sequenced so as to create an experiment. For example, Mexico’s PROGRESA (Programa de Educación, Salud y Alimentación), which is a comprehensive antipoverty program combining health care, education, and nutrition, was implemented gradually, with the first locations chosen randomly and the outcomes compared to locations in which PROGRESA had not yet been implemented. Although there is always some possibility that the districts in which PROGRESA was first introduced differed in some other relevant regard from districts in which PROGRESA had not yet been implemented, such a study should count as a randomized control trial. The fact that the experiment is part of the implementation of a policy may create a risk of bias, but it is otherwise irrelevant.
Natural experiments are not easy to find, because of the importance of confounders, but sometimes there are ways around the problems. For example, consider Angrist’s (Reference Angrist1990) study of the effects of serving in the Vietnam War on lifetime earnings. The possibility that young men from less affluent families were more likely to enlist rules out any simple attribution of the lesser earnings of veterans to having served in the military. But an individual’s number in the draft lottery, which determined whether the individual was “draft eligible,” counts as an intervention. Because draft eligibility is determined by a random process, one can attribute the earnings differences between those who were and those who were not draft eligible to the difference in draft eligibility. (Of course, it is possible by chance that those whose numbers in the draft lottery made them eligible to be drafted all happened to be less valuable employees, but this is enormously improbable.) Knowing the effect of draft eligibility on lifetime earnings, Angrist could then use information concerning the proportions of draft-eligible and draft-ineligible men who wound up serving in the military to reach a conclusion concerning the effect of military service on lifetime earnings.Footnote 17 Draft eligibility serves as an instrument that permits inferences concerning the effects on earnings of military service.
In explicitly causal language, one can explain the definition and use of instrumental variables as follows. If economists want to determine whether
causes
, in circumstances in which there are likely to be unmeasured confounders, they cannot draw any causal conclusions from observing a correlation between
and
. But if there is some other variable
that is (1) a cause of
, that is (2) independent of any confounders of the relationship (if any) between
and
and (3) has no causal connection to
unless there is a causal path from
to
via
, then
counts as an instrumental variable, and the discovery of a correlation between
and
is strong evidence that
causes
.

Figure 13.2 An instrumental variable.
In Figure 13.2, economists cannot determine whether
causes
from information concerning whether
and
are correlated, owing to the possibility of unknown confounders,
. But if there is a cause of
, such as
, that has no causal relationship to
apart from a causal relationship that it might have in virtue of causing
, then the discovery of a correlation between
and
is evidence that
causes
. The causal inferences economists can draw on the basis of measuring the correlation between
and
presuppose causal knowledge concerning the relationship between
and the other variables. For example, Angrist’s conclusions rest on the premises that draft eligibility does not affect lifetime earnings, except via its effects on military service, that draft eligibility increases the probability of military service, and that draft eligibility is not caused by another of the other variables including possible confounders.
Although the ethical and feasibility constraints on economic experimentation remain, economists are now able and willing to confront economic hypotheses with evidence. For example, instead of concluding (plausibly) that an increase in the minimum wage increases unemployment among unskilled workers, because firms can substitute machinery or more highly skilled workers for the now more expensive unskilled labor, economists have taken advantage of natural experiments to discover that relatively small increases in the minimum wage have little effect on the employment of unskilled workers (Card and Kreuger Reference Card and Kreuger1994).
13.5 The Economists’ Deductive Method
We are now in a position to formulate a schema sketching a “deductive” method of theory appraisal that is both justifiable and consistent with existing theoretical practice in economics, insofar as that practice aims to appraise theories empirically. For, as I stressed earlier (§§6.4, 7.3, and 7.4), a good deal of theoretical work in economics is still concerned with conceptual exploration, not with empirical theorizing.
To facilitate the comparison of what I am calling the economists’ deductive method with Mill’s inexact method a priori, I have juxtaposed sketches of the two methods in Table 13.4.
Table 13.4 Deductive methods
Mill’s inexact deductive method | Economist’s deductive method |
|---|---|
Borrow proven (ceteris paribus) laws concerning the operation of relevant causal factors | Formulate credible (ceteris paribus) and pragmatically convenient generalizations concerning the operation of relevant causal factors |
Deduce from these laws and statements of initial conditions, simplifications, etc., predictions concerning relevant phenomena | Deduce from these generalizations, and statements of initial conditions, simplifications, etc., predictions concerning relevant phenomena |
Test the predictions | Test the predictions |
If the predictions are correct, then regard the whole amalgam as confirmed. If the predictions are not correct, then judge (1) whether there is any mistake in the deduction, (2) what sort of interferences occurred, and (3) how central the borrowed laws are (how major the causal factors they identify are) and whether the set of borrowed laws should be expanded or contracted | If the predictions are correct, then regard the whole amalgam as confirmed; if the predictions are not correct, then compare alternative accounts of the failure on the basis of explanatory success, empirical progress, and pragmatic usefulness |
The economist’s deductive method, unlike the inexact method a priori, is consistent with standard views of confirmation. What justifies continuing to call it a deductive method, despite its concessions that the inexact laws with which one begins are not proven and that they can be refuted by economic evidence? First (in sharp contrast to the methodological views discussed in Chapters 11 and 12), independent direct confirmation of the basic inexact laws plays a crucial role.Footnote 18 Second, refutation is largely proscribed, albeit by the circumstances, not by methodological rule. Since economists are typically dealing with complex phenomena in which many simplifications are required and in which interferences are to be expected, the evidential weight of predictive failure will be small. It will rarely be rational to surrender a well-supported hypothesis because of a predictive failure in circumstances such as these. The Allais problem exaggerates the weight of evidence because of its quasi-experimental basis.
The simplified account of the economist’s deductive method sketched earlier follows the hypothetico-deductive method (§10.1) precisely in steps 2 and 3 and is consistent with it in steps 1 and 4, where it is merely more specific. The hypothetico-deductive method is mute on where hypotheses to be tested come from and permits one to begin with a theory with known empirical and pragmatic virtues.
The fourth step of the economist’s deductive method abbreviates Section 13.3.2. It is consistent with the hypothetico-deductive method, which merely requires that the correctness or incorrectness of the predictions contribute to the appraisal of the hypothesis tested. The empirical grounds for discriminating between theories in the economist’s deductive method remind one of Lakatos’ formulations, and they direct one to consider what theory modifications or qualifications best explain the data and best increase the confirmed empirical content of the theory. It will be extremely difficult to judge theory modifications on empirical grounds, because of the acute practical Duhem–Quine problem in economics (§A.7), which is a consequence of how dubious are the various auxiliary hypotheses that are necessary in order to perform most tests. Pragmatic grounds may consequently play a large role. If one cannot tell which theory modification is empirically better, it is sensible to choose the one that has greater pragmatic virtues – that is, the one that it is easier to use, gives sharper advice, lends itself to cleaner mathematical expression, and so forth.Footnote 19
Yet, when experiments are possible and when alternatives are available that inherit the initial credibility of the accepted theory and offer similar pragmatic advantages, then the economist’s deductive method favors theory change. If one studies how economists respond to experimental anomalies, one can see that they are not committed to a dogmatic view of confirmation, such as the inexact method a priori. The case study in Chapter 14 provides evidence for this claim. The dogmatism one used to find in economists’ responses to anomalies, which resulted more from a commitment to an image of economics as a separate science than from any theory of confirmation, has not disappeared, but it has faded. Many modifications are proposed, discussed, and tested.
13.6 The Deductive Method and the Demands of Policy
The one remaining criticism of the deductive method is practical: in following a deductive method, economists allegedly condemn their work to practical futility. This criticism is mistaken. The economist’s deductive method does not rule out theory changes when doing so will increase the empirical content of the theory – on the contrary, it mandates them. Nor does it – or any other variety of the deductive method – condemn empirical generalization. Mill is explicit in endorsing common sense on this point: if something works, use it (though with due caution). Moreover, the development of empirical generalizations, for which no deductive derivation is currently possible, is of great theoretical importance too, for such generalizations constitute the most important data for which theories need to account. There are, as we saw in Section 7.5, methodological rules against employing ad hoc generalizations – that is, generalizations that do not permit rational choice explanations, that do not give pride of place to consumerism, that have narrow scope, or that rule out the possibility of equilibrium. But these result from the vision of economics as a separate science, not from distinctive views of theory appraisal.
The economist’s deductive method does not recommend repudiating useful empirical generalizations or abandoning accurate predictive devices. Instead, it condemns naive reliance on unreliable empirical generalizations, and it offers an additional means of getting a predictive grasp on the phenomena. Whether the best way to aim an artillery piece is by firing it in various circumstances and fitting a curve to the data points or by calculating from fundamental laws is an empirical question. Rather than forbidding the first procedure, the deductive method offers a way of improving, correcting, and extending the results one gets by it.
If the standard theory of the firm had all the empirical virtues claimed for it by Milton Friedman and others, then one should make use of it for relevant practical purposes. The economist’s deductive method does not recommend theoretical purism that spurns useful tools that are not in perfect condition or perfectly understood. By considering the realism of a theory’s assumptions – the constituent causal processes and their laws – one may be able to get some guidance concerning when the predictions of the theory are likely to break down and concerning how to modify the theory in the face of apparent disconfirmation.
13.7 Conclusion: Economics as a Decreasingly Separate Science
What may stand in the way of developing generalizations that are of practical utility is not the deductive method per se but Mill’s vision of economics as a separate science: as a discipline that is concerned with a domain in which a small number of causal factors predominate. I argued in Chapter 7 that this vision of economics as a separate science, although not often expressed in this terminology, remains central to contemporary microeconomics. Mainstream microeconomics, macroeconomics, and general equilibrium theory presuppose that a single set of causal factors underlies economic phenomena and determines their broad features. Other relevant causal factors are countenanced typically only as disturbing causes, whose influence must be acknowledged, but which do not form a part of the central theory. Their effects are allegedly significant with respect to a narrow range of cases; while, without its many specific qualifications, the basic theory is still purportedly a good general guide. As already noted, Mill makes such a pragmatic case:
[T]he ascertainment of the effect due to the one class of circumstances alone is a sufficiently intricate and difficult business to make it expedient to perform it once for all, and then allow for the effect of the modifying circumstances; especially as certain fixed combinations of the former are apt to recur often, in conjunction with every-varying circumstances of the latter class.
To surrender the understanding of economics as a separate science would be to part with the grand vision that a single theory could provide one with a basic grasp of the subject matter. A considerable number of economists are now willing to pay the price, and they have paired up with psychologists, political scientists, neurologists, and much less frequently with sociologists and anthropologists. However, the temper and character of modern economics still embodies the Millian vision of the discipline as a separate science.
Can one better understand economies by applying equilibrium theory, or would economists do better to develop a variety of different theories with narrower domains and a larger repertory of causes? The latter alternative would lower the barriers between economics and other social sciences, since the causal factors with which sociologists and psychologists have been concerned may be important in particular economic subdomains. Although the question is an empirical one, the answer also depends on the objectives and uses of economic theories. For a separate science of economics has aesthetic appeal, heuristic power, and normative force, none of which economists will willingly sacrifice unless the more fragmented and less purely “economic” alternatives have similar virtues and fit the data much better. So long as the data consist of noisy economic statistics, I doubt that the sacrifice will often appear worthwhile.
But, with the development of experimental economics and with increasingly sophisticated field research, this situation is changing; and if twenty-five years from now I undertake a third edition of this book, its title may no longer be appropriate. The central economic theories may have more structural similarities to the sorts of theories favored by institutionalist economistsFootnote 20 than to contemporary microeconomics, macroeconomics, behavioral economics, and general equilibrium theory. But I have no crystal ball, and it may be impossible to generate additional significant theories that provide any appreciably better grip on the data than do contemporary mainstream theories. In that case, economics will go on as it has; and critics may continue to complain that economists are not behaving as responsible scientists should. But before criticizing prematurely, they should recognize that the apparent dogmatism can arise from the circumstances in which economists find themselves – blessed with behavioral postulates that are plausible, powerful, and convenient, and cursed with the inability to learn much from experience.
Economists are committed to equilibrium theory because they regard its basic laws as credible and as possessing heuristic and pragmatic virtues. Their response to anomalous market data, which mimics the inexact method a priori, is not illegitimately dogmatic. It is, on the contrary, consistent with standard views of theory assessment, once one takes account of how bad these data are. The problem is not a moral failing among economists – their inability to live up to their Popperian convictions – but a reflection of how hard it is to learn about complex phenomena if one does not know a great deal already and can do few controlled experiments.



