Misinformation in Experimental Political Science

The American Political Science Association recently cautioned against the use of misinformation (giving research participants false information about the state of the world) in research with human subjects. This recommendation signals a growing recognition, as experimental research itself grows in prevalence in political science, that deceptive practices pose ethical problems. But what is wrong with misinformation in particular? I argue that while this question certainly has an ethical dimension, misinformation is bad for inference too. Misinformation moves us away from answering questions about the political world effectively. I propose a straightforward, intuitive solution to this twofold problem: tell the truth.

The American Political Science Association recently cautioned against the use of misinformation (giving research participants false information about the state of the world) in research with human subjects.This recommendation signals a growing recognition, as experimental research itself grows in prevalence in political science, that deceptive practices pose ethical problems.But what is wrong with misinformation in particular?I argue that while this question certainly has an ethical dimension, misinformation is bad for inference too.Misinformation moves us away from answering questions about the political world effectively.I propose a straightforward, intuitive solution to this twofold problem: tell the truth.
When I pretend or engage in make-believe, I close my eyes to the world around me, sometimes literally, the better to imagine a world that isn't actually there.-Agnes Callard, Aspiration: The Agency of Becoming (2018, 84) R ecent decades have seen political scientists increasingly adopt experimental approaches to answering causal questions about politics (Druckman and Green 2021).Such questions are, by their nature, counterfactual: their causal logic depends on the idea that things could be otherwise, if some causal force did or did not operate (Imbens and Rubin 2015).In cases where the causal force takes the form of an informational intervention-which, in political science, it often does-this counterfactual logic can seem to imply that we need to make up some (counterfactual) information and see whether things would be otherwise if that information were true.This intuition can lead researchers down the path of misinformation.According to new guidelines, published in 2020 by the American Political Science Association's (APSA) Ad Hoc Committee on Human Subjects Research (ACHSR 2020, 7), misinformation is a type of "deception" in which researchers give participants "false information about the state of the world."The guidelines advise researchers to "carefully consider" any use of misinformation.This paper does exactly that.I argue that misinformation raises not only ethical, but also inferential concerns, and researchers would be better off telling the truth.

What Is Misinformation?
Misinformation is the provision of false information about the state of the world (ACHSR 2020, 7).Two criteria must therefore be met for an experiment to involve misinformation: false information must be provided, and that information must be about the state of the world. 1 In principle, researchers could employ misinformation for any purpose, but it is safe to assume that political scientists have little interest in carelessly feeding research subjects random pieces of misinformation for no reason -a practice that would be akin to "bullshit" (Frankfurt 1998).Rather, researchers use misinformation because they are interested in the effect of the misinformation itself.The misinformation is the experimental treatment.
Imagine we are interested in learning about the bandwagon effect (Barnfield 2020): the idea that the popularity of a political candidate or party makes people more likely to vote for them.We conduct an experiment: we show a random selection of people the results of a vote intention poll (the treatment) and test whether they are more likely to vote for a more popular candidate than people who do not see the information.This structure is common across political science experiments that "concern how people react to the content and format of information presented to them" (Dafoe, Zhang and Caughey 2018, 400). 2 We want freedom to manipulate this information-here, the different candidates' vote shares-to ensure that we target Matthew Barnfield (UK) is a Postdoctoral Research Fellow in the Department of Politics at the University of Exeter, UK (m.g.barnfield@gmail.com).He began this project as a Doctoral Candidate in Political Science at Queen Mary University of London, UK.
the effect of interest.The desire to manipulate information encourages us to make up that information.
Examples of experiments that use made-up polling information as their treatment abound (e.g., Dahlgaard et al. 2017;Dizney and Roskens 1962;Goidel and Shields 1994;Madson and Hillygus 2019).However, only some fit the bill for misinformation.Others invoke a purely made-up context to go with their made-up treatment.For instance, they invite participants to imagine they are voting in a made-up election contested between two made-up candidates.Here, as the "false information" researchers provide is not about the actual "state of the world", they constitute more of a thought experiment, or science fiction, rather than deception. 3By misinformation, I refer to cases where the experimental context is the actual political world, and the treatment is false information about that context.
For example, consider the following experimental design: A recent poll of vote intentions at the upcoming presidential election, conducted in your state of New York, put Donald Trump and Joe Biden on the following vote shares: Trump 62%-32% Biden Please indicate which candidate you intend to vote for.
This example constitutes misinformation because the contextual setup establishes that we are talking about the actual political world, but the treatment information is false-Biden had a large lead in the polls in New York throughout the 2020 campaign.

The Ethics of Misinformation
Misinformation is deceptive.Two common ways of framing why deception matters ethically, and of assessing experimental ethics more broadly, are the Kantian deontological view and the utilitarian view (Berghmans 2007).

The Kantian View
Kant's ethical theory asserts that people are "ends"' in themselves (Kant 1978(Kant , 1996;;Korsgaard 1996).Treating a person as a "mere means," by using them to "promote your own ends in a way to which [they] could not possibly consent," violates their "dignity" (Korsgaard 2018, 77).This violation contravenes the "Formula of Humanity," which demands that "we avoid all use of force, coercion, and deception, that is, all devices that are intended to override or redirect the free and autonomous choices of other people" (Korsgaard 2018, 78, emphasis added).
A predominant concern with deception-especially with "identity," "motivation," and "activity" deception (ACHSR 2020, 7)-is that it overrides participants' autonomy by compromising informed consent (Bok 1995).In Humphreys' (2015, 101) terms, informed consent serves both a "diagnostic" and an "effective" function: it provides evidence that respondents were not treated as "mere means" (the diagnostic function) while also actively enhancing their autonomy (the effective function).Since deception typically involves misrepresenting "what you are doing" (activity deception), "who you are" (identity deception), and "the reasons for the research" (motivation deception), it fundamentally prevents participants knowing what they are choosing to do when they consent to participate in an experiment.Under activity deception, respondents are unaware that they are participating in research.Under identity deception, they are unaware of who is conducting that research.Under motivation deception, they are unaware why the research is being conducted.Misinformation does not necessarily involve any of these types of deception.The only deception it necessarily involves is a false piece of information about the state of the world.For we could even, in principle, tell respondents, "you are about to read the result of a completely made-up poll."If we also state who we are, what we are doing, and why, the diagnostic and effective functions of informed consent should be fulfilled, and respondents' dignity and autonomy therefore respected.
However, admitting the treatment information is false would compromise the "integrity" of the research design (ACHSR 2020, 8).Ideally, if we are studying what people would do if some false information were true, we should not forewarn them that it is false.But when omitting this disclaimer, we cannot assume that respondents are consenting to being deceived.They might not consent if they knew they were about to be exposed to misinformation.By participating, they are not acting in accordance with their (would-be) "self-chosen plan" (Beauchamp and Childress 2019;Phillips 2021).
Moreover, under misinformation, participants cannot autonomously make the decision they are asked to make in the experiment itself.In the bandwagon example, respondents cast a vote.Misinformation interferes with their ability to make this choice autonomously or rationally.And importantly, this experimental choice often mimics a real-world choice respondents will go on to make in their real lives.In experiments that ask people to express an opinion as an experimental outcome, this measure might reflect the opinion they have formed about the real-world matter at hand.By extension then, misinformation could lead people away from making autonomous, rational decisions in their lives outside of the experimental context (Phillips 2021, 284).When we treat research participants with misinformation, we deceive them into making the choice in a way that achieves our ends-say, the falsification or verification of some hypothesis-and not theirs.On the Kantian view, misinformation therefore violates respondents' dignity by disrespecting their autonomy.

The Utilitarian View
The utilitarian view, meanwhile, emphasizes a trade-off "between the wrong or harm of deception and the benefits resulting from the research" (Berghmans 2007, 14).What is the "wrong or harm" of misinformation?
Firstly, the loss of decision-making autonomy just discussed is assimilable, in a utilitarian framework, to what Humphreys (2011) calls a "participation cost"-a direct cost to the participant incurred by taking part in the experiment.Insofar as people value their capacity for autonomy and rationality, denying them of this capacity is harmful.
Secondly though, the most obvious harms of misinformation are "process costs"-those "broader" costs "associated with the implementation of the research, including the immediate social effects of the intervention" (Humphreys 2011).One process cost of misinformation is the harm it may do to science by compromising participants' trust in researchers.As Hertwig and Ortmann (2008, 63; see also Wilson and Hunter 2010) note, "a likely outcome of deceptive practices is participants' future resistance to other research efforts."Given rising rates of survey non-response and the "overextraction" (Leeper 2019) of human respondents by social research-which increasingly relies on limited groups of regular survey-takers (Krupnikov, Nam, and Style 2021)-it is wise to avoid deterring participants.Even if respondents go on to participate in future experiments, they may be distrusting of the information they receive in these experiments (Humphreys 2014;Svorenčík 2016), compromising research quality. 4 Beyond this, the treatment effect measured in the experiment represents a potential process cost if it is based on a falsehood.In the bandwagon example, the treatment is designed to change people's voting behavior.Participants might end up voting for Trump where they would have voted for Biden.Influencing the vote through misinformation is detrimental to democracy (Brown 2018;De Ridder 2021).Participating in misleading experiments measuring preferences for war crimes "predisposes citizens toward behavior that would violate international humanitarian law but also can undermine their understanding of what is permissible under the laws of war" (Carpenter, Montgomery, and Nylen 2021, 8).In other words, misinformation can increase confusion about, and boost support for, war crimes.Zimmerman (2016, 185-187) argues that political science experiments "often" cause "great" and "lasting" harm to "at least one person or group."Indeed, political science applications of misinformation seem especially likely to incur such harmful process costs, compared to other disciplines.In canonical cases of deception discussed in the psychological sciences (see, e.g., Kelman 1967), the effects only reach the individuals who participated in the experiment-that is, they tend to take the form of participation costs.By contrast, changing someone's vote intention alters the result of an election and thereby the livelihoods of the members of a polity (Desposato 2016, 277).Making people more supportive of war crimes shifts public opinion about whether potentially thousands of citizens should undergo atrocities (Carpenter, Montgomery, and Nylen 2021).The questions political scientists address have collective implications.Correspondingly, there is a greater "collective wrong" (Whitfield 2019, 532-534) that the benefits of research must outweigh.
The False Promise of Debriefing Some, including APSA (ACHSR 2020, 8), propose "debriefing"-providing participants "information about the study [and] their role in it" (Desposato 2016, 282) after the fact-as a means of ameliorating the ethical pitfalls of deception (Berghmans 2007).From a Kantian perspective, when employing misinformation, researchers should debrief respondents.Debriefing is a way of showing a person respect by telling them what happened to them, even if it will change nothing in terms of concrete participation or process costs.
From a utilitarian perspective, debriefing may offset the process costs of misinformation discussed earlier.First, debriefing may help "to increase public trust and understanding of what we are doing" (Desposato 2016, 283).Arguably, confessing to having misinformed participants makes scientists seem more trustworthy than they would seem if they did not admit it.But it seems equally plausible that debriefing will compromise trust.Participants sometimes lack the political knowledge required to recognize that an experiment is deceiving them.If they do not learn after the experiment that the treatment was untrue, they will see no reason to believe it was.The debrief guarantees that they will find out.Debriefing, itself, could therefore incur its own process cost by damaging future research efforts (Hertwig and Ortmann 2008, 63).
Second, debriefing may offset process costs by overturning treatment effects.Suppose respondents are told after the bandwagon experiment that the poll result they saw was fake.People will see that they should not have been persuaded by the treatment, cueing them to reverse that change.This argument only holds if the treatment independently produces the behavior observed in the experiment.Yet when people process new or surprising information, they draw on the wider information environment, to understand or rationalize it (see, e.g., Lodge and Taber 2013). 5In research on the bandwagon effect, this tendency is called the "cognitive response mechanism" (Mutz 1998, 212): people change their mind in response to poll results because learning that a candidate is popular causes them to rationalize the information, by reflecting on other arguments supportive of that candidate. 6These other arguments convince them to jump on the bandwagon.The behavior observed in the bandwagon experiment, then, would be the product of reflection on arguments supportive of candidates other than just their popularity, brought about by false information about their popularity.Debriefing cannot guarantee that these unknown additional arguments will cease to be relevant to the respondent.Indeed, evidence suggests that factual corrections alter people's factual beliefs, but not their political attitudes (Nyhan et al. 2020).
If treatments are "transformative" (Paul and Healy 2018), the inability to overturn treatment effects through debriefing becomes an ontological problem.Experimental treatments can effectively replace participants' "psychological selves" (Paul and Healy 2018, 326).Under this framework, the people who receive a debriefing message at T 2 are a fundamentally different population from the group who entered the experiment at T 1 .Then, a debrief is not really a debrief, but rather a new message delivered to a group of people who are different from the group we misinformed.The difference is theoretically salient because the changes were brought about by a theoryderived intervention.The cognitive response problem is a (perhaps weaker) version of this challenge.The debrief is an attempt to persuade a group, more of whom will now be Trump supporters (according to the bandwagon effect), not to vote for Trump based on the information we previously provided to a group that was then composed of more Biden supporters-an importantly different group.
My point is not to reject the idea of debriefing outright, but rather that debriefing does not eliminate the ethical problems with misinformation.On a Kantian view, we should debrief, but we should still not misinform-there is no clear sense in which the former injunction overrides the latter.On a utilitarian view, debriefing cannot be expected to offset the process costs of misinformation and may even produce process costs of its own.Research subjects and scholars share the view that debriefing is not a solution to deception: absent informed consent, neither group judges experiments with debriefing more acceptable than those without (Desposato 2018, 745).

Inference under Misinformation
A simple response to these ethical challenges would be that misinformation, as a type of deception, is a necessary way of acquiring important knowledge (Penslar 1995).As Bok (1995, 2) puts it, researchers see a "trade-off between minor violations of the truth for the sake of access to far greater truths."With misinformation, the unique combination of a randomized treatment and a realistic context should enable researchers to get informative estimates of how people would really react if the treatment came true in the actual world-knowledge that can be applied to the benefit of society.In Humphreys' (2011) terms, though misinformation incurs "participation" and "process"' costs, it produces "outcome benefits"-the benefits of the knowledge produced by the findings and how it can be applied.So, in utilitarian terms, it may be worth it if these outcome benefits outweigh the costs discussed earlier.
There is, however, debate over whether the outcome benefits from experiments are valuable enough to justify their process and participation costs (e.g., Carlson 2020; Phillips 2021) and, indeed, whether such a calculation is even possible (Zimmerman 2016).Findley and Nielson (2016, 152) propose one approach, the "half-doubled rule" in which researchers "sincerely" estimate the benefits to human knowledge of a given experiment, and its costs to subjects, then reduce the estimated benefits by half and double the estimated costs.Such an adjustment may be necessary because researchers deceive themselves into believing their research is more valuable than it is (Desposato 2020).An extreme case of this logic, raised and rejected by Teele (2014, 116), is that ethics "are beside the point" when "experimental research is the only way to ensure valid causal inference." The trouble with misinformation is that it tends to compromise the validity of such causal inference.So its putative outcome benefits are largely illusive.It therefore looks unlikely that the general calculus of weighing up outcome benefits against process and participation costs will favor employing misinformation in many cases.This argument can be seen at three levels of abstraction: the research design level, the causal inferential level, and the metaphysical level.

Research Design
It is a well-recognized tenet of good research design that effects should be externally valid-it should be possible to generalize them to the "population of interest" (Toshkov 2016, 173).The real contexts invoked by misinformation experiments would appear to make them more likely to find an externally valid effect; the realistic situation means that the sample should behave more like the population out in the actual world.But when seeking to maximize external validity, it not only matters that the context reflects the actual world; it also matters that treatments do.In the bandwagon example, there is a population of polls out in the world.The experimental design must maximize the representativeness of the sample of polls used in the experiment with respect to this population, if its results are to be generalizable to the relevant context.In my example, that context is the 2020 U.S. presidential election.Misinformation inherently falls short here, because it is not sampled from a population, but rather made up.If the design and analysis of the experiment implicitly treat Trump's lead in New York as no less likely than Biden being in the lead (which the polls would really have suggested), then the effect of this scenario can be seriously miscalculated when mapping it onto the actual world (see de la Cuesta, Egami, and Imai 2021).
Realism also affects internal validity-whether "the independent variable produced the observed effect" (Halperin and Heath 2017, 149).Misinformation compromises internal validity by contravening the tenet of "experimental realism": to the extent that subjects become psychologically engaged in the process they confront, internal validity intensifies.Similarly, internal validity diminishes in the face of subject disengagement … .If subjects approach a task with skepticism or detachment, then genuine responses fade … .This raises the possibility that measures obtained do not accurately reflect the process being manipulated, but rather manifest a different underlying construct altogether.(McDermott 2011, 28) In the face of misinformation, skeptical respondents might not look at the experimental task as something to take seriously.Rather than thinking about the election and how they would behave, instead the false treatment leads them to behave in arbitrary ways.They might also see the false treatment as deliberately pushing them in a certain direction, and then do what they think the researcher "wants" them to do (Jimenez-Buedo and Guala 2016; Zizzo 2010).
Such concerns can be traced back to the influential work of psychologist Sidney Siegel (1961; see also Siegel and Goldstein 1959), in whose view deception undermined the relationship between the experimenter and the subjects, who, in consequence, are often too preoccupied with the true agenda of the experiment or doubt the announced relation between actions and rewards, which therefore ultimately jeopardizes experimental control.(Svorenčík 2016, 279) Arguably, this concern is only valid if participants know they are being misinformed.If they are deceived, then they do not realize that the treatment is unrealistic.But, in this case, the ethical problems discussed earlier presumably must be taken more seriously.Essentially, if a treatment is deceptive, it must pass a threshold of realism.If it is not realistic, it must not be deceptive.Misinformation depends on being deceptive, with all the attendant ethical concerns, to maximize its scientific value.

Causal Inference
In causal inferential terms, these problems arise because misinformation is "ill-defined" or "ambiguous" (VanderWeele 2018, 1). 7Misinformational treatments are a "hypothetical intervention," defined as "the specification, possibly contrary to fact, of the event or state X = x"-for instance, a hypothetical situation in the world where the polls (X) show Trump winning in New York (x) (VanderWeele 2018, 1).Such a hypothetical intervention is "well-defined" if "for each individual i in population P, there is a unique … distribution of values … such that the event or state X = x, along with the state of the universe and the laws of nature, jointly entail" the observed outcome-otherwise "the hypothetical intervention is said to be ill-defined or ambiguous'"(VanderWeele 2018, 1).
Misinformation is ambiguous by design.As Vander-Weele (2018, 4) goes on to explain, we could speak of a well-defined hypothetical intervention on a genetic variant even if we cannot at present alter it … it is relatively straightforward to think of just that one variant being other than it was and the rest of the universe being the same.In contrast, if we were to consider the temperature being 30 degrees versus 40 degrees in a specific town, there may have to be numerous things that would have to be different for the temperature to be different, such as the wind speed and air pressure in that town, along with the temperature in spatially contiguous towns, etc. … and these different states of the universe may have very different implications for the outcome.
For Donald Trump to poll at 62% in New York, lots of other things would have to be different too.Importantly, many of these differences are direct implications of the theories that we (for the most part) subscribe to as political scientists.For example, the economic vote (e.g., Pickup and Evans 2013) would predict that, had Trump's presidency brought about considerable economic growth and prosperity in the United States, he may have performed better at the 2020 election.If we lend strong credence to economic voting theory, we might argue that the economy would have to be in a better state for Trump's counterfactual polling to be well-defined.But of course, the theory also implies that if the economy were doing better our research subjects would already be more likely to vote for Trump anyway.This is just one possible scenario, and there are many others that could give rise to the counterfactual we set up in our treatment.All those scenarios, if they truly came to pass, would produce a different distribution of responses to the experiment.The hypothetical intervention is ill-defined.
Such ambiguity can give rise to "information equivalence": manipulating information about a particular attribute will generally alter respondents' beliefs about background attributes in the scenario as well … .Manipulating whether a country is described as "a democracy" or "not a democracy," for example, is likely to affect subjects' beliefs about such background features as the country's geographic location or demographic composition.If it does, then any differences between experimental groups cannot be reliably attributed to the effects of the beliefs of interest.(Dafoe, Zhang and Caughey 2018, 400-401) Here then, the "things that would have to be different" themselves directly affect participants' behavior in the experiment because the participants assume those things are different in order to make sense of the ambiguous intervention.In the bandwagon example, participants assume that the economy must be doing better, and this assumption affects their vote choice.As a result, the experiment does not uniquely isolate the effect of interest, 1214 Perspectives on Politics Reflection | Misinformation in Experimental Political Science but rather any number of other effects of the background beliefs that respondents form to reduce the ambiguity of the misinformation (Landgrave and Weller 2022).

Metaphysics
In metaphysical terms, such problems arise because misinformation assumes the existence of impossible worlds.On a prominent view, counterfactuals are grounded in distinctions between the actual world and possible worlds (Lewis 1973): if something is false in our world, then counterfactual reasoning assumes there is a possible world where this thing is true-essentially, an alternate universe.Misinformation experiments attempt to create a possible world that is a "perfect duplicate" (Paul and Healy 2018, 321) of the actual world in all respects until the treatment is delivered.The truth of the treatment is the only thing that differs between the actual world and the possible world created by the experiment.
To see why we cannot do this, consider with David Lewis (1973, 9, emphasis original) the statement "if kangaroos had no tails, they would topple over": We might think it best to confine our attention to worlds where kangaroos have no tails and everything else is as it actually is; but there are no such worlds.Are we to suppose that kangaroos have no tails but that their tracks in the sand are as they actually are?Then we shall have to suppose that these tracks are produced in a way quite different from the actual way.Are we to suppose that kangaroos have no tails but that their genetic makeup is as it actually is?Then we shall have to suppose that genes control growth in a way quite different from the actual way (or else that there is something, unlike anything there actually is, that removes the tails).And so it goes; respects of similarity and difference trade off.If we try too hard for exact similarity to the actual world in one respect, we will get excessive differences in some other respect.
We cannot assume that there is a possible world whose only difference from our actual world is that the misinformation is true.There are no such worlds.There is no possible world where everything was the same right up to the point that 62% of voters in New York decided to support Donald Trump.There are possible worlds where Trump polls at 62% in New York; those worlds just differ from ours in other ways too.
Our respondents are not in any of those other worlds and have not undergone the processes that produced the differences in those worlds relative to the actual world.So misinformation cannot speak to the worlds where the treatment is true and there are other necessary differences from our own, because the sample on which it bases its findings, and the population to which it generalizes its effects, do not come from those worlds.They live in the actual world.
But what if, when cast into a possible world by misinformation, participants imagine they are in that world and then act accordingly?A fake poll result might make them assume broader differences in the election contextin effect, projecting themselves into some possible world where the treatment makes sense.Of course, again, the trouble with this solution is precisely that people are likely to care about those differences and respond accordingly.Even if no unobserved factors affected responses in this way, people would have to be able to project themselves into infinitely many possible worlds to resolve the problem fully, because otherwise the question remains unanswered in many possible worlds-all those other worlds into which they do not "project themselves".

Approximating the Truth
Arguably, these inferential concerns are less pressing the more misinformation approximates the truth.Consider the bandwagon case again.Polls have a margin of error.If we were interested in studying what would happen if Trump had the lead in state polls, we could choose a state where this situation is within the margin of error of current polling.We could then make up a poll within this margin and use this as our treatment.For example, we could ask: A recent poll of vote intentions at the upcoming presidential election, conducted in your state of Arizona, put Donald Trump and Joe Biden on the following vote shares: Trump 49%-45% Biden Please indicate which candidate you intend to vote for.
These made-up vote shares would be within the margin of error of a Morning Consult poll conducted in late October 2020 suggesting Biden was in the lead on approximately 48% to Trump's 46%. 8Because the margin of error is largely brought about by random noise, we could claim that nothing else would have to change to produce this alternative outcome-we would just have to get lucky.But as the treatment is still made-up, this is still a case of misinformation-just one we can analyze without worrying about inferential problems.
A first issue with this response is that such a quantitative idea of approximation is not always applicable.Countless potential applications of misinformation are not as malleable as the bandwagon example.As soon as we move beyond conveniently blurry cases, all the ethical and inferential issues raised above rear their heads.A second issue is that going to the effort of approximating the truth begs the question-why not just tell the truth?

True Treatments
That is, why not take data from the actual world and deploy it as a treatment in an experiment that is firmly grounded in the actual world-as Desposato (2016, 282) puts it, "tell the whole truth"?Call this the "true treatments" approach.Indeed, in the Arizona example, December 2023 | Vol. 21/No. 4 1215the very same reason that the workaround makes sensethe fact that it could be produced randomly-also means that a poll could actually be conducted that comes up with the situation whose effects we want to test.A true treatments approach would involve taking such actual polling data and exposing respondents in the treatment group to it (e.g., van der Meer, Hakhverdian, and Aaldering 2016).This enables researchers to intervene in the political world without facing the charge of unduly influencing it by false pretenses.
For example, we could ask: 9 A recent poll of vote intentions at the upcoming presidential election, conducted  (Dafoe, Zhang and Caughey 2018, 400).The earlier arguments demonstrate that answering such questions raises considerable ethical and inferential challenges when the treatment information is not true.It works perfectly well with true treatments.

The Ethics of True Treatments
True treatments are not deceptive.As Humphreys (2014) claims, "interventions that seek to inject expert information into public debate around elections can sometimes be implemented using an entirely transparent design, with consent and no deception."On Kant's deontological view, this means that true treatments better respect the autonomy of the participant.On the utilitarian view, it means that they reduce the process and participation costs discussed earlier.Of course, they still produce consequences: if I report a real poll in a bandwagon experiment, the point is still to change people's opinions or behavioral intentions.But there are crucial differences.First, because the information is true, any effect it has could have happened anyway.Second, because the information is true, any effect it has may be normatively and politically desirable in some cases.Arendt (2005), for example, argues that truth-telling is essential to political life precisely because it informs people's beliefs and promotes a legitimate plurality of opinions.Legislating between the cases where Arendt's claim applies may require stronger assumptions about what is politically desirable that are beyond the scope of this paper-and on which political science lacks any strong consensus (Whitfield 2019, 530).Third though, a commitment to true treatments could shift debates away from the ethics of deception and towards -potentially more productive-discussions about how to resolve such "normative ambiguity" (Phillips 2021, 281).Perhaps there are political domains into which experimental political scientists should not intervene, but regardless, we are likely to be better off telling the truth when we do so.

Inference under True Treatments
Correspondingly, true treatments produce the outcome benefits that misinformation fails to produce, because they provide valid causal inferences.At the research design level, true treatments are representative of the information out in the actual world, so their effects are more externally valid than those of misinformation.Ideally, to maximize external validity, where different versions of the same information are available (as is often the case with polls) the versions should also be presented to respondents in proportion to their likelihood of exposure to those versions in their lives. 10As for internal validity, true treatments are realistic, so should minimize skepticism and other sources of disengagement.
At the causal inferential level, the hypothetical intervention is well-defined in the case of true treatments because the experiment makes the intervention nonhypothetical.We want to estimate the effect of information exposure as a counterfactual state, and the experiment directly intervenes to bring about this state-i.e., to make it no longer counterfactual for some people.As this exposure is determined purely randomly, it does not rely on anything else changing to produce it.That is, in VanderWeele's (2018, 4) terms, the "state of the universe" is held constant.There is no ambiguity in the intervention.We directly isolate and manipulate the intervention of interest and observe the effect.There is, then, no intrinsic reason for participants to "update their beliefs about background attributes," reducing concerns about "information equivalence" (Dafoe, Zhang and Caughey 2018, 400-401).
At the metaphysical level, true treatments do not require the existence of worlds that could not possibly exist.Instead, the treatment is true in the actual world.The possible world we are interested in is one where people are 1216 Perspectives on Politics exposed to this information.We know that this possible world exists because we create it by exposing them to the information.And randomization into treatment and control groups guarantees that there is a possible world in which each respondent was exposed to the information.In these possible worlds, everything else was the same up to the point that they were exposed to the information, because the process by which this exposure is determined is random.The counterfactual makes sense.Causal inference is possible.

Conclusion
In response to APSA's call to "carefully consider" any use of misinformation in experimental design, this paper has made three contributions.First, it has clarified how we can think about the ethics of misinformation, as a specific type of deception.Second, it has argued that misinformation tends to compromise inference.Third, it has invited researchers to adopt "true treatments" as a solution to these problems.
By providing these reflections, I raise what Paul and Healy (2018, 334) refer to as a "methodological challenge" to experimental political science.While such challenges often require that scholars look for "innovations" (Paul and Healy 2018, 334) to continue doing experimental research, I have argued that no innovation is necessary.If political scientists still see misinformation as the only way to tackle their research questions, then in addition to APSA's (ACHSR 2020, 7) suggestion to disclose, justify, and explain steps taken to mitigate the ethical cost of deception, researchers should also address whether and how their design overcomes the inferential problems set out here.But a better solution is right under our noses.By simply telling the truth, political scientists not only do more ethical research, but also do causal inference that makes more sense.truthful statements to convey a misleading impression") (Rogers et al. 2017, 456).Experimental treatments can, in principle, be deceptive in all these ways.For expositional clarity, I focus on cases of "lying by commission" (the "active use of false statements") (Rogers et al. 2017, 456), as this form of deception fits best with APSA's original definition of misinformation.Resolving debates about whether lies of omission and paltering pose the same ethical and inferential problems as lies of commission is beyond the scope of this paper.Future work may fruitfully address these nuances-indeed, these are exactly the kinds of debate I hope to spark. 2 Zimmerman (2016,(184)(185) provides an extensive list of such experiments, on a wide range of topics. 3Science fiction writer Ursula K Le Guin (2017, xiv) highlighted this analogy when explaining that "you can read … a lot of … science fiction, as a thought experiment … .Let's say this or that is such and so, and see what happens … .In a story so conceived, the moral complexity proper of the modern novel need not be sacrificed … .Thought and intuition can move freely within bounds set only by the terms of the experiment."She concludes, elsewhere, that "fiction is invention, but it is not lies"-however "if you begin to fake [the facts], to pretend things happened in a way that makes a nice neat story, you're misusing imagination.You're passing invention off as fact, which is, among children at least, called lying" (2019, 108). 4Concerns about subject disengagement and overextraction are particularly pertinent in the case of "elite field experiments" or "audit experiments," where potential participants are savvier to the deceptive practices used by researchers (Landgrave 2020, 490)not to mention, there are fewer of them.Misinformation is a common feature of such experiments (Bischof et al. 2021) and may produce unique process costs that extend beyond the effects on the individual taking the survey (Desposato 2022). 5Some theories emphasize "hot" or otherwise "affectively charged" cognition (Lodge and Taber 2013).In Affective Intelligence Theory (Marcus, Neuman, and MacKuen 2000), attitude-incongruent information induces "anxiety," motivating a "search" for information.Debriefing may reassure people that they do not need to have been anxious, but it cannot undo the experience.Kelman (1967) questions at length the researcher's right to affect the psychology of participants in such ways.
explicitly causal language, or the dominant frameworks of causal inference, are nonetheless likely to endorse the importance of validity. 8See https://morningconsult.com/form/2020-u-s-elec tion-tracker/. 9This is the result of a Morning Consult poll conducted in late October 2020: https://morningconsult.com/ form/2020-u-s-election-tracker/. 10 This is also a clear way to limit the presence of "lies of omission" and "paltering" (Rogers et al. 2017, 456), which true treatments do not fully rule out per se.
This approach might be criticized on the grounds that it does not do what experiments are supposed to do.Experiments allow us to answer counterfactual questions by drawing on possible political worlds; keeping the treatment within the range of values it takes within the actual world misses the potential to observe these counterfactuals.But this is the wrong way to think about what experiments do.The question is not strictly about what would happen if the political world were different.It is about what would happen if people's exposure to the political world were different-this is what we actually manipulate.We ask what people would do upon receiving some information about the political world.This is what the theories tested by informational interventions are ultimately about