The Politics of Police Data: State Legislative Capacity and the Transparency of State and Substate Agencies

Police, like other bureaucratic agencies, are responsible for collecting and disseminating policy-relevant data. Nonetheless, critical data, including killings by police, often go unreported. We argue that this is due in part to the limited oversight capacity of legislative bodies to whom police are accountable. Although many local assemblies lack the means for effective oversight, well-resourced state legislatures may induce transparency from state and substate agencies. This argument is evaluated in two studies of police transparency in the United States. First, we examine the compliance of 19,095 state, county, and municipal police agencies with official data requests over five decades, finding strong positive effects of state legislative capacity on transparency. Second, we examine the accuracy of transmitted data on killings by police, finding that lethality is systematically underreported in states with lower-capacity legislatures. Collectively, our study has implications for research on policing, legislatures, agency control, and analyses of government data.

T he ability of policy-making, elected representatives to manage the behavior of policy-implementing, unelected agents is vital to democratic responsiveness and accountability. Accurate information is necessary for effective agency management, enabling representatives to evaluate agent behavior, assess outcomes, and update legislation. Unfortunately, legislators are typically unable to observe bureaucratic actions directly, relying instead on reports from the agents themselves. Thus, the relationship between representatives and bureaucrats is often defined by stark informational asymmetries that disadvantage legislators, allowing agencies to operate with undesirably high levels of autonomy and low levels of accountability.
Well-resourced legislatures can mitigate these asymmetries by using their oversight and budgeting powers to compel transparency from their respective agencies. However, many representative bodies are resource deprived-lacking time, expertise, etc.-and therefore unable to coerce agencies to share information on their actions and outcomes. These shortcomings are particularly salient in local (i.e., city, county) government, where assemblies are often acutely underresourced. The inability of local legislatures to compel transparency has consequences for both local and state legislative bodies in the US, as state policy is routinely implemented by local agencies (e.g., public health clinics, schools, water authorities). Therefore, state representatives rely on local agencies for policy-relevant information, yet local assemblies often lack the capacity to compel this information.
We argue that in hierarchical governing systemsfederalism (e.g., Argentina, India) or unitary systems with devolved competencies (e.g., Denmark, Japan)-higher-level legislative bodies may engage in direct oversight of lower-level agencies to ameliorate this problem, using traditional legislative instruments to compel transparency among local agents. That is, the very resources that enable state legislatures to effectively oversee state agencies can also be deployed downward, allowing state legislatures to provide oversight of, and compel information sharing by, agencies organized at the county and municipal levels. As a result, where state legislatures have the capacity to effectively pursue these avenues we should observe more accurate agency reporting.
To demonstrate this, we study the transparency of American police agencies, a canonical "street-level" bureaucracy (Lipsky 1980) and an arm of government that remains understudied in political science (Soss and Weaver 2017). Police enjoy extraordinary discretion in exercising their power-including life and death decisions-making oversight and management crucial. Yet, most police agencies are organized at the municipal or county level, where the oversight capacity of elected representatives may be limited. Drawing from the literature on legislative capacity and oversight as well as realworld examples of legislative investigation, we argue that high-capacity state legislatures can use their oversight authority to compel state and substate police agencies to report relevant information. By increasing the transparency of state, county, and municipal agencies, state legislatures can reduce informational asymmetries between bureaucrats and elected representatives at each of these levels of government, providing support for policy making and agency control.
We test the implications of our argument in a multipart analysis. First, we analyze administrative records of 19,095 state, county, and municipal police agencies' compliance with official requests for information over nearly six decades, finding strong evidence for a positive relationship between state legislative capacity and agency transparency. This is the largest agency-level analysis of police to date. Second, to assess the accuracy of agency data transmissions, we compare self-reports of police lethality to crowdsourced estimates. We find that where state legislatures lack the capacity for effective oversight, police agencies systematically underreport their lethality, annually concealing hundreds of killings from the official record. Finally, we analyze the effect of a state policy intervention on deadly force reporting, demonstrating that substate agencies are responsive to state legislative intercession.
In summary, we present an intuitive argument for how hierarchical governing structures can reduce informational asymmetries that may inhibit policy making or allow unelected agents to operate with undesirable levels of autonomy. We then present evidence that the capacity for state legislative oversight can increase the transparency of state, county, and municipal agencies. Our argument and findings add to a growing literature on how legislative capacity shapes both the influence of bureaucrats over policy outcomes (e.g., Kroeger 2022) and the ability of legislators to manage agency behavior (e.g., Boehmke and Shipan 2015;Lillvis and McGrath 2017). Additionally, our analysis contributes to research on "democratic deficits" in the American states (e.g., Landgrave and Weller 2020;Lax and Phillips 2012) by demonstrating that deficiencies in legislative capacity allow for less transparent and therefore less accountable agencies. Finally, we believe that our argument begins to rectify a shortcoming in existing interdisciplinary "institutional" analyses (Thompson [1967] 2017) of police, highlighting the dearth of attention paid to electoral institutions and the potential for "legal accountability" (Romzek and Dubnick 1987) vis-à-vis police. We hope that others will follow in this vein, and we encourage our colleagues to devote greater attention to the "second face of the American state" by both applying classic political economic models of governance and innovating new approaches to the study of police (Soss and Weaver 2017, 567).
On police behavior more directly, our results illuminate the extent of police lethality and reveal systematic deficiencies in public records. Though government manipulation of official data is a wellrecognized problem in autocracies (Hollyer, Rosendorff, and Vreeland 2011), we show that these problems are also manifest in the United States-official statistics on crime and policing are systematically biased and cannot be taken at face value. Unfortunately, this means that thousands of academic articles on crime and policing are likely to be fundamentally flawed. This underscores the importance of understanding the potential limitations of official data to scientific research, especially where government officials may have incentives to conceal or manipulate data. The Trump administration provides a recent example as evidence indicates attempts to manipulate data on climate change (Friedman 2019), crime (Malone and Asher 2017), and public health (Stolberg 2020). We conclude the manuscript by encouraging our colleagues to study the political economy of data gathering and dissemination.

ENGAGING INTERDISCIPLINARY POLICE RESEARCH
Research on policing is extensive and interdisciplinary, yet it largely fails to consider how interbranch politics affect law enforcement. Borrowing the Thompson ([1967] 2017) model of organizations, this research can be classified into three broad categories: "technical," how police perform discrete tasks; "managerial," how police organize themselves; and "institutional," how police interact with nonpolice stakeholders, which is our focus. To date, the vast majority of institutional research focuses on what Romzek and Dubnick (1987) call political accountability-answering to residents-as opposed to legal accountability-answering to elected representatives -despite the fact that police (excepting elected sheriffs) are not directly accountable to residents and elected representatives have greater power to shape police behaviors. The existing literature's relative inattention to how political institutions influence policing presents an opportunity for substantial contribution by political scientists. In the following, we briefly review police research in each of these main categories, arguing that more attention needs to be paid to role of elected representatives as central actors in the police ecosystem, especially by political scientists.
Institutional (or interactive) studies seek to situate our understanding of police within a broader societal context. Most of this work focuses on sentiment toward police in the mass public, including research in criminology (Hawdon, Ryan, and Griffin 2003), law (Sunshine and Tyler 2003), sociology (Desmond, Papachristos, and Kirk 2016), and more. 1 The political science research on policing also tends to focus primarily on this police-resident perspective (see Soss and Weaver 2017)-for example, studying the sociopolitical implications of police contact (e.g., White 2019)rather than on police interactions with the legislatures (or even executives) to whom police are accountable. Even when research specifically targets police oversight, one does not find application of general theory on the role of elected representatives and their oversight competencies. Instead, this research tends to analyze specific case studies of civilian complaint processing (e.g., Goldsmith 1991).
To demonstrate this point, consider Prenzler and Ronken's 2001 typology of police oversight models. The authors outright dismiss the possibility of oversight by elected representatives, instead classifying models into three categories: (a) internal affairs (i.e., oversight bodies housed within-agency), (b) civilian review (i.e., external bodies reviewing internal oversight practices), and (c) civilian control (i.e., external bodies with investigation and sanctioning authority). To a political scientist, dismissing the role of elected representatives is a critical omission because representatives are the voice and hands of citizens in government-the possibility of civilian control of police, or any other bureaucracy, in absence of action by elected representatives is vanishingly small. As we elaborate below, though many local elected representatives lack the capacity for meaningful oversight, state legislatures often do have sufficient power to influence police behaviors. Here we follow and echo Soss and Weaver's 2017 advice to study police as part of government, incentivized by and accountable to other governmental actors.
Political science engagement with policing research should not be limited to institutional studies alone, as each arm of policing literature builds upon and complements the others. For example, technical studies focus on the execution and efficacy of police activities, including research on deescalation tactics (Todak and James 2018), race-gender bias in policing (Ba et al. 2021), the effects of police staffing on crime intervention and deterrence (Levitt 1997), and so on. These studies inform institutional research in general by showing us what police actually do and our research in particular by clarifying the need to consider features of communities (e.g., racial composition) and agencies (e.g., staffing) in our empirical analysis. With few exceptions (including Levitt 1997), these studies tend to take policy as given but, moving forward, must consider the broader political environment that produces policy and structures incentives for compliance. For example, Mughan, Li, and Nicholson-Crotty (2020) present evidence that electoral accountability may deter sheriffs from abusing property-seizure policies.
Managerial studies, which focus on within-agency organizational choices (e.g., Goldstein 1963), have clear connections with institutional research. For example, Farris and Holman (2017) show the political orientations of sheriffs shape their departments' immigration enforcement strategies. But these studies, too, tend to think of agencies in a vacuum. Even reform-oriented research, including analyses of "community policing" (Weisburd and Braga 2019) or sociologically oriented research on "police culture" (Campeau 2015;Crank 2014), come largely from the managerial perspective, implicitly assuming only police themselves have influence. Ultimately, however, agency heads are externally accountable, and we need to consider how this structures their incentives in making decisions that guide subordinate behavior. For example, recent work by Eckhouse (2021) nicely demonstrates the potential gains from marrying managerial and institutional approaches, offering a theory of how police (re)shape their behavior in response to monitoring regimes.
Moving police research forward requires deeper integration across these literatures and more meaningful cross-field collaboration. Political scientists can aid these efforts by recognizing police departments as government agencies and offering theories of how their location within this broader institutional context structures police incentives. For example, legislatures can issue policy-specific mandates, oversee police operations, alter police budgets, etc. Therefore, political scientists must bring the legislature into our understanding of policing. In the next sections, we participate in this effort by considering the hierarchical structure of government in which police operate, the potential policy tools that are at the disposal of legislature vis-à-vis police, and the de facto budget constraints on potential legislative oversight activity.

LEGISLATIVE CAPACITY AND BUREAUCRATIC CONTROL
Under nearly all modern governing arrangements, the elected representatives that write, propose, and pass policies into law delegate the enforcement of these laws to unelected agents. The level of autonomy enjoyed by these agents is a function of the ex ante, de jure discretion afforded to them in implementing policy. Importantly, however, this autonomy is also a function of these agents' de facto ability to resist ex post oversight. Although nearly all democratic legislatures possess constitutional authority for both ex ante design and ex post oversight, a legislature's ability to use these tools is subject to a budget constraint: legislative capacity. Legislatures' endowment of resources (i.e., capacity) for conducting legislative work such as gathering information, writing comprehensive proposals, practicing oversight, and, if necessary, issuing sanctions to exert control over wayward agencies vary considerably. This is familiar in US state politics research, where state legislative capacity has been tracked for several decades (e.g., Bowen and Greene 2014;Squire 2007).
Recent research indicates that higher-capacity legislatures design more detailed proposals that leave less room for bureaucratic discretion (Huber and Shipan 2002;Vakilifathi 2019). Higher-capacity legislatures also conduct more rigorous oversight of the bureaucracy, monitoring implementation for concordance with legislative intent (Boehmke and Shipan 2015;Lillvis and McGrath 2017). As a result of this design and oversight capacity, voters in states with wellresourced legislatures are more likely to get the policy outcomes they desire (Fortunato and Turner 2018;Lax and Phillips 2012). We continue in this emergent stream of research, identifying the potential role of state legislative capacity for oversight of county and municipal agencies.

STATE LEGISLATIVE OVERSIGHT OF POLICE AGENCIES
The level of discretion police officers enjoy is extraordinary, as are the potential consequences of their actions (Lipsky 1980). Even routine, relatively minor decisions-for example, assessing parking fines-can be costly for citizens. At the extreme, officer discretion is a matter of life and death, as police are permitted to subdue civilians with force, including the use of electroshock weapons and firearms. Although discretion may be necessary given the complex and unpredictable environments in which police operate, the severity of the potential consequences for residents from engagement with police also demands robust oversight. This need for discretion and oversight creates a well-known problem with a similarly well-known potential solution in principal-agent arrangements: it can be prohibitively costly to monitor police in real time, but ex post scrutiny may be applied to compel preferred behaviors (McCubbins and Schwartz 1984).
Transparent reporting is one such behavior. We assume that legislatures generally prefer transparency because they (a) require information for policy design and (b) cannot manage agency actions in a state of ignorance. This information includes statistics on crime, records of calls for assistance and corresponding dispatch, agency expenditures, the use of force by officers, etc. Without this information, representatives cannot develop high-quality policy instruments or ensure the conformity of policy execution with legislative intent. Furthermore, legislators have a strong interest-both intrinsic and electoral-in protecting the public welfare. This requires not only creating effective anticrime policy but also constraining abuse of the state's monopoly on violence. Faithful reporting by police is necessary to successfully pursue these objectives.
Conversely, police agencies have strong incentives toward opacity. First, existing theoretical research argues that the executive can leverage its informational advantage to obtain greater resources (e.g., Banks and Weingast 1992). Qualitative studies of police budgeting strategy support this theoretical consensus: informational advantage is critical and agency heads report exploiting it several times over in interviews with Coe and Wiesel (2001). Second, keeping the legislature in the dark allows agencies to operate with autonomy, able to enforce policy as they see fit, and therefore extract benefits from drift. This is true for both on-theground execution by individual officers and broad administrative priorities by agencies, as research indicates that police are averse to information-intensive, line-item budgets that constrain spending discretion (Rubin 2019). Third, opacity allows agencies to operate with impunity, free from fear of reprisal when drifting from their mandate. In the case of police agencies such drift is particularly onerous given the potential costs of officer abuse.
Police have strong incentives for opacity regarding use of force in particular, as the revelation of police brutality or killings may draw public ire and, with it, the potential for official sanctions; offending officers and their supervisors more likely to be disciplined, dismissed, or face criminal charges in such climates. These incentives are so strong that police agencies have been accused of falsifying coroner's reports to conceal their lethality (Egel, Chabria, and Garrison 2017). Evidence indicates police willingness to conceal abuses in other areas as well: inflating clearance rates (Yeung et al. 2018), timecard manipulation (Rocheleau 2019), indiscriminate property seizure (Mughan, Li, and Nicholson-Crotty 2020), and so on. Incentives for opacity extend beyond concealing abuse of office, as recent research by Eckhouse (2021) shows that police systematically manipulate crime statistics, reclassifying rapes as "unfounded," to improve their metrics.
Apart from deselecting agency leaders, a power that state legislatures only enjoy directly in reference to state-level agents, 2 what can legislatures do to encourage transparency? Generally, legislatures have two instruments: sunshine and budget. With the former (sunshine), legislatures may use investigatory powers to discover and publicize missteps to shame agencies into better performance. All state legislative chambers have the authority to subpoena records on expenditures and assets, formal and informal procedure, internal communication, and other data critical to overcoming their informational disadvantage. Using subpoena power, they may also compel testimony from rank-and-file bureaucrats, agency leaders, and outside experts in order to help decipher the information they have gathered or to report on specific decisions, events, and procedures. These testimonial periods serve to generate additional information and/or shame bureaucrats, which elsewhere has been shown to be effective in promoting compliance (e.g., Hafner-Burton 2008).
These are well-designed coercive instruments; thus, the threat of subpoena-backed by the power of contempt-is typically sufficient for compliance with the legislature's demands. Indeed, a "direct threat" of scrutiny or an explicit ultimatum is rarely required for agency compliance. Instead an "indirect threat" of scrutiny, produced by a legislature that has, over time, created a culture of compliance by engaging in rigorous oversight, can compel preferred behaviors. In other words, oversight of any agency in the state creates positive externalities for compliance among every agency in the state, as it establishes a legislature's reputation for scrutiny. 3 Of course, legislatures are at times forced to use their coercive powers. For example, in 2018, the Maryland General Assembly formed a special committee to investigate the Baltimore Police Department's gun-tracing task force that used its subpoena power extensively (Fenton 2020). That same year, the Florida House issued a series of subpoenas to the Broward County and Palm Beach County Sheriff's Offices in its investigation of police conduct during the Parkland shooting (Koh 2018) and the Nebraska legislature subpoenaed the head of its State Department of Correctional Services to deliver extensive information and testimony regarding its lethal injection protocol (Duggan 2018).
These investigations may reveal issues that require legislative solutions to guide future behavior including the provision of additional resources to agencies, the closure of statutory loopholes, etc. Should agencies remain recalcitrant, however, the legislature may employ its ultimate coercive power: reduced agency funding (or the threat thereof). Importantly, the budgetary powers of state legislatures apply to state and substate agencies because states provide direct funding to county and municipal police, as they do for county and municipal public health clinics, schools, etc. In 2017, 32% of total local government revenue came from state transfers (Tax Policy Center 2017), providing great leverage to state legislatures in coercing preferred behaviors from substate actors. Moreover, local governments actively lobby states for these resources on a regular basis (Payson 2020), affording legislatures repeated opportunities for influence.
Although all state legislatures possess the de jure powers of oversight and budgeting, their de facto capacity to engage in these behaviors-their time and resources for legislative work-varies considerably. Squire (2007) documents these resource endowments by tracking the degree to which legislatures differ in session length, representative salaries, and staff allocations, which determine (respectively) the time available for legislative activity, members' ability to invest in expertise (or become "full-time" legislators), and the labor available to legislators for information gathering, etc. For example, although representatives in California have ample resources for investigating and scrutinizing various agencies, representatives in New Hampshire are resource starved. California's legislators are paid an annual salary of $114,877 plus $206 for each sessionday, have an average of 18 staffers, and meet nearly every business day. New Hampshire General Court members, on the other hand, are paid just $100 per year total, share 150 staffers between 424 members, and meet fewer than 25 days per year. Our central expectation is that agencies will be more forthcoming -that is, police will share more information-when the legislature's capacity for compliance-inducing behaviors is greater. Therefore, when the legislature's resource endowment increases, so too will agencies' tendency to transmit information. 4 We have made the case that state legislatures require agency transparency, particularly in the case of crime and policing, but must also establish why compelling transparency from local agencies requires state legislatures-why must state legislatures direct their oversight capacity downward to county and municipal agencies in order to get these data? After all, county boards and municipal councils are the representative bodies to which these agencies are immediately accountable. First, many city and county legislatures may lack the capacity for meaningful oversight, as these bodies meet infrequently and their members are often part-time workers. For example, data on city councilor compensation in California and Florida reveal that the median councilor is paid just $5,562 annually, with 78% falling under the national poverty line and 11% paid nothing at all. Second, local bodies may lack the political will to oversee their own police agencies, necessitating state legislative involvement. Councilors may be unwilling to scrutinize their police agencies for fear of electoral reprisals, producing a kind of regulatory capture. Zimring (2017) discusses these local conflicts of interest regarding police accountability at length, and political science research shows that police use their localized influence to extract substantial rents from municipal governments (Anzia and Moe 2015). These problems dissipate at the state level as membership is geographically diverse, allowing representatives from one city to lead scrutiny of police in another.
Taken together, these factors reveal that local oversight may prove insufficient and state intercession may often be necessary, especially given the inherent complexities in evaluating police behaviors and budgeting practices. This may help explain why previous research has shown that local policing is responsive to state-level political processes at similar (Chaney and Saltzstein (DOJ) Uniform Crime Reporting (UCR) program, which is administered by the Federal Bureau of Investigation (FBI 2020): The FBI's Uniform Crime Reporting (UCR) Program serves as the national repository for crime data collected by law enforcement. Its primary objective is to generate reliable information for use in law enforcement administration, operation, and management. Considered a leading social indicator, UCR data is used for monitoring fluctuations in crime levels, evaluating policies, and regulating staffing levels. In addition to the American public relying on the data for information, criminologists, sociologists, legislators, city planners, the media, and other students of criminal justice use them for a variety of research and planning purposes.
Each year, the UCR provides every police agency in the US with a detailed framework for classifying, counting, and reporting events, including, but not limited to, homicide, so-called justifiable homicide by a police officer, sexual assault, and "clearances." Importantly, participation is voluntary-at the discretion of individual agencies-with many agencies choosing not to participate. Moreover, the DOJ does not audit these data or sanction individual agencies for nonparticipation. This is important for the research design because it means that all police agencies receive an identical stimulus from the DOJ but no centralized incentive to comply.
Given the voluntary nature of the program, UCR data are often (rightly) maligned as inadequate for assessing patterns of crime. For example, Maltz and Targonski (2002, 299) note the extent of data missingness and conclude that the "crime data, as they are currently constituted, should not be used, especially in policy studies." Regarding killings by police, specifically, Zimring (2017, 29) writes that "the voluntary nature of the reporting system means that a significant number of killings by police do not get included in the official numbers." Although we discuss these limitations in the following sections, it is important to note that, for the purposes of our study, this is a feature and not a bug. We are interested not in explaining variation in crime itself but variation in data sharing by police. Therefore, it is imperfect compliance with the UCR program that enables us to effectively test our argument on the role of legislative capacity. Specifically, compliance with UCR requests should be greater for agencies facing the threat of legislative scrutiny, the credibility of which is increasing in the state legislature's oversight capacity. From this perspective, the UCR data requests are analogous to an annual audit study of all police agencies where the treatment is time-and state-varying legislative capacity.
We outline above why legislatures would prefer police agencies to share data, but why would legislatures want to compel participation in the UCR in particular? UCR participation has substantial efficiency advantages over alternative means of data collection, at least from the legislature's perspective.
When legislatures want information on crime, policing, and public safety, they may either solicit this information on an ad hoc basis-requiring the design, funding, and implementation of data-collection regimes-or they may encourage their agencies to participate in the UCR at a substantial resource savings. Second, UCR data allow for comparison that may be of interest to legislatures. Because it is a national program, states are able to compare outcomes on important metrics using common definitions and criteria. This allows legislatures to, for example, evaluate whether innovations used in other states have proven effective and are worth deeper consideration. With ad hoc, individualized data alone, such comparison is difficult. 7 Taken together, UCR participation has the potential to yield information that is more valuable than innovating a new program, at a lower cost. This is likely why most states delegate the entire enterprise of crime and policing data collection and dissemination to the UCR. We see this manifest in reliance on UCR statistics in state legislative reports on crime and policing (e.g., Joint Legislative Audit and Review Commission 2008), or state legislative directives to begin reporting certain information into the UCR (Texas SB 349 [2021]), of course, these directives only have teeth in proportion to the legislature's capacity to compel compliance. When states do create their own programs for data dissemination, they tend to be a state-specific reporting of UCR-submitted data, as in Maryland and South Carolina (Criminal Justice Information Services 2019; Williams 2020), or UCR data supplemented by collecting additional information that the UCR does not request. For example, North Carolina's Highway Patrol traffic stop database was designed to collect a specific type of data-driver demographics and various outcomes of traffic stops made by state troopers-outside the purview of the UCR. 8 Stated alternatively, the existence and low cost (to states) of the UCR has effectively preempted the development of alternative, state-designed crime and policing data infrastructure.
To estimate the effect of state legislative capacity on police agency transparency, we undertake two empirical studies. First, we evaluate whether agencies are more likely to comply with DOJ data requests as their state legislature's capacity for oversight increases, by analyzing administrative records indicating UCR participation for all municipal, county, and state police agencies in the 50 states 9 between 1960 and 2017 in a 7 We acknowledge that comparisons relying on the UCR data may also be faulty; however, whether such comparisons should be undertaken is a normative question outside of the scope of our argument, that state legislatures demonstrate a preference for UCR data is an empirical claim supported below. 8 Note that this database was created in response to the passage of Senate Bill 76 by the state legislature in April 1999, providing additional evidence that state legislatures are interested in and can act to promote transparent reporting from police. 9 That is, all municipal, county, and state police agencies in the United States excepting those in Native American Reservations, US Territories, and Washington, DC. generalized difference-in-differences (DiD) framework. The data reveal strong evidence for a plausibly causal effect of capacity on UCR compliance.
Second, we take advantage of several recent initiatives to crowdsource reliable data on police lethality, including well-known projects by The Washington Post and The Guardian, to assess the accuracy of data reported by agencies. 10 It is widely accepted that these crowdsourced data provide more accurate information on the extent of killings by police than do official statistics (Zimring 2017). 11 Moreover, these projects are among the only systematic efforts by nonpolice to compile information overlapping the UCR and therefore present a rare opportunity to systematically scrutinize the accuracy of UCR data. Our expectation is that higher legislative capacity will be associated with more transparent reporting of deadly force. The data support this, indicating fewer reporting discrepancies in states with high-capacity legislatures. Finally, we assess the effect of a policy intervention that mandates external investigation of deadly force in seven states and find plausibly causal evidence that state legislative intercession increases police transparency.

STUDY 1: UCR COMPLIANCE
Our administrative data include all state, county, and municipal police agencies in the 50 states from 1960 to 2017, indicating whether or not agencies complied with annual UCR data requests. The data are split into two sample periods-1960-1994 and 1995-2017 -because the records before 1995 were obtained as physical media and then parsed from a series of quite large, unformatted (or unusually formatted) text documents. 12 Therefore, we cannot be completely certain that we have an exhaustive accounting for the 1960-1994 period and analyze the periods separately to avoid contaminating the latter, exhaustive sample. There are 19,095 active agencies across the two periods. 13 We do not analyze the scope or accuracy of the data submitted in this section, merely the most basic level of compliance: whether the agency provided any information at all in response to the DOJ's request. This is common in the audit studies that are similar in design to ours. In general, compliance trends upward over the sample period, with about 37% of agencies reporting in 1960 compared with 76% in 2017. Parsing these data further, in the older sample  only 1% of agencies comply every year, 15% never comply, and 84% comply intermittently. In the more recent sample , 48% of agencies comply every year, 8% never comply, and 44% comply intermittently.
Our capacity measure is an adaptation of the Squire (2007) index-the industry standard measure of capacity-which combines the institutional endowments discussed above: legislator compensation, length of legislative sessions, and legislative staff. As Squire notes, the original measure is calculated intermittently and always relative to Congress, making the index inappropriate for dynamic analyses like ours. Given these limitations, we instead gather raw data on legislator compensation, session length, and staff expenditures for each legislative session and scale the values into a summary measure using Quinn's (2004) factor analytic model. 14 Because our adapted measure uses inflation-adjusted dollar amounts and scales all years jointly, it is better suited for our over-time analysis. 15 In addition to the modified Squire index, we also include the imposition of legislative term limits as an additional proxy for capacity, as term limits cause a sudden and permanent reduction to aggregate experience and institutional memory.
To sum up, the outcome in Study 1 is a police agency's choice to comply with a UCR data request. The treatment variables are the resource-based legislative capacity measure and term limit imposition, and we use a generalized difference-in-differences design, relying on two-way (state and year) fixed effects, for identification (Angrist and Pischke 2008). 16 We estimate and report the results for several variations of this model on both samples, starting with reduced-form models where capacity and term limits are regressed on UCR participation with only state and year fixed effects.
Next, we generalize and estimate more fully specified models. First, we separately evaluate the influence of capacity on agencies with appointed and elected heads. DeHart (2020) notes that sheriffs are unusual among American police agency heads, as nearly all of them are elected, whereas state and municipal agency heads are appointed. Being elected rather than appointed should partially insulate sheriffs from state legislative pressures because elections both change the relevant principal and endow sheriffs with greater legitimacy, enhancing public support for the office and the individuals who hold it. To assess whether elected (relative to 10 Sumner, Farris, and Holman (2020) advocate for, and demonstrate the validity of, crowdsourced data in studies of local politics more generally. 11 For those interested in details on the different methodologies employed in these distinct crowdsourced data-collection efforts, we recommend chapter 2 of Zimring (2017). 12 This split also corresponds to an update to the Paperwork Reduction Act, which changed the rules governing the FBI's internal handling of the information in the earlier and latter periods. 13 This is an unbalanced panel because agencies are created, decommissioned, or merged over time.
14 The data are made available by Bowen and Greene (2014). We use the Quinn (2004) model because it is robust to missing values (and there are a handful). An equal-weights average of imputed resources produces substantively equivalent results. 15 Additional details on the construction of our preferred measure and summary statistics are provided in Appendix B. 16 Legislative resources and term limits are chosen by legislatures or voters and are therefore exogenously imposed on the police agencies whose compliance we are modeling. Term limits, in particular, are imposed by voters via referendum with arbitrary timing and may be safely considered exogenous to police behavior. appointed) agency heads are less sensitive to statelegislative scrutiny, we use DeHart's 2020 classification of all US counties as electing or appointing their sheriffs and interact an indicator for elections on our covariates of interest.
Second, given the low risk of simultaneity (i.e., UCR compliance having an effect on legislative capacity) 17 and the careful attention paid to measurement, the biggest threat to sound inference in our analysis is potential confounding. Although the DiD estimator should be reasonably robust to the omission of potential confounders (e.g., Angrist and Pischke 2008), we sought to further assuage potential concerns that our results are driven by omitted variable bias. Therefore, we gathered data to account for potential confounders including indicators for agency type (city, county, state, and whether or not the sheriff's office is noted in the constitution, following Falcone and Wells [1995]), 18 the size of the population to which the agency is accountable, how many officers and civilians each agency employs, government partisanship, and the percentage of Black residents. In addition to our measure of government partisanship, we also account for timevarying perturbations from party control by including Party Â Year fixed effects-where the former indicates either Democratic or Republican (relative to divided) control of the legislature and the latter indicates the year of observation. Collectively, these covariates help account for local oversight capacity, agencies' resources for compliance, and political preferences, among other factors.
In Table 1 we present estimates from reduced-form and fully specified DiD models for each of the two samples. The estimates reveal strong evidence for an effect of state legislative capacity on UCR participation and therefore strong support for our argument that capacity encourages police transparency. In all models, the point estimates are positive, large, and many times greater than their standard error estimate. Substantively, the effect of capacity is remarkable. For example, in the latter sample, if all legislatures had resources equivalent to those of Texas, currently about the mean level of capacity, and increased their resources to the level of one of the higher-capacity legislatures, like Michigan-which would not be an unprecedented change (Squire and Hamm 2005)-the predicted effect on the probability of UCR participation would be an increase of over 0.06. This translates to an expectation of between 1,200 and 1,300 more agencies complying with UCR data requests each year. 19 Term limit imposition also yields the anticipated effect, sharply reducing compliance probability by about 0.07 in the earlier sample (where there is less variability on the covariate) and 0.17 in the latter sample. Furthermore, the effect of legislative capacity is muted for agencies with elected heads in the latter sample and washed out completely in the earlier sample. Likewise, term limit effects are muted for elected heads in the latter sample, but the interaction is null in the earlier sample. The data suggest election, relative to appointment, seems to insulate some heads from legislative scrutiny.
In Appendix B, we report the results for all control variables and show that central result is robust to an array of alternative specifications including interacting capacity with partisan legislative control; analyzing state, county, and municipal agencies separately; and employing an alternative measure of capacity derived from Vannoni, Ash, and Morelli (2021). The effect of capacity remains positive, substantively large, and statistically significant under all conditions: increasing 17 Note that because UCR data are filed after the reporting period, legislative capacity is effectively lagged, making anticipation the only threat of reverse causality-i.e., that legislative capacity increases in expectation of increased transparency. We find this implausible. 18 Forty-one states have noted the sheriff's office in their constitutions at any time and thirty-seven retain the note as of 2021. A full accounting is given in Appendix D, where we also discuss constitutional funding mandates for sheriffs. 19 Calculation uses 2008 scores. Taking a conservative approach, as advocated in Mummolo and Peterson (2018), the effect of an average within-unit standard-deviation (0.275) increase to capacity produces an increase of 0.013, which translates to an additional 247 agencies complying with UCR data requests annually. Although there may, of course, be some additional risk of confounding despite our various checks, the robustness of our results to the inclusion of a variety of observed potential confounds gives us confidence in the direction and relative size of our effects. Overturning these results would require an unmeasured timevarying variable that (a) is jointly related to participation and capacity (i.e., relevant to the main effect), (b) has heterogeneous over-time variation across states (i.e., does not produce a trend proxied by the year fixed effects), and (c) is not already proxied for by the current time-varying controls. Furthermore, sensitivity analyses following Cinelli and Hazlett (2020) reveal that such a confounder would need to predict both capacity and UCR participation over 10 times better than does our most predictive control variable (population, which is correlated with compliance: β > 0.06 and t > 50 and in both samples) to wash out the capacity effect. Although this is possible, we cannot think of particular covariates satisfying these conditions; otherwise, we would have included them in the model. At minimum, we have shown that there are systematic differences in UCR participation across states and that these are robustly correlated with state legislative capacity. Although the magnitude of this effect may be better identified in subsequent studies, we believe that first revealing this relationship provides substantial value for researchers in two ways. First, our analysis suggests that political oversight is a fruitful area for future research, as political scientists increasingly analyze policing. Second, our analysis provides analysts with information on where, and a plausible explanation for why, data on crime are likely to be of lower or higher quality. Although we do not study crime itself as an outcome here, our analysis has clear implications for research that does.

STUDY 2: UNDERREPORTING DEADLY FORCE
The UCR is the only comprehensive, national effort by the US government to build an accounting of police lethality, but these data are inaccurate. 20 The intermittent UCR compliance discussed above makes it clear that any accounting of crime or police activity contained in the UCR is incomplete. Furthermore, given the evidence that noncompliance is determined in part by political concerns and recognition of the substantial incentive police agencies have to obfuscate (or exaggerate) certain behaviors and outcomes, the information that is contained in the UCR is almost certainly inaccurate and biased toward reflecting favorably on police agencies. Such bias is likely acute in the case of police use of lethal force or what the UCR calls "justifiable homicide by law enforcement," and indeed we find that underreporting is pervasive.
Unfortunately, UCR estimates of police lethality have been the standard in academic research for decades. Jacobs and O'Brien (1998, 846) went so far as to defend the data as "likely to be accurate because such homicides are difficult to conceal." As recently as Holmes, Painter, and Smith (2019), these data have been used as credible estimates of police lethality despite discussions of the well-known limitations of these data in academic research (e.g., Klinger 2012) and in the documentation of the data themselves. 21 Though previous research recognizes UCR shortcomings, recognition alone is insufficient. The fact that the wealthiest and most durable democracy in history does not know how many of its residents are killed by agents of the state is alarming and warrants study such that we may understand the political processes that produce (or permit) this downward bias in the official statistics and how the bias may potentially be alleviated.
We argue that one political factor driving this opacity, or bias, is the inability of legislatures to shape the behavior of police agencies. We evaluate this argument by comparing the UCR's official statistics to deadly force counts from several crowdsourced initiatives organized in recent years: The Guardian's The Counted  (2014-2016), and Fatal Encounters (2013-2016. 22 Although not ground-truth data, these projects represent the best available estimates on the extent of police killings. Even former FBI Director James Comey has recognized the greater accuracy of these crowdsourced data, calling the UCR statistics "embarrassing and ridiculous" by comparison (Kindy 2015). A central feature of these data is that each case is independently vetted and confirmed, and indeed an 20 For police killings there are two other potential official sources: the Centers for Disease Control's National Vital Statistics System (NVSS, which is not designed for police lethality accounting) and the Bureau of Justice Statistics' (BJS) Arrest Related Deaths (ARD). ARD was discontinued in 2014 after the BJS concluded that ARD did not meet its data-quality standards and systematic underestimation in both ARD and NVSS was confirmed by Barber (2016). Although a new program by the CDC-the National Violent Death Reporting System-offers some improvements, this project covers only a selection of states and still appears to substantially underestimate. 21 We note that many scholars are now turning to city-level data, often gathered internally and reported voluntarily or through Freedom of Information Act requests. Knox, Lowe, and Mummolo (2020) provide an excellent explanation of the bias in the event-generating processes that underlie these data, but we urge further caution regarding bias in the data-generating process-they are subject to the same selection biases when being recorded and reported as UCR data. 22 Additional details on these data are available in Appendix C. Our sample stops at 2016 because UCR data on police killings disaggregated to the state level are presently unavailable from 2017 onward. independent report issued by the DOJ suggests they provide a superior accounting (Banks 2016). Thus, discrepancies between UCR and the crowdsourced data provide an estimate of (known) undercountingthe number of known police killings that could have been officially reported but were not. Assessing difference across sources in this manner is common in event data analysis of political violence (e.g., Davenport and Ball 2002).
We plot these data in Figures 1 and 2 to demonstrate the systematic underreporting of police killings in the UCR. Figure 1 gives aggregate lethality figures by year, and there are two clear patterns. First, the figures are relatively consistent across time within reporting group. Second, the UCR underestimates police killings by at least 50% relative to crowdsourced estimates every year. We note that Fatal Encounters is a high outlier because it counts "deaths in custody" (e.g., dying of a heart attack while in custody awaiting arraignment) in addition to "justifiable homicides." Though Fatal Encounters' counts are likely to be biased upward as a result, these estimates are still useful to include in our comparison set, especially because we know the source and direction of the bias. Figure 2 decomposes the 2015 estimates to demonstrate state-level differences in reporting on the incidence of police killings per 100,000 residents. Again, there are two central patterns. The first is the striking degree of variability in the incidence of police killings across the United States. For example, police killings are a relatively rare occurrence in Connecticut but an order of magnitude higher in New Mexico. In 2015, police lethality in all crowdsourced data sources is also substantially greater than the official estimates of drugrelated homicides (452) and gang-related homicides (757). The second main pattern is the substantial variability in degree of underreporting. For example, Michigan reports almost all known police killings, whereas Florida fails to report any.
This variability in underreporting is the focus of Study 2, as we argue that a significant portion of it can be attributed to systematic differences in capacity for legislative oversight. To test this, we undertake two additional analyses. First, we compare rates of underreporting of police lethality across states to the states' legislative capacity. The data indicate that states lacking the capacity for oversight have significantly higher numbers of unreported police killings. Second, we provide evidence for a plausibly causal effect of state legislative intercession on police transparency by analyzing agency-reported deadly force numbers pre-post a policy intervention in seven states. This analysis suggests that the intervention significantly reduces the number of unreported police killings.
To measure the underreporting of police lethality, we calculate the difference between the publicly gathered data and the data contained in the UCR in each state-year (e.g., WaPost AZ,2016 -UCR AZ,2016 )-the number of known police killings unreported in official statistics, scaled per 100,000 residents. 23 The effective sample period for these data is 2013-2016, which limits our modeling options because capacity does not vary meaningfully within states over this period and term limits do not vary at all. Given the lack of within-unit variation in these main predictors, we focus on modeling the cross-sectional variation and, therefore, do not include state fixed effects. Instead, we use state random effects to account for within-unit correlation in unobservables and additional control variables to account for potential state-specific confounders. Theoretically, these are the state's available resources for collecting and disseminating information on police killings and the preferences of police agencies and legislatures for transparency. Operationally, these enter the model as the state's wealth (median income), the average number of agency employees (officers and civilians), share of Black population, the partisan control of government, and the political preferences of police agencies. Agency conservatism is captured using Bonica's (2016) preference estimates derived from campaign contribution data, and we use the mean estimated ideal point for all contributors working in law enforcement, where greater values indicate greater conservatism. 24 Research on the connection between political preferences and preferences for police militarism or violence suggest this is the best available measure (Christie, Petrie, and Timmins 1996;Gerber and Jackson 2017  Guardian with 0 opportunities (not 0 counts)-conflating cases in which all deaths were faithfully reported and cases in which there were no deaths to report. 24 We use the 2014 electoral cycle estimates. 25 Nebraska is coded as Republican, reflecting candidate endorsements. Switching this coding does not change the results, nor does substituting median legislator ideology for partisan control. This can be seen in Appendix C. 26 We do not include a measure of UCR participation because it is posttreatment. However, the only role UCR participation could play is to mediate the effect of capacity on unreported police killings -i.e., identify the extent of capacity-related underreporting that is due to nonparticipation. In Study 2, we are interested in the total effect of capacity on unreported police killings. This is what our coefficient captures, with an understanding that some of this effect is likely due to UCR nonparticipation because participation is positively related to capacity (Study 1).
Before turning to our results, it is important to recall precisely what we are modeling: a police agencies' propensity to report a killing, given that the killing has occurred. Therefore, the relevant covariates to include are predictors of the choice to report-that is, preferences and resources-not predictors of the event itself, such as local rates of violent crime. Importantly, in addition to being an irrelevant control variable, any official measure of crime would be subject to the same deficiencies as are the UCR estimates of police lethality and would therefore induce estimation bias.
The results are provided in Table 2, with separate models for each crowdsourced database. All covariates are standardized to facilitate interpretation, with each coefficient indicating the effect of a one-standarddeviation (or binary) increase in the predictor. As can be seen, state legislative capacity is consistently associated with less underreporting in official statistics, as agencies in states with higher legislative capacity are more likely to enter police killings in official reports across all data sources. The results indicate that a onestandard-deviation increase in legislative capacity is associated with around 0.035-0.075 fewer unreported police killings per 100,000 residents-or an average of 2.2-4.8 per state annually. Converting all states from their true capacity to the sample maximum (California) is predicted to reduce unreported killings by 284-670 per annum. Also, as expected, the estimate from the Fatal Encounters analysis is smaller and has a larger relative standard error, reflecting our qualitative understanding that these data bias undercounts upward from the true value. 27 Finally, the term limits indicator, which we would expect to have a positive effect on underreporting, has an effective zero estimate in all models. The strong negative relationship between transparency and term limits revealed in the first study is not manifest here.
To be clear, causal identification in this analysis is complicated by the lack of random treatment assignment, the sparsity of the outcome data, the time invariance of the main predictors, and our inability to include state fixed effects. Each precludes standard design-or model-based strategies for casual inference. Although reasonable people can disagree on what can be drawn from observational data analysis without ideal identification strategies, at minimum our analysis makes novel use of crowdsourced data to identify the set of states where police killings are most likely to be underreported and demonstrates that this underreporting is significantly negatively correlated with legislative capacity. Should researchers have alternative (i.e., noncapacity) theories explaining this relationship, we strongly encourage them to pursue this in future research. To further investigate our capacity-based argument, however, we can assess the causal effect of legislative intercession on the transparent reporting of deadly force, even if we cannot assess the causal effect of legislative capacity per se, which is of course an antecedent factor to intercession. Since 2011, California, Colorado, Connecticut, Illinois, New York, Utah, and Wisconsin have implemented policies requiring and regulating the investigation of officer-involved deaths. 28 Consistent with our general argument, these laws are positively correlated with legislative capacity (p = 0:017). Each of these laws prescribes a systematic investigation protocol for officerinvolved deaths (or use of force) by actors outside of the officer's agency and mandates the application of the protocol. Therefore, these laws prescribe a level of transparency and formal memorialization of the events. This both reduces the incentive to obfuscation in official statistics by expanding the number of actors with knowledge of the event and increases the probability of being held accountable for underreporting. Our empirical expectation is that this policy intervention should reduce the number of unreported police killings. To test this, we estimate a series of DiD models where the outcome is the difference between the publicly gathered data and UCR data in each state year (e.g., WaPost AZ,2016 -UCR AZ,2016 ). Our covariate of interest is the intervention status for the state-year. We expect the treatment to have a negative effect-to reduce the number of unreported killings. 29 Table 3 presents the DiD analyses of the effects of the investigation policy. Every model recovers a negative and significant estimate, revealing evidence for a plausibly causal effect of state legislative intercession on the transparent reporting of deadly force. Substantively, our results suggest that, by imposing investigation policies, the treated states improved the accuracy of police self-reporting by an average of 5 to 10 deaths annually. These results, taken together with the robust negative correlations between capacity and underreporting in Table 2 provide compelling evidence for our overarching argument that endowing state legislatures with the resources for rigorous oversight (or policy intervention) can substantially improve bureaucratic transparency across levels of government. This, in turn, should help legislatures acquire information that is necessary for policy design as well as help direct and constrain the behavior of agencies, ultimately helping align government action with voters' preferences.

DISCUSSION
We have argued that state legislatures-which depend upon state, county, and municipal agencies for policyrelevant information-can direct their scrutiny powers downward, helping to compel bureaucratic transparency of state and substate agencies. Thus, increasing state legislative capacity can increase information sharing and ultimately improve legislative management of agencies at each of these levels. An analysis of administrative records of 19,095 police agencies' compliance with official data requests over nearly six decades yielded strong evidence for a plausibly causal effect of capacity on state, county, and municipal agency transparency. We then assessed the accuracy of reported data by comparing the number of known unreported police killings in each state over recent years to states' legislative capacity, finding a robust negative association-states with higher-capacity legislatures had significantly lower rates of unreported police lethality. A DiD analysis of a transparency intervention provided complementary evidence that state legislative intercession improves reporting of police lethality.
Taken together, we believe this provides compelling evidence that (a) oversight can ameliorate informational asymmetries between unelected government agents and elected representatives and (b) state legislative oversight capacity can be directed downward to scrutinize substate agencies. This contributes to our understanding of interbranch conflict across levels of government, the broader implications of investing in legislative capacity, and the link between these investments and democratic responsiveness. Furthermore, police present a difficult test of our argument on the potential for state legislative bodies to affect localagency behaviors. Relative to other agencies, police are generally popular (Gallup 2019) and possess strong Hawaii has also passed such a law, but its passage falls outside our sample period. Data on investigation laws were obtained from the National Conference of State Legislatures. 29 Note that the outcome is not scaled in these models, as this effectively interacts the treatment with population changes, breaking the identification strategy. Controlling for population on the righthand side of the model does not substantively change the estimates on the treatment, nor does including the other control variables from Table 2. unions, which may make police less sensitive to legislative demands for transparency. This suggests that our findings here should generalize well to other departments of the executive. Even absent these broader implications for oversight, we believe that the specific analysis of policing here is important. As Zimring (2017) notes, the volume of killings by police is immense, with roughly 1,000 deaths per year, over half of which go unreported in official statistics. This is nearly 40 times the lethality threshold used in some measures of civil conflict. Yet, this particular form of political violence receives comparatively little attention in political science research. We believe that our approach and findings here, complementing Soss and Weaver's (2017) argument to study police within the broader context of governance and democracy, suggest a potentially fruitful path toward better understanding how police behaviors may be brought in line with public sentiment. Our work also contributes to a fast-growing literature on the deep and wide-ranging implications of legislative capacity, an area of legislative research of which scholars have barely begun to scratch the surface. Furthermore, research on the relationship between governing bodies across different levels is growing but overwhelmingly focused on bottom-up interactions (e.g., the lobbying of Congress or state legislatures by municipal governments: Loftis and Kettler 2015; Payson 2020), whereas this article suggests that top-down interactions also deserve attention.
In addition, we provide evidence that the gathering and dissemination of official data is politically motivated, obscuring our understanding of government action. There are several implications. First, if agency behaviors are conditioned by expectations of scrutiny, then coercing transparency may effectively decrease police violence. Second, insights from previous studies using official statistics on crime and policing should not be taken on faith and may warrant reexamination, especially studies that use UCR data without accounting for the considerable selection bias in their generation and dissemination: given the above analyses, is difficult to think of any circumstance under which UCR data are appropriate for studying the causes of crime. Third, researchers need to consider the dataproduction process explicitly and apart from the eventgeneration process (Cook and Weidmann 2019), especially when data providers have interests that are inconsistent with transparency.
To see how this matters, consider research by Knox, Lowe, and Mummolo (2020) showing official statistics can mask racially biased policing by recording only the conditions of interactions (e.g., the race and speed of a driver an officer has pulled over) and not recording the conditions of interaction opportunities (e.g., the race and speed of all drivers an officer observed). Importantly, the authors' estimates of the extent to which recorded data obscure the true nature of the eventgenerating process assumes that the data-generating process is unbiased (e.g., the race and speed of all drivers an officer has pulled over is faithfully recorded). Our analyses show that this almost certainly is not the case, so we may consider the bias estimated by Knox, Lowe, and Mummolo (2020) a lower bound, at least in the case of behaviors or events, like racially biased use of force, that bear potential disciplinary costs.
Carefully considering how and why our realized data deviate from ground-truth events is important in areas aside from crime as well. Just as police have incentives to underreport their lethality, school personnel have incentives to manipulate test scores (Jacob and Levitt 2003) and the Environmental Protection Agency may face pressure to underestimate pollution-related deaths (Friedman 2019). Moreover, as news media consumption becomes increasingly segmented and local media divestment continues (Peterson 2021), the ability of the press to serve as a watchdog declines. Therefore, better understanding political avenues for ensuring more accurate government data reporting is increasingly important. This matters both for less salient issue areas where auxiliary media-based data are not easily and readily available and, in the future, for high-salience issues as media resources continue to be diminished. Good governance and good science require good data, and we simply cannot rely on media to fill in gaps in the official statistics. Thus, we urge our colleagues to study the political economy of official statistics-critically scrutinizing the accuracy, reliability, and political determinants of official data needed for government accountability and scientific inquiry.

SUPPLEMENTARY MATERIALS
To view supplementary material for this article, please visit http://doi.org/10.1017/S0003055422000624.

DATA AVAILABILITY STATEMENT
Research documentation and data that support the findings of this study are openly available at the American Political Science Review Dataverse: https://doi. org/10.7910/DVN/MFFV3C.