Hostname: page-component-8448b6f56d-xtgtn Total loading time: 0 Render date: 2024-04-23T06:15:58.147Z Has data issue: false hasContentIssue false

The Politics of Police Data: State Legislative Capacity and the Transparency of State and Substate Agencies

Published online by Cambridge University Press:  20 July 2022

SCOTT J. COOK*
Affiliation:
Texas A&M University, United States
DAVID FORTUNATO*
Affiliation:
University of California, San Diego, United States, and Copenhagen Business School, Denmark
*
Scott J. Cook, Associate Professor, Department of Political Science, Texas A&M University, United States, sjcook@tamu.edu.
David Fortunato, Associate Professor, School of Global Policy and Strategy, University of California, San Diego, United States, and Associate Professor, Department of International Economics, Government and Business, Copenhagen Business School, Denmark, dfortunato@ucsd.edu.
Rights & Permissions [Opens in a new window]

Abstract

Police, like other bureaucratic agencies, are responsible for collecting and disseminating policy-relevant data. Nonetheless, critical data, including killings by police, often go unreported. We argue that this is due in part to the limited oversight capacity of legislative bodies to whom police are accountable. Although many local assemblies lack the means for effective oversight, well-resourced state legislatures may induce transparency from state and substate agencies. This argument is evaluated in two studies of police transparency in the United States. First, we examine the compliance of 19,095 state, county, and municipal police agencies with official data requests over five decades, finding strong positive effects of state legislative capacity on transparency. Second, we examine the accuracy of transmitted data on killings by police, finding that lethality is systematically underreported in states with lower-capacity legislatures. Collectively, our study has implications for research on policing, legislatures, agency control, and analyses of government data.

Type
Research Article
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2022. Published by Cambridge University Press on behalf of the American Political Science Association

The ability of policy-making, elected representatives to manage the behavior of policy-implementing, unelected agents is vital to democratic responsiveness and accountability. Accurate information is necessary for effective agency management, enabling representatives to evaluate agent behavior, assess outcomes, and update legislation. Unfortunately, legislators are typically unable to observe bureaucratic actions directly, relying instead on reports from the agents themselves. Thus, the relationship between representatives and bureaucrats is often defined by stark informational asymmetries that disadvantage legislators, allowing agencies to operate with undesirably high levels of autonomy and low levels of accountability.

Well-resourced legislatures can mitigate these asymmetries by using their oversight and budgeting powers to compel transparency from their respective agencies. However, many representative bodies are resource deprived—lacking time, expertise, etc.—and therefore unable to coerce agencies to share information on their actions and outcomes. These shortcomings are particularly salient in local (i.e., city, county) government, where assemblies are often acutely underresourced. The inability of local legislatures to compel transparency has consequences for both local and state legislative bodies in the US, as state policy is routinely implemented by local agencies (e.g., public health clinics, schools, water authorities). Therefore, state representatives rely on local agencies for policy-relevant information, yet local assemblies often lack the capacity to compel this information.

We argue that in hierarchical governing systems—federalism (e.g., Argentina, India) or unitary systems with devolved competencies (e.g., Denmark, Japan)—higher-level legislative bodies may engage in direct oversight of lower-level agencies to ameliorate this problem, using traditional legislative instruments to compel transparency among local agents. That is, the very resources that enable state legislatures to effectively oversee state agencies can also be deployed downward, allowing state legislatures to provide oversight of, and compel information sharing by, agencies organized at the county and municipal levels. As a result, where state legislatures have the capacity to effectively pursue these avenues we should observe more accurate agency reporting.

To demonstrate this, we study the transparency of American police agencies, a canonical “street-level” bureaucracy (Lipsky Reference Lipsky1980) and an arm of government that remains understudied in political science (Soss and Weaver Reference Soss and Weaver2017). Police enjoy extraordinary discretion in exercising their power—including life and death decisions—making oversight and management crucial. Yet, most police agencies are organized at the municipal or county level, where the oversight capacity of elected representatives may be limited. Drawing from the literature on legislative capacity and oversight as well as real-world examples of legislative investigation, we argue that high-capacity state legislatures can use their oversight authority to compel state and substate police agencies to report relevant information. By increasing the transparency of state, county, and municipal agencies, state legislatures can reduce informational asymmetries between bureaucrats and elected representatives at each of these levels of government, providing support for policy making and agency control.

We test the implications of our argument in a multipart analysis. First, we analyze administrative records of 19,095 state, county, and municipal police agencies’ compliance with official requests for information over nearly six decades, finding strong evidence for a positive relationship between state legislative capacity and agency transparency. This is the largest agency-level analysis of police to date. Second, to assess the accuracy of agency data transmissions, we compare self-reports of police lethality to crowdsourced estimates. We find that where state legislatures lack the capacity for effective oversight, police agencies systematically underreport their lethality, annually concealing hundreds of killings from the official record. Finally, we analyze the effect of a state policy intervention on deadly force reporting, demonstrating that substate agencies are responsive to state legislative intercession.

In summary, we present an intuitive argument for how hierarchical governing structures can reduce informational asymmetries that may inhibit policy making or allow unelected agents to operate with undesirable levels of autonomy. We then present evidence that the capacity for state legislative oversight can increase the transparency of state, county, and municipal agencies. Our argument and findings add to a growing literature on how legislative capacity shapes both the influence of bureaucrats over policy outcomes (e.g., Kroeger Reference Kroeger2022) and the ability of legislators to manage agency behavior (e.g., Boehmke and Shipan Reference Boehmke and Shipan2015; Lillvis and McGrath Reference Lillvis and Mcgrath2017). Additionally, our analysis contributes to research on “democratic deficits” in the American states (e.g., Landgrave and Weller Reference Landgrave and Weller2020; Lax and Phillips Reference Lax and Phillips2012) by demonstrating that deficiencies in legislative capacity allow for less transparent and therefore less accountable agencies. Finally, we believe that our argument begins to rectify a shortcoming in existing interdisciplinary “institutional” analyses (Thompson [1967] Reference Thompson2017) of police, highlighting the dearth of attention paid to electoral institutions and the potential for “legal accountability” (Romzek and Dubnick Reference Romzek and Dubnick1987) vis-à-vis police. We hope that others will follow in this vein, and we encourage our colleagues to devote greater attention to the “second face of the American state” by both applying classic political economic models of governance and innovating new approaches to the study of police (Soss and Weaver Reference Soss and Weaver2017, 567).

On police behavior more directly, our results illuminate the extent of police lethality and reveal systematic deficiencies in public records. Though government manipulation of official data is a well-recognized problem in autocracies (Hollyer, Rosendorff, and Vreeland Reference Hollyer, Rosendorff and Vreeland2011), we show that these problems are also manifest in the United States—official statistics on crime and policing are systematically biased and cannot be taken at face value. Unfortunately, this means that thousands of academic articles on crime and policing are likely to be fundamentally flawed. This underscores the importance of understanding the potential limitations of official data to scientific research, especially where government officials may have incentives to conceal or manipulate data. The Trump administration provides a recent example as evidence indicates attempts to manipulate data on climate change (Friedman Reference Friedman2019), crime (Malone and Asher Reference Malone and Asher2017), and public health (Stolberg Reference Stolberg2020). We conclude the manuscript by encouraging our colleagues to study the political economy of data gathering and dissemination.

Engaging Interdisciplinary Police Research

Research on policing is extensive and interdisciplinary, yet it largely fails to consider how interbranch politics affect law enforcement. Borrowing the Thompson ([1967] Reference Thompson2017) model of organizations, this research can be classified into three broad categories: “technical,” how police perform discrete tasks; “managerial,” how police organize themselves; and “institutional,” how police interact with nonpolice stakeholders, which is our focus. To date, the vast majority of institutional research focuses on what Romzek and Dubnick (Reference Romzek and Dubnick1987) call political accountability—answering to residents—as opposed to legal accountability—answering to elected representatives—despite the fact that police (excepting elected sheriffs) are not directly accountable to residents and elected representatives have greater power to shape police behaviors. The existing literature’s relative inattention to how political institutions influence policing presents an opportunity for substantial contribution by political scientists. In the following, we briefly review police research in each of these main categories, arguing that more attention needs to be paid to role of elected representatives as central actors in the police ecosystem, especially by political scientists.

Institutional (or interactive) studies seek to situate our understanding of police within a broader societal context. Most of this work focuses on sentiment toward police in the mass public, including research in criminology (Hawdon, Ryan, and Griffin Reference Hawdon, Ryan and Griffin2003), law (Sunshine and Tyler Reference Sunshine and Tyler2003), sociology (Desmond, Papachristos, and Kirk Reference Desmond, Papachristos and Kirk2016), and more.Footnote 1 The political science research on policing also tends to focus primarily on this police–resident perspective (see Soss and Weaver Reference Soss and Weaver2017)—for example, studying the sociopolitical implications of police contact (e.g., White Reference White2019)—rather than on police interactions with the legislatures (or even executives) to whom police are accountable. Even when research specifically targets police oversight, one does not find application of general theory on the role of elected representatives and their oversight competencies. Instead, this research tends to analyze specific case studies of civilian complaint processing (e.g., Goldsmith Reference Goldsmith1991).

To demonstrate this point, consider Prenzler and Ronken’s Reference Prenzler and Ronken2001 typology of police oversight models. The authors outright dismiss the possibility of oversight by elected representatives, instead classifying models into three categories: (a) internal affairs (i.e., oversight bodies housed within-agency), (b) civilian review (i.e., external bodies reviewing internal oversight practices), and (c) civilian control (i.e., external bodies with investigation and sanctioning authority). To a political scientist, dismissing the role of elected representatives is a critical omission because representatives are the voice and hands of citizens in government—the possibility of civilian control of police, or any other bureaucracy, in absence of action by elected representatives is vanishingly small. As we elaborate below, though many local elected representatives lack the capacity for meaningful oversight, state legislatures often do have sufficient power to influence police behaviors. Here we follow and echo Soss and Weaver’s Reference Soss and Weaver2017 advice to study police as part of government, incentivized by and accountable to other governmental actors.

Political science engagement with policing research should not be limited to institutional studies alone, as each arm of policing literature builds upon and complements the others. For example, technical studies focus on the execution and efficacy of police activities, including research on deescalation tactics (Todak and James Reference Todak and James2018), race–gender bias in policing (Ba et al. Reference Ba, Knox, Mummolo and Rivera2021), the effects of police staffing on crime intervention and deterrence (Levitt Reference Levitt1997), and so on. These studies inform institutional research in general by showing us what police actually do and our research in particular by clarifying the need to consider features of communities (e.g., racial composition) and agencies (e.g., staffing) in our empirical analysis. With few exceptions (including Levitt Reference Levitt1997), these studies tend to take policy as given but, moving forward, must consider the broader political environment that produces policy and structures incentives for compliance. For example, Mughan, Li, and Nicholson-Crotty (Reference Mughan, Li and Nicholson-Crotty2020) present evidence that electoral accountability may deter sheriffs from abusing property-seizure policies.

Managerial studies, which focus on within-agency organizational choices (e.g., Goldstein Reference Goldstein1963), have clear connections with institutional research. For example, Farris and Holman (Reference Farris and Holman2017) show the political orientations of sheriffs shape their departments’ immigration enforcement strategies. But these studies, too, tend to think of agencies in a vacuum. Even reform-oriented research, including analyses of “community policing” (Weisburd and Braga Reference Weisburd and Braga2019) or sociologically oriented research on “police culture” (Campeau Reference Campeau2015; Crank Reference Crank2014), come largely from the managerial perspective, implicitly assuming only police themselves have influence. Ultimately, however, agency heads are externally accountable, and we need to consider how this structures their incentives in making decisions that guide subordinate behavior. For example, recent work by Eckhouse (Reference Eckhouse2021) nicely demonstrates the potential gains from marrying managerial and institutional approaches, offering a theory of how police (re)shape their behavior in response to monitoring regimes.

Moving police research forward requires deeper integration across these literatures and more meaningful cross-field collaboration. Political scientists can aid these efforts by recognizing police departments as government agencies and offering theories of how their location within this broader institutional context structures police incentives. For example, legislatures can issue policy-specific mandates, oversee police operations, alter police budgets, etc. Therefore, political scientists must bring the legislature into our understanding of policing. In the next sections, we participate in this effort by considering the hierarchical structure of government in which police operate, the potential policy tools that are at the disposal of legislature vis-à-vis police, and the de facto budget constraints on potential legislative oversight activity.

Legislative Capacity and Bureaucratic Control

Under nearly all modern governing arrangements, the elected representatives that write, propose, and pass policies into law delegate the enforcement of these laws to unelected agents. The level of autonomy enjoyed by these agents is a function of the ex ante, de jure discretion afforded to them in implementing policy. Importantly, however, this autonomy is also a function of these agents’ de facto ability to resist ex post oversight. Although nearly all democratic legislatures possess constitutional authority for both ex ante design and ex post oversight, a legislature’s ability to use these tools is subject to a budget constraint: legislative capacity. Legislatures’ endowment of resources (i.e., capacity) for conducting legislative work such as gathering information, writing comprehensive proposals, practicing oversight, and, if necessary, issuing sanctions to exert control over wayward agencies vary considerably. This is familiar in US state politics research, where state legislative capacity has been tracked for several decades (e.g., Bowen and Greene Reference Bowen and Greene2014; Squire Reference Squire2007).

Recent research indicates that higher-capacity legislatures design more detailed proposals that leave less room for bureaucratic discretion (Huber and Shipan Reference Huber and Shipan2002; Vakilifathi Reference Vakilifathi2019). Higher-capacity legislatures also conduct more rigorous oversight of the bureaucracy, monitoring implementation for concordance with legislative intent (Boehmke and Shipan Reference Boehmke and Shipan2015; Lillvis and McGrath Reference Lillvis and Mcgrath2017). As a result of this design and oversight capacity, voters in states with well-resourced legislatures are more likely to get the policy outcomes they desire (Fortunato and Turner Reference Fortunato and Turner2018; Lax and Phillips Reference Lax and Phillips2012). We continue in this emergent stream of research, identifying the potential role of state legislative capacity for oversight of county and municipal agencies.

State Legislative Oversight of Police Agencies

The level of discretion police officers enjoy is extraordinary, as are the potential consequences of their actions (Lipsky Reference Lipsky1980). Even routine, relatively minor decisions—for example, assessing parking fines—can be costly for citizens. At the extreme, officer discretion is a matter of life and death, as police are permitted to subdue civilians with force, including the use of electroshock weapons and firearms. Although discretion may be necessary given the complex and unpredictable environments in which police operate, the severity of the potential consequences for residents from engagement with police also demands robust oversight. This need for discretion and oversight creates a well-known problem with a similarly well-known potential solution in principal-agent arrangements: it can be prohibitively costly to monitor police in real time, but ex post scrutiny may be applied to compel preferred behaviors (McCubbins and Schwartz Reference McCubbins and Schwartz1984).

Transparent reporting is one such behavior. We assume that legislatures generally prefer transparency because they (a) require information for policy design and (b) cannot manage agency actions in a state of ignorance. This information includes statistics on crime, records of calls for assistance and corresponding dispatch, agency expenditures, the use of force by officers, etc. Without this information, representatives cannot develop high-quality policy instruments or ensure the conformity of policy execution with legislative intent. Furthermore, legislators have a strong interest—both intrinsic and electoral—in protecting the public welfare. This requires not only creating effective anticrime policy but also constraining abuse of the state’s monopoly on violence. Faithful reporting by police is necessary to successfully pursue these objectives.

Conversely, police agencies have strong incentives toward opacity. First, existing theoretical research argues that the executive can leverage its informational advantage to obtain greater resources (e.g., Banks and Weingast Reference Banks and Weingast1992). Qualitative studies of police budgeting strategy support this theoretical consensus: informational advantage is critical and agency heads report exploiting it several times over in interviews with Coe and Wiesel (Reference Coe and Wiesel2001). Second, keeping the legislature in the dark allows agencies to operate with autonomy, able to enforce policy as they see fit, and therefore extract benefits from drift. This is true for both on-the-ground execution by individual officers and broad administrative priorities by agencies, as research indicates that police are averse to information-intensive, line-item budgets that constrain spending discretion (Rubin Reference Rubin2019). Third, opacity allows agencies to operate with impunity, free from fear of reprisal when drifting from their mandate. In the case of police agencies such drift is particularly onerous given the potential costs of officer abuse.

Police have strong incentives for opacity regarding use of force in particular, as the revelation of police brutality or killings may draw public ire and, with it, the potential for official sanctions; offending officers and their supervisors more likely to be disciplined, dismissed, or face criminal charges in such climates. These incentives are so strong that police agencies have been accused of falsifying coroner’s reports to conceal their lethality (Egel, Chabria, and Garrison Reference Egel, Chabria and Garrison2017). Evidence indicates police willingness to conceal abuses in other areas as well: inflating clearance rates (Yeung et al. Reference Yeung, Greenblatt, Fahey and Harris2018), timecard manipulation (Rocheleau Reference Rocheleau2019), indiscriminate property seizure (Mughan, Li, and Nicholson-Crotty Reference Mughan, Li and Nicholson-Crotty2020), and so on. Incentives for opacity extend beyond concealing abuse of office, as recent research by Eckhouse (Reference Eckhouse2021) shows that police systematically manipulate crime statistics, reclassifying rapes as “unfounded,” to improve their metrics.

Apart from deselecting agency leaders, a power that state legislatures only enjoy directly in reference to state-level agents,Footnote 2 what can legislatures do to encourage transparency? Generally, legislatures have two instruments: sunshine and budget. With the former (sunshine), legislatures may use investigatory powers to discover and publicize missteps to shame agencies into better performance. All state legislative chambers have the authority to subpoena records on expenditures and assets, formal and informal procedure, internal communication, and other data critical to overcoming their informational disadvantage. Using subpoena power, they may also compel testimony from rank-and-file bureaucrats, agency leaders, and outside experts in order to help decipher the information they have gathered or to report on specific decisions, events, and procedures. These testimonial periods serve to generate additional information and/or shame bureaucrats, which elsewhere has been shown to be effective in promoting compliance (e.g., Hafner-Burton Reference Hafner-Burton2008).

These are well-designed coercive instruments; thus, the threat of subpoena—backed by the power of contempt—is typically sufficient for compliance with the legislature’s demands. Indeed, a “direct threat” of scrutiny or an explicit ultimatum is rarely required for agency compliance. Instead an “indirect threat” of scrutiny, produced by a legislature that has, over time, created a culture of compliance by engaging in rigorous oversight, can compel preferred behaviors. In other words, oversight of any agency in the state creates positive externalities for compliance among every agency in the state, as it establishes a legislature’s reputation for scrutiny.Footnote 3 Of course, legislatures are at times forced to use their coercive powers. For example, in 2018, the Maryland General Assembly formed a special committee to investigate the Baltimore Police Department’s gun-tracing task force that used its subpoena power extensively (Fenton Reference Fenton2020). That same year, the Florida House issued a series of subpoenas to the Broward County and Palm Beach County Sheriff’s Offices in its investigation of police conduct during the Parkland shooting (Koh Reference Koh2018) and the Nebraska legislature subpoenaed the head of its State Department of Correctional Services to deliver extensive information and testimony regarding its lethal injection protocol (Duggan Reference Duggan2018).

These investigations may reveal issues that require legislative solutions to guide future behavior including the provision of additional resources to agencies, the closure of statutory loopholes, etc. Should agencies remain recalcitrant, however, the legislature may employ its ultimate coercive power: reduced agency funding (or the threat thereof). Importantly, the budgetary powers of state legislatures apply to state and substate agencies because states provide direct funding to county and municipal police, as they do for county and municipal public health clinics, schools, etc. In 2017, 32% of total local government revenue came from state transfers (Tax Policy Center 2017), providing great leverage to state legislatures in coercing preferred behaviors from substate actors. Moreover, local governments actively lobby states for these resources on a regular basis (Payson Reference Payson2020), affording legislatures repeated opportunities for influence.

Although all state legislatures possess the de jure powers of oversight and budgeting, their de facto capacity to engage in these behaviors—their time and resources for legislative work—varies considerably. Squire (Reference Squire2007) documents these resource endowments by tracking the degree to which legislatures differ in session length, representative salaries, and staff allocations, which determine (respectively) the time available for legislative activity, members’ ability to invest in expertise (or become “full-time” legislators), and the labor available to legislators for information gathering, etc. For example, although representatives in California have ample resources for investigating and scrutinizing various agencies, representatives in New Hampshire are resource starved. California’s legislators are paid an annual salary of $114,877 plus $206 for each session-day, have an average of 18 staffers, and meet nearly every business day. New Hampshire General Court members, on the other hand, are paid just $100 per year total, share 150 staffers between 424 members, and meet fewer than 25 days per year. Our central expectation is that agencies will be more forthcoming—that is, police will share more information—when the legislature’s capacity for compliance-inducing behaviors is greater. Therefore, when the legislature’s resource endowment increases, so too will agencies’ tendency to transmit information.Footnote 4

We have made the case that state legislatures require agency transparency, particularly in the case of crime and policing, but must also establish why compelling transparency from local agencies requires state legislatures—why must state legislatures direct their oversight capacity downward to county and municipal agencies in order to get these data? After all, county boards and municipal councils are the representative bodies to which these agencies are immediately accountable. First, many city and county legislatures may lack the capacity for meaningful oversight, as these bodies meet infrequently and their members are often part-time workers. For example, data on city councilor compensation in California and Florida reveal that the median councilor is paid just $5,562 annually, with 78% falling under the national poverty line and 11% paid nothing at all. Second, local bodies may lack the political will to oversee their own police agencies, necessitating state legislative involvement. Councilors may be unwilling to scrutinize their police agencies for fear of electoral reprisals, producing a kind of regulatory capture. Zimring (Reference Zimring2017) discusses these local conflicts of interest regarding police accountability at length, and political science research shows that police use their localized influence to extract substantial rents from municipal governments (Anzia and Moe Reference Anzia and Moe2015). These problems dissipate at the state level as membership is geographically diverse, allowing representatives from one city to lead scrutiny of police in another.

Taken together, these factors reveal that local oversight may prove insufficient and state intercession may often be necessary, especially given the inherent complexities in evaluating police behaviors and budgeting practices. This may help explain why previous research has shown that local policing is responsive to state-level political processes at similar (Chaney and Saltzstein Reference Chaney and Saltzstein1998) or greater levels (Levitt Reference Levitt1997) than to municipal politics. Thus, if state legislatures play the role we argue, the effects of legislative capacity should be detectable at the state, county, and municipal levels.Footnote 5

Further discussion of this relationship, including more detailed description of state legislatures’ investigatory powers and additional examples of state legislative oversight of police agencies, at all levels, across 11 states, is given in Appendix A. We also show a strong, positive relationship between legislative capacity and legislative responsiveness to the George Floyd protests in May–June of 2020. Collectively, this discussion supports our argument that strong state legislatures often assume an oversight role vis-à-vis local agencies, which we test empirically in the following sections.

Research Design

We evaluate our central expectation—that the transparency of state and substate police agencies is increasing with state legislative capacity—in two main studies.Footnote 6 Both rely on data from the Department of Justice’s (DOJ) Uniform Crime Reporting (UCR) program, which is administered by the Federal Bureau of Investigation (FBI 2020):

The FBI’s Uniform Crime Reporting (UCR) Program serves as the national repository for crime data collected by law enforcement. Its primary objective is to generate reliable information for use in law enforcement administration, operation, and management. Considered a leading social indicator, UCR data is used for monitoring fluctuations in crime levels, evaluating policies, and regulating staffing levels. In addition to the American public relying on the data for information, criminologists, sociologists, legislators, city planners, the media, and other students of criminal justice use them for a variety of research and planning purposes.

Each year, the UCR provides every police agency in the US with a detailed framework for classifying, counting, and reporting events, including, but not limited to, homicide, so-called justifiable homicide by a police officer, sexual assault, and “clearances.” Importantly, participation is voluntary—at the discretion of individual agencies—with many agencies choosing not to participate. Moreover, the DOJ does not audit these data or sanction individual agencies for nonparticipation. This is important for the research design because it means that all police agencies receive an identical stimulus from the DOJ but no centralized incentive to comply.

Given the voluntary nature of the program, UCR data are often (rightly) maligned as inadequate for assessing patterns of crime. For example, Maltz and Targonski (Reference Maltz and Targonski2002, 299) note the extent of data missingness and conclude that the “crime data, as they are currently constituted, should not be used, especially in policy studies.” Regarding killings by police, specifically, Zimring (Reference Zimring2017, 29) writes that “the voluntary nature of the reporting system means that a significant number of killings by police do not get included in the official numbers.” Although we discuss these limitations in the following sections, it is important to note that, for the purposes of our study, this is a feature and not a bug. We are interested not in explaining variation in crime itself but variation in data sharing by police. Therefore, it is imperfect compliance with the UCR program that enables us to effectively test our argument on the role of legislative capacity. Specifically, compliance with UCR requests should be greater for agencies facing the threat of legislative scrutiny, the credibility of which is increasing in the state legislature’s oversight capacity. From this perspective, the UCR data requests are analogous to an annual audit study of all police agencies where the treatment is time- and state-varying legislative capacity.

We outline above why legislatures would prefer police agencies to share data, but why would legislatures want to compel participation in the UCR in particular? UCR participation has substantial efficiency advantages over alternative means of data collection, at least from the legislature’s perspective. When legislatures want information on crime, policing, and public safety, they may either solicit this information on an ad hoc basis—requiring the design, funding, and implementation of data-collection regimes—or they may encourage their agencies to participate in the UCR at a substantial resource savings. Second, UCR data allow for comparison that may be of interest to legislatures. Because it is a national program, states are able to compare outcomes on important metrics using common definitions and criteria. This allows legislatures to, for example, evaluate whether innovations used in other states have proven effective and are worth deeper consideration. With ad hoc, individualized data alone, such comparison is difficult.Footnote 7

Taken together, UCR participation has the potential to yield information that is more valuable than innovating a new program, at a lower cost. This is likely why most states delegate the entire enterprise of crime and policing data collection and dissemination to the UCR. We see this manifest in reliance on UCR statistics in state legislative reports on crime and policing (e.g., Joint Legislative Audit and Review Commission 2008), or state legislative directives to begin reporting certain information into the UCR (Texas SB 349 [2021]), of course, these directives only have teeth in proportion to the legislature’s capacity to compel compliance. When states do create their own programs for data dissemination, they tend to be a state-specific reporting of UCR-submitted data, as in Maryland and South Carolina (Criminal Justice Information Services 2019; Williams Reference Williams2020), or UCR data supplemented by collecting additional information that the UCR does not request. For example, North Carolina’s Highway Patrol traffic stop database was designed to collect a specific type of data—driver demographics and various outcomes of traffic stops made by state troopers—outside the purview of the UCR.Footnote 8 Stated alternatively, the existence and low cost (to states) of the UCR has effectively preempted the development of alternative, state-designed crime and policing data infrastructure.

To estimate the effect of state legislative capacity on police agency transparency, we undertake two empirical studies. First, we evaluate whether agencies are more likely to comply with DOJ data requests as their state legislature’s capacity for oversight increases, by analyzing administrative records indicating UCR participation for all municipal, county, and state police agencies in the 50 statesFootnote 9 between 1960 and 2017 in a generalized difference-in-differences (DiD) framework. The data reveal strong evidence for a plausibly causal effect of capacity on UCR compliance.

Second, we take advantage of several recent initiatives to crowdsource reliable data on police lethality, including well-known projects by The Washington Post and The Guardian, to assess the accuracy of data reported by agencies.Footnote 10 It is widely accepted that these crowdsourced data provide more accurate information on the extent of killings by police than do official statistics (Zimring Reference Zimring2017).Footnote 11 Moreover, these projects are among the only systematic efforts by nonpolice to compile information overlapping the UCR and therefore present a rare opportunity to systematically scrutinize the accuracy of UCR data. Our expectation is that higher legislative capacity will be associated with more transparent reporting of deadly force. The data support this, indicating fewer reporting discrepancies in states with high-capacity legislatures. Finally, we assess the effect of a policy intervention that mandates external investigation of deadly force in seven states and find plausibly causal evidence that state legislative intercession increases police transparency.

Study 1: UCR Compliance

Our administrative data include all state, county, and municipal police agencies in the 50 states from 1960 to 2017, indicating whether or not agencies complied with annual UCR data requests. The data are split into two sample periods—1960–1994 and 1995–2017—because the records before 1995 were obtained as physical media and then parsed from a series of quite large, unformatted (or unusually formatted) text documents.Footnote 12 Therefore, we cannot be completely certain that we have an exhaustive accounting for the 1960–1994 period and analyze the periods separately to avoid contaminating the latter, exhaustive sample. There are 19,095 active agencies across the two periods.Footnote 13

We do not analyze the scope or accuracy of the data submitted in this section, merely the most basic level of compliance: whether the agency provided any information at all in response to the DOJ’s request. This is common in the audit studies that are similar in design to ours. In general, compliance trends upward over the sample period, with about 37% of agencies reporting in 1960 compared with 76% in 2017. Parsing these data further, in the older sample (1960–1994) only 1% of agencies comply every year, 15% never comply, and 84% comply intermittently. In the more recent sample (1995–2017), 48% of agencies comply every year, 8% never comply, and 44% comply intermittently.

Our capacity measure is an adaptation of the Squire (Reference Squire2007) index—the industry standard measure of capacity—which combines the institutional endowments discussed above: legislator compensation, length of legislative sessions, and legislative staff. As Squire notes, the original measure is calculated intermittently and always relative to Congress, making the index inappropriate for dynamic analyses like ours. Given these limitations, we instead gather raw data on legislator compensation, session length, and staff expenditures for each legislative session and scale the values into a summary measure using Quinn’s (Reference Quinn2004) factor analytic model.Footnote 14 Because our adapted measure uses inflation-adjusted dollar amounts and scales all years jointly, it is better suited for our over-time analysis.Footnote 15 In addition to the modified Squire index, we also include the imposition of legislative term limits as an additional proxy for capacity, as term limits cause a sudden and permanent reduction to aggregate experience and institutional memory.

To sum up, the outcome in Study 1 is a police agency’s choice to comply with a UCR data request. The treatment variables are the resource-based legislative capacity measure and term limit imposition, and we use a generalized difference-in-differences design, relying on two-way (state and year) fixed effects, for identification (Angrist and Pischke Reference Angrist and Pischke2008).Footnote 16 We estimate and report the results for several variations of this model on both samples, starting with reduced-form models where capacity and term limits are regressed on UCR participation with only state and year fixed effects.

Next, we generalize and estimate more fully specified models. First, we separately evaluate the influence of capacity on agencies with appointed and elected heads. DeHart (Reference DeHart2020) notes that sheriffs are unusual among American police agency heads, as nearly all of them are elected, whereas state and municipal agency heads are appointed. Being elected rather than appointed should partially insulate sheriffs from state legislative pressures because elections both change the relevant principal and endow sheriffs with greater legitimacy, enhancing public support for the office and the individuals who hold it. To assess whether elected (relative to appointed) agency heads are less sensitive to state-legislative scrutiny, we use DeHart’s 2020 classification of all US counties as electing or appointing their sheriffs and interact an indicator for elections on our covariates of interest.

Second, given the low risk of simultaneity (i.e., UCR compliance having an effect on legislative capacity)Footnote 17 and the careful attention paid to measurement, the biggest threat to sound inference in our analysis is potential confounding. Although the DiD estimator should be reasonably robust to the omission of potential confounders (e.g., Angrist and Pischke Reference Angrist and Pischke2008), we sought to further assuage potential concerns that our results are driven by omitted variable bias. Therefore, we gathered data to account for potential confounders including indicators for agency type (city, county, state, and whether or not the sheriff’s office is noted in the constitution, following Falcone and Wells [Reference Falcone and Wells1995]),Footnote 18 the size of the population to which the agency is accountable, how many officers and civilians each agency employs, government partisanship, and the percentage of Black residents. In addition to our measure of government partisanship, we also account for time-varying perturbations from party control by including Party × Year fixed effects—where the former indicates either Democratic or Republican (relative to divided) control of the legislature and the latter indicates the year of observation. Collectively, these covariates help account for local oversight capacity, agencies’ resources for compliance, and political preferences, among other factors.

In Table 1 we present estimates from reduced-form and fully specified DiD models for each of the two samples. The estimates reveal strong evidence for an effect of state legislative capacity on UCR participation and therefore strong support for our argument that capacity encourages police transparency. In all models, the point estimates are positive, large, and many times greater than their standard error estimate. Substantively, the effect of capacity is remarkable. For example, in the latter sample, if all legislatures had resources equivalent to those of Texas, currently about the mean level of capacity, and increased their resources to the level of one of the higher-capacity legislatures, like Michigan—which would not be an unprecedented change (Squire and Hamm Reference Squire and Hamm2005)—the predicted effect on the probability of UCR participation would be an increase of over 0.06. This translates to an expectation of between 1,200 and 1,300 more agencies complying with UCR data requests each year. Footnote 19

Table 1. Legislative Capacity on UCR Participation

Term limit imposition also yields the anticipated effect, sharply reducing compliance probability by about 0.07 in the earlier sample (where there is less variability on the covariate) and 0.17 in the latter sample. Furthermore, the effect of legislative capacity is muted for agencies with elected heads in the latter sample and washed out completely in the earlier sample. Likewise, term limit effects are muted for elected heads in the latter sample, but the interaction is null in the earlier sample. The data suggest election, relative to appointment, seems to insulate some heads from legislative scrutiny.

In Appendix B, we report the results for all control variables and show that central result is robust to an array of alternative specifications including interacting capacity with partisan legislative control; analyzing state, county, and municipal agencies separately; and employing an alternative measure of capacity derived from Vannoni, Ash, and Morelli (Reference Vannoni, Ash and Morelli2021). The effect of capacity remains positive, substantively large, and statistically significant under all conditions: increasing state legislative capacity is associated with increases in state, county, and municipal police agency transparency. Finally, we replicate our analysis by examining compliance with the 2016 iteration of the Law Enforcement Management and Administrative Statistics survey (administered by the Bureau of Justice Statistics) and find the central relationship reported here manifest in those data as well: police agencies in states with higher-capacity legislatures are significantly more likely to comply with requests for data.

Although there may, of course, be some additional risk of confounding despite our various checks, the robustness of our results to the inclusion of a variety of observed potential confounds gives us confidence in the direction and relative size of our effects. Overturning these results would require an unmeasured time-varying variable that (a) is jointly related to participation and capacity (i.e., relevant to the main effect), (b) has heterogeneous over-time variation across states (i.e., does not produce a trend proxied by the year fixed effects), and (c) is not already proxied for by the current time-varying controls. Furthermore, sensitivity analyses following Cinelli and Hazlett (Reference Cinelli and Hazlett2020) reveal that such a confounder would need to predict both capacity and UCR participation over 10 times better than does our most predictive control variable (population, which is correlated with compliance: β > 0.06 and t > 50 and in both samples) to wash out the capacity effect. Although this is possible, we cannot think of particular covariates satisfying these conditions; otherwise, we would have included them in the model.

At minimum, we have shown that there are systematic differences in UCR participation across states and that these are robustly correlated with state legislative capacity. Although the magnitude of this effect may be better identified in subsequent studies, we believe that first revealing this relationship provides substantial value for researchers in two ways. First, our analysis suggests that political oversight is a fruitful area for future research, as political scientists increasingly analyze policing. Second, our analysis provides analysts with information on where, and a plausible explanation for why, data on crime are likely to be of lower or higher quality. Although we do not study crime itself as an outcome here, our analysis has clear implications for research that does.

Study 2: Underreporting Deadly Force

The UCR is the only comprehensive, national effort by the US government to build an accounting of police lethality, but these data are inaccurate.Footnote 20 The intermittent UCR compliance discussed above makes it clear that any accounting of crime or police activity contained in the UCR is incomplete. Furthermore, given the evidence that noncompliance is determined in part by political concerns and recognition of the substantial incentive police agencies have to obfuscate (or exaggerate) certain behaviors and outcomes, the information that is contained in the UCR is almost certainly inaccurate and biased toward reflecting favorably on police agencies. Such bias is likely acute in the case of police use of lethal force or what the UCR calls “justifiable homicide by law enforcement,” and indeed we find that underreporting is pervasive.

Unfortunately, UCR estimates of police lethality have been the standard in academic research for decades. Jacobs and O’Brien (Reference Jacobs and O’Brien1998, 846) went so far as to defend the data as “likely to be accurate because such homicides are difficult to conceal.” As recently as Holmes, Painter, and Smith (Reference Holmes, Painter and Smith2019), these data have been used as credible estimates of police lethality despite discussions of the well-known limitations of these data in academic research (e.g., Klinger Reference Klinger2012) and in the documentation of the data themselves.Footnote 21 Though previous research recognizes UCR shortcomings, recognition alone is insufficient. The fact that the wealthiest and most durable democracy in history does not know how many of its residents are killed by agents of the state is alarming and warrants study such that we may understand the political processes that produce (or permit) this downward bias in the official statistics and how the bias may potentially be alleviated.

We argue that one political factor driving this opacity, or bias, is the inability of legislatures to shape the behavior of police agencies. We evaluate this argument by comparing the UCR’s official statistics to deadly force counts from several crowdsourced initiatives organized in recent years: The Guardian’s The Counted (2015–2016), The Washington Post’s Fatal Force (2015–2016), Mapping Police Violence (2013–2016), Killed by Police (2014–2016), and Fatal Encounters (2013–2016).Footnote 22 Although not ground-truth data, these projects represent the best available estimates on the extent of police killings. Even former FBI Director James Comey has recognized the greater accuracy of these crowdsourced data, calling the UCR statistics “embarrassing and ridiculous” by comparison (Kindy Reference Kindy2015). A central feature of these data is that each case is independently vetted and confirmed, and indeed an independent report issued by the DOJ suggests they provide a superior accounting (Banks Reference Banks, Ruddle, Kennedy and Planty2016). Thus, discrepancies between UCR and the crowdsourced data provide an estimate of (known) undercounting—the number of known police killings that could have been officially reported but were not. Assessing difference across sources in this manner is common in event data analysis of political violence (e.g., Davenport and Ball Reference Davenport and Ball2002).

We plot these data in Figures 1 and 2 to demonstrate the systematic underreporting of police killings in the UCR. Figure 1 gives aggregate lethality figures by year, and there are two clear patterns. First, the figures are relatively consistent across time within reporting group. Second, the UCR underestimates police killings by at least 50% relative to crowdsourced estimates every year. We note that Fatal Encounters is a high outlier because it counts “deaths in custody” (e.g., dying of a heart attack while in custody awaiting arraignment) in addition to “justifiable homicides.” Though Fatal Encounters’ counts are likely to be biased upward as a result, these estimates are still useful to include in our comparison set, especially because we know the source and direction of the bias.

Figure 1. Police Killings by Source (2013–2016)

Figure 2. Comparing 2015 Police Killings per 100,000 Residents across the United States

Figure 2 decomposes the 2015 estimates to demonstrate state-level differences in reporting on the incidence of police killings per 100,000 residents. Again, there are two central patterns. The first is the striking degree of variability in the incidence of police killings across the United States. For example, police killings are a relatively rare occurrence in Connecticut but an order of magnitude higher in New Mexico. In 2015, police lethality in all crowdsourced data sources is also substantially greater than the official estimates of drug-related homicides (452) and gang-related homicides (757). The second main pattern is the substantial variability in degree of underreporting. For example, Michigan reports almost all known police killings, whereas Florida fails to report any.

This variability in underreporting is the focus of Study 2, as we argue that a significant portion of it can be attributed to systematic differences in capacity for legislative oversight. To test this, we undertake two additional analyses. First, we compare rates of underreporting of police lethality across states to the states’ legislative capacity. The data indicate that states lacking the capacity for oversight have significantly higher numbers of unreported police killings. Second, we provide evidence for a plausibly causal effect of state legislative intercession on police transparency by analyzing agency-reported deadly force numbers pre–post a policy intervention in seven states. This analysis suggests that the intervention significantly reduces the number of unreported police killings.

To measure the underreporting of police lethality, we calculate the difference between the publicly gathered data and the data contained in the UCR in each state-year (e.g., WaPost AZ,2016 UCR AZ,2016 )—the number of known police killings unreported in official statistics, scaled per 100,000 residents.Footnote 23 The effective sample period for these data is 2013–2016, which limits our modeling options because capacity does not vary meaningfully within states over this period and term limits do not vary at all. Given the lack of within-unit variation in these main predictors, we focus on modeling the cross-sectional variation and, therefore, do not include state fixed effects. Instead, we use state random effects to account for within-unit correlation in unobservables and additional control variables to account for potential state-specific confounders. Theoretically, these are the state’s available resources for collecting and disseminating information on police killings and the preferences of police agencies and legislatures for transparency. Operationally, these enter the model as the state’s wealth (median income), the average number of agency employees (officers and civilians), share of Black population, the partisan control of government, and the political preferences of police agencies. Agency conservatism is captured using Bonica’s (Reference Bonica2016) preference estimates derived from campaign contribution data, and we use the mean estimated ideal point for all contributors working in law enforcement, where greater values indicate greater conservatism.Footnote 24 Research on the connection between political preferences and preferences for police militarism or violence suggest this is the best available measure (Christie, Petrie, and Timmins Reference Christie, Petrie and Timmins1996; Gerber and Jackson Reference Gerber and Jackson2017). Baseline categories for legislative and gubernatorial control are divided control and independent, respectively.Footnote 25 , Footnote 26

Before turning to our results, it is important to recall precisely what we are modeling: a police agencies’ propensity to report a killing, given that the killing has occurred. Therefore, the relevant covariates to include are predictors of the choice to report—that is, preferences and resources—not predictors of the event itself, such as local rates of violent crime. Importantly, in addition to being an irrelevant control variable, any official measure of crime would be subject to the same deficiencies as are the UCR estimates of police lethality and would therefore induce estimation bias.

The results are provided in Table 2, with separate models for each crowdsourced database. All covariates are standardized to facilitate interpretation, with each coefficient indicating the effect of a one-standard-deviation (or binary) increase in the predictor. As can be seen, state legislative capacity is consistently associated with less underreporting in official statistics, as agencies in states with higher legislative capacity are more likely to enter police killings in official reports across all data sources. The results indicate that a one-standard-deviation increase in legislative capacity is associated with around 0.035–0.075 fewer unreported police killings per 100,000 residents—or an average of 2.2–4.8 per state annually. Converting all states from their true capacity to the sample maximum (California) is predicted to reduce unreported killings by 284–670 per annum. Also, as expected, the estimate from the Fatal Encounters analysis is smaller and has a larger relative standard error, reflecting our qualitative understanding that these data bias undercounts upward from the true value.Footnote 27 Finally, the term limits indicator, which we would expect to have a positive effect on underreporting, has an effective zero estimate in all models. The strong negative relationship between transparency and term limits revealed in the first study is not manifest here.

Table 2. Police Killings Unreported to UCR

To be clear, causal identification in this analysis is complicated by the lack of random treatment assignment, the sparsity of the outcome data, the time invariance of the main predictors, and our inability to include state fixed effects. Each precludes standard design- or model-based strategies for casual inference. Although reasonable people can disagree on what can be drawn from observational data analysis without ideal identification strategies, at minimum our analysis makes novel use of crowdsourced data to identify the set of states where police killings are most likely to be underreported and demonstrates that this underreporting is significantly negatively correlated with legislative capacity. Should researchers have alternative (i.e., noncapacity) theories explaining this relationship, we strongly encourage them to pursue this in future research. To further investigate our capacity-based argument, however, we can assess the causal effect of legislative intercession on the transparent reporting of deadly force, even if we cannot assess the causal effect of legislative capacity per se, which is of course an antecedent factor to intercession.

Since 2011, California, Colorado, Connecticut, Illinois, New York, Utah, and Wisconsin have implemented policies requiring and regulating the investigation of officer-involved deaths.Footnote 28 Consistent with our general argument, these laws are positively correlated with legislative capacity ( $ p\hskip.18em = 0.017 $ ). Each of these laws prescribes a systematic investigation protocol for officer-involved deaths (or use of force) by actors outside of the officer’s agency and mandates the application of the protocol. Therefore, these laws prescribe a level of transparency and formal memorialization of the events. This both reduces the incentive to obfuscation in official statistics by expanding the number of actors with knowledge of the event and increases the probability of being held accountable for underreporting. Our empirical expectation is that this policy intervention should reduce the number of unreported police killings. To test this, we estimate a series of DiD models where the outcome is the difference between the publicly gathered data and UCR data in each state year (e.g., WaPost AZ,2016 UCR AZ,2016 ). Our covariate of interest is the intervention status for the state-year. We expect the treatment to have a negative effect—to reduce the number of unreported killings.Footnote 29

Table 3 presents the DiD analyses of the effects of the investigation policy. Every model recovers a negative and significant estimate, revealing evidence for a plausibly causal effect of state legislative intercession on the transparent reporting of deadly force. Substantively, our results suggest that, by imposing investigation policies, the treated states improved the accuracy of police self-reporting by an average of 5 to 10 deaths annually. These results, taken together with the robust negative correlations between capacity and underreporting in Table 2 provide compelling evidence for our overarching argument that endowing state legislatures with the resources for rigorous oversight (or policy intervention) can substantially improve bureaucratic transparency across levels of government. This, in turn, should help legislatures acquire information that is necessary for policy design as well as help direct and constrain the behavior of agencies, ultimately helping align government action with voters’ preferences.

Table 3. Difference-in-Differences Analysis of State Investigation Laws

Discussion

We have argued that state legislatures—which depend upon state, county, and municipal agencies for policy-relevant information—can direct their scrutiny powers downward, helping to compel bureaucratic transparency of state and substate agencies. Thus, increasing state legislative capacity can increase information sharing and ultimately improve legislative management of agencies at each of these levels. An analysis of administrative records of 19,095 police agencies’ compliance with official data requests over nearly six decades yielded strong evidence for a plausibly causal effect of capacity on state, county, and municipal agency transparency. We then assessed the accuracy of reported data by comparing the number of known unreported police killings in each state over recent years to states’ legislative capacity, finding a robust negative association—states with higher-capacity legislatures had significantly lower rates of unreported police lethality. A DiD analysis of a transparency intervention provided complementary evidence that state legislative intercession improves reporting of police lethality.

Taken together, we believe this provides compelling evidence that (a) oversight can ameliorate informational asymmetries between unelected government agents and elected representatives and (b) state legislative oversight capacity can be directed downward to scrutinize substate agencies. This contributes to our understanding of interbranch conflict across levels of government, the broader implications of investing in legislative capacity, and the link between these investments and democratic responsiveness. Furthermore, police present a difficult test of our argument on the potential for state legislative bodies to affect local-agency behaviors. Relative to other agencies, police are generally popular (Gallup 2019) and possess strong unions, which may make police less sensitive to legislative demands for transparency. This suggests that our findings here should generalize well to other departments of the executive.

Even absent these broader implications for oversight, we believe that the specific analysis of policing here is important. As Zimring (Reference Zimring2017) notes, the volume of killings by police is immense, with roughly 1,000 deaths per year, over half of which go unreported in official statistics. This is nearly 40 times the lethality threshold used in some measures of civil conflict. Yet, this particular form of political violence receives comparatively little attention in political science research. We believe that our approach and findings here, complementing Soss and Weaver’s (Reference Soss and Weaver2017) argument to study police within the broader context of governance and democracy, suggest a potentially fruitful path toward better understanding how police behaviors may be brought in line with public sentiment. Our work also contributes to a fast-growing literature on the deep and wide-ranging implications of legislative capacity, an area of legislative research of which scholars have barely begun to scratch the surface. Furthermore, research on the relationship between governing bodies across different levels is growing but overwhelmingly focused on bottom-up interactions (e.g., the lobbying of Congress or state legislatures by municipal governments: Loftis and Kettler Reference Loftis and Kettler2015; Payson Reference Payson2020), whereas this article suggests that top-down interactions also deserve attention.

In addition, we provide evidence that the gathering and dissemination of official data is politically motivated, obscuring our understanding of government action. There are several implications. First, if agency behaviors are conditioned by expectations of scrutiny, then coercing transparency may effectively decrease police violence. Second, insights from previous studies using official statistics on crime and policing should not be taken on faith and may warrant reexamination, especially studies that use UCR data without accounting for the considerable selection bias in their generation and dissemination: given the above analyses, is difficult to think of any circumstance under which UCR data are appropriate for studying the causes of crime. Third, researchers need to consider the data-production process explicitly and apart from the event-generation process (Cook and Weidmann Reference Cook and Weidmann2019), especially when data providers have interests that are inconsistent with transparency.

To see how this matters, consider research by Knox, Lowe, and Mummolo (Reference Knox, Lowe and Mummolo2020) showing official statistics can mask racially biased policing by recording only the conditions of interactions (e.g., the race and speed of a driver an officer has pulled over) and not recording the conditions of interaction opportunities (e.g., the race and speed of all drivers an officer observed). Importantly, the authors’ estimates of the extent to which recorded data obscure the true nature of the event-generating process assumes that the data-generating process is unbiased (e.g., the race and speed of all drivers an officer has pulled over is faithfully recorded). Our analyses show that this almost certainly is not the case, so we may consider the bias estimated by Knox, Lowe, and Mummolo (Reference Knox, Lowe and Mummolo2020) a lower bound, at least in the case of behaviors or events, like racially biased use of force, that bear potential disciplinary costs.

Carefully considering how and why our realized data deviate from ground-truth events is important in areas aside from crime as well. Just as police have incentives to underreport their lethality, school personnel have incentives to manipulate test scores (Jacob and Levitt Reference Jacob and Levitt2003) and the Environmental Protection Agency may face pressure to underestimate pollution-related deaths (Friedman Reference Friedman2019). Moreover, as news media consumption becomes increasingly segmented and local media divestment continues (Peterson Reference Peterson2021), the ability of the press to serve as a watchdog declines. Therefore, better understanding political avenues for ensuring more accurate government data reporting is increasingly important. This matters both for less salient issue areas where auxiliary media-based data are not easily and readily available and, in the future, for high-salience issues as media resources continue to be diminished. Good governance and good science require good data, and we simply cannot rely on media to fill in gaps in the official statistics. Thus, we urge our colleagues to study the political economy of official statistics—critically scrutinizing the accuracy, reliability, and political determinants of official data needed for government accountability and scientific inquiry.

Supplementary Materials

To view supplementary material for this article, please visit http://doi.org/10.1017/S0003055422000624.

DATA AVAILABILITY STATEMENT

Research documentation and data that support the findings of this study are openly available at the American Political Science Review Dataverse: https://doi.org/10.7910/DVN/MFFV3C.

ACKNOWLEDGMENTS

This paper benefited from feedback from participants at the 2019 Annual Meeting of the Southern Political Science Association; workshops at American University, ETH-Zurich, Stony Brook University, University of California, San Diego, and University of St. Gallen; comments from Mallory Compton, Andrew Flores, Gerald Gamm, Mary Kroeger, Casey LaFrance, Josh McCrain, Rob McGrath, Ken Meier, Jonathan Mummolo, Tessa Provins, and Ian Turner; and consultation with Elizabeth Davis and Zhen Zeng at the Bureau of Justice Statistics. We are also grateful to the editor and three anonymous reviewers for constructive comments, criticisms, and suggestions. Any errors of omission or commission are of the authors’ making.

CONFLICT OF INTEREST

The authors declare no ethical issues or conflicts of interest in this research.

ETHICAL STANDARDS

The authors affirm this research did not involve human subjects.

Footnotes

1 The notable exceptions that do focus on agency–legislature interactions tend to focus specifically on budgeting. For example, LaFrance (Reference LaFrance2012) shows sheriffs act more antagonistically than cooperatively in budgeting disputes with legislatures, whereas Coe and Wiesel (Reference Coe and Wiesel2001) provide interview evidence that police covet their informational advantages in budgeting negotiations.

2 State legislatures may be able to deselect substate leaders indirectly by shaming them into resigning, threatening to cut budget transfers, etc.

3 This is an adaptive expectations argument. A more traditional, rational expectations argument would simply be that agencies observe the legislature’s resource endowment, forecast the probability of recourse, and make their transparency choice accordingly.

4 The imposition of term limits also alters a dimension of legislative capacity not captured by these resources: the chamber’s aggregate experience, which is critical to institutional memory and overall expertise. Therefore, we also consider term limits in our analysis.

5 Of course, some local legislatures do have the resources for scrutinizing their police (e.g., Choi and Hong Reference Choi and Hong2021), and we attempt to account for this in our analysis.

6 Cook and Fortunato (Reference Cook and Fortunato2022) provide the replication materials for all analyses presented here and in the Appendix.

7 We acknowledge that comparisons relying on the UCR data may also be faulty; however, whether such comparisons should be undertaken is a normative question outside of the scope of our argument, that state legislatures demonstrate a preference for UCR data is an empirical claim supported below.

8 Note that this database was created in response to the passage of Senate Bill 76 by the state legislature in April 1999, providing additional evidence that state legislatures are interested in and can act to promote transparent reporting from police.

9 That is, all municipal, county, and state police agencies in the United States excepting those in Native American Reservations, US Territories, and Washington, DC.

10 Sumner, Farris, and Holman (Reference Sumner, Farris and Holman2020) advocate for, and demonstrate the validity of, crowdsourced data in studies of local politics more generally.

11 For those interested in details on the different methodologies employed in these distinct crowdsourced data-collection efforts, we recommend chapter 2 of Zimring (Reference Zimring2017).

12 This split also corresponds to an update to the Paperwork Reduction Act, which changed the rules governing the FBI’s internal handling of the information in the earlier and latter periods.

13 This is an unbalanced panel because agencies are created, decommissioned, or merged over time.

14 The data are made available by Bowen and Greene (Reference Bowen and Greene2014). We use the Quinn (Reference Quinn2004) model because it is robust to missing values (and there are a handful). An equal-weights average of imputed resources produces substantively equivalent results.

15 Additional details on the construction of our preferred measure and summary statistics are provided in Appendix B.

16 Legislative resources and term limits are chosen by legislatures or voters and are therefore exogenously imposed on the police agencies whose compliance we are modeling. Term limits, in particular, are imposed by voters via referendum with arbitrary timing and may be safely considered exogenous to police behavior.

17 Note that because UCR data are filed after the reporting period, legislative capacity is effectively lagged, making anticipation the only threat of reverse causality—i.e., that legislative capacity increases in expectation of increased transparency. We find this implausible.

18 Forty-one states have noted the sheriff’s office in their constitutions at any time and thirty-seven retain the note as of 2021. A full accounting is given in Appendix D, where we also discuss constitutional funding mandates for sheriffs.

19 Calculation uses 2008 scores. Taking a conservative approach, as advocated in Mummolo and Peterson (Reference Mummolo and Peterson2018), the effect of an average within-unit standard-deviation (0.275) increase to capacity produces an increase of 0.013, which translates to an additional 247 agencies complying with UCR data requests annually.

20 For police killings there are two other potential official sources: the Centers for Disease Control’s National Vital Statistics System (NVSS, which is not designed for police lethality accounting) and the Bureau of Justice Statistics’ (BJS) Arrest Related Deaths (ARD). ARD was discontinued in 2014 after the BJS concluded that ARD did not meet its data-quality standards and systematic underestimation in both ARD and NVSS was confirmed by Barber (Reference Barber, Azrael, Cohen, Miller, Thymes, Wang and Hemenway2016). Although a new program by the CDC—the National Violent Death Reporting System—offers some improvements, this project covers only a selection of states and still appears to substantially underestimate.

21 We note that many scholars are now turning to city-level data, often gathered internally and reported voluntarily or through Freedom of Information Act requests. Knox, Lowe, and Mummolo (Reference Knox, Lowe and Mummolo2020) provide an excellent explanation of the bias in the event-generating processes that underlie these data, but we urge further caution regarding bias in the data-generating process—they are subject to the same selection biases when being recorded and reported as UCR data.

22 Additional details on these data are available in Appendix C. Our sample stops at 2016 because UCR data on police killings disaggregated to the state level are presently unavailable from 2017 onward.

23 We aggregate to the state level because most agencies do not have an instance of deadly force in any given year, overdispersing the data with 0 opportunities (not 0 counts)—conflating cases in which all deaths were faithfully reported and cases in which there were no deaths to report.

24 We use the 2014 electoral cycle estimates.

25 Nebraska is coded as Republican, reflecting candidate endorsements. Switching this coding does not change the results, nor does substituting median legislator ideology for partisan control. This can be seen in Appendix C.

26 We do not include a measure of UCR participation because it is posttreatment. However, the only role UCR participation could play is to mediate the effect of capacity on unreported police killings—i.e., identify the extent of capacity-related underreporting that is due to nonparticipation. In Study 2, we are interested in the total effect of capacity on unreported police killings. This is what our coefficient captures, with an understanding that some of this effect is likely due to UCR nonparticipation because participation is positively related to capacity (Study 1).

27 Although not the focus of our analysis, the other results lend validity to our findings. Income, proxying for collection and dissemination resources, is negatively related to underreported killings. Furthermore, although, as expected, partisan control of government has no discernible effect on underreporting, the effect of police ideology is positive, large, and statistically robust in all models.

28 Hawaii has also passed such a law, but its passage falls outside our sample period. Data on investigation laws were obtained from the National Conference of State Legislatures.

29 Note that the outcome is not scaled in these models, as this effectively interacts the treatment with population changes, breaking the identification strategy. Controlling for population on the right-hand side of the model does not substantively change the estimates on the treatment, nor does including the other control variables from Table 2.

References

REFERENCES

Angrist, Joshua D., and Pischke, Jörn-Steffen. 2008. Mostly Harmless Econometrics: An Empiricist’s Companion. Princeton, NJ: Princeton University Press.CrossRefGoogle Scholar
Anzia, Sarah F., and Moe, Terry M.. 2015. “Public Sector Unions and the Costs of Government.” The Journal of Politics 77 (1): 114–27.CrossRefGoogle Scholar
Ba, Bocar A., Knox, Dean, Mummolo, Jonathan, and Rivera, Roman. 2021. “The Role of Officer Race and Gender in Police-Civilian Interactions in Chicago.” Science 371 (6530): 696702.CrossRefGoogle ScholarPubMed
Banks, Duren, Ruddle, Paul, Kennedy, Erin, and Planty, Michael G.. 2016. Arrest-Related Deaths Program Redesign Study, 2015-16: Preliminary Findings. Washington, DC: Bureau of Justice Statistics.Google Scholar
Banks, Jeffrey S., and Weingast, Barry R.. 1992. “The Political Control of Bureaucracies under Asymmetric Information.” American Journal of Political Science 36 (2): 509–24.CrossRefGoogle Scholar
Barber, Catherine, Azrael, Deborah, Cohen, Amy, Miller, Matthew, Thymes, Deonza, Wang, David Enze, and Hemenway, David. 2016. “Homicides by Police: Comparing Counts from the National Violent Death Reporting System, Vital Statistics, and Supplementary Homicide Reports.” American Journal of Public Health 106 (5): 922–27.CrossRefGoogle ScholarPubMed
Boehmke, Frederick J., and Shipan, Charles R.. 2015. “Oversight Capabilities in the States: Are Professionalized Legislatures Better at Getting What They Want?State Politics & Policy Quarterly 15 (3): 366–86.CrossRefGoogle Scholar
Bonica, Adam. 2016. Database on Ideology, Money in Politics, and Elections [computer file]. Public Version 2.0. Social Science Data Collection, Stanford University [distributor].Google Scholar
Bowen, Daniel C., and Greene, Zachary. 2014. “Should We Measure Professionalism with an Index? A Note on Theory and Practice in State Legislative Professionalism Research.” State Politics & Policy Quarterly 14 (3): 277–96.CrossRefGoogle Scholar
Campeau, Holly. 2015. “‘Police Culture’ at Work: Making Sense of Police Oversight.” British Journal of Criminology 55 (4): 669–87.CrossRefGoogle Scholar
Chaney, Carole Kennedy, and Saltzstein, Grace Hall. 1998. “Democratic Control and Bureaucratic Responsiveness: The Police and Domestic Violence.” American Journal of Political Science 42 (3): 745–68.CrossRefGoogle Scholar
Choi, Heasun, and Hong, Sounman. 2021. “Linking Political and Bureaucratic Representation: Does Minority Representation on City Councils Affect the Outcomes of Street-Level Service?Public Administration 99 (2): 405–21.CrossRefGoogle Scholar
Christie, Gayre, Petrie, Simon, and Timmins, Perri. 1996. “The Effect of Police Education, Training and Socialisation on Conservative Attitudes.” Australian & New Zealand Journal of Criminology 29 (3): 299314.CrossRefGoogle Scholar
Cinelli, Carlos, and Hazlett, Chad. 2020. “Making Sense of Sensitivity: Extending Omitted Variable Bias.” Journal of the Royal Statistical Society: Series B (Statistical Methodology) 82 (1): 3967.CrossRefGoogle Scholar
Coe, Charles K., and Wiesel, Deborah Lamm. 2001. “Police Budgeting: Winning Strategies.” Public Administration Review 61 (6): 718–27.CrossRefGoogle Scholar
Cook, Scott J., and Weidmann, Nils B.. 2019. “Lost in Aggregation: Improving Event Analysis with Report-Level Data.” American Journal of Political Science 63 (1): 250–64.CrossRefGoogle Scholar
Cook, Scott J., and Fortunato, David. 2022. “Replication Data for: The Politics of Police Data: State Legislative Capacity and the Transparency of State and Substate Agencies.” Harvard Dataverse. Dataset. https://doi.org/10.7910/DVN/MFFV3C.CrossRefGoogle Scholar
Crank, John P. 2014. Understanding Police Culture. New York: Routledge.CrossRefGoogle Scholar
Criminal Justice Information Services. 2019. “Crime in South Carolina 2019.South Carolina State Law Enforcement Division.Google Scholar
Davenport, Christian, and Ball, Patrick. 2002. “Views to a Kill: Exploring the Implications of Source Selection in the Case of Guatemalan State Terror, 1977-1995.” Journal of Conflict Resolution 46 (3): 427–50.CrossRefGoogle Scholar
DeHart, Cameron. 2020. “The Rise (and Fall) of Elected Sheriffs.” APSA Preprints. doi: 10.33774/apsa-2020-gr1cm.CrossRefGoogle Scholar
Desmond, Matthew, Papachristos, Andrew V., and Kirk, David S.. 2016. “Police Violence and Citizen Crime Reporting in the Black Community.” American Sociological Review 81 (5): 857–76.CrossRefGoogle Scholar
Duggan, Joe. 2018. “Nebraska Lawmakers Ready to Subpoena State Prisons Director to Testify about Death Penalty Protocol.” Omaha World-Herald, April 25. https://www.omaha.com/news/state_and_regional/nebraska-lawmakers-ready-to-subpoena-state-prisons-director-to-testify-about-death-penalty-protocol/article_052dc6eb-f5ad-5114-9e88-4cf58ed6e1be.html.Google Scholar
Eckhouse, Laurel. 2021. “Metrics Management and Bureaucratic Accountability: Evidence from Policing.” American Journal of Political Science 66 (2): 385401.CrossRefGoogle Scholar
Egel, Benjy, Chabria, Anita, and Garrison, Ellen. 2017. “Hands Removed, Findings Changed: Pathologists Say San Joaquin Sheriff 'Does Whatever He Feels Like Doing.’” Sacramento Bee, December 5. https://www.sacbee.com/news/local/article188145129.html.Google Scholar
Falcone, David N., and Wells, L. Edward. 1995. “The County Sheriff as a Distinctive Policing Modality.” American Journal of Police 14 (3/4): 123–49.CrossRefGoogle Scholar
Farris, Emily M., and Holman, Mirya R.. 2017. “All Politics Is Local? County Sheriffs and Localized Policies of Immigration Enforcement.” Political Research Quarterly 70 (1): 142–54.CrossRefGoogle Scholar
FBI. 2020. “A Word about UCR Data.” Federal Bureau of Investigation: Uniform Crime Reporting Program. https://ucr.fbi.gov/a-word-about-ucr-data.Google Scholar
Fortunato, David, and Turner, Ian R.. 2018. “Legislative Capacity and Credit Risk.” American Journal of Political Science 62 (3): 623–36.CrossRefGoogle Scholar
Friedman, Lisa. 2019. “E.P.A. Plans to Get Thousands of Pollution Deaths off the Books by Changing Its Math.” New York Times, May 20. https://www.nytimes.com/2019/05/20/climate/epa-air-pollution-deaths.html.Google Scholar
Gerber, Monica M., and Jackson, Jonathan. 2017. “Justifying Violence: Legitimacy, Ideology and Public Support for Police Use of Force.” Psychology, Crime & Law 23 (1): 7995.CrossRefGoogle Scholar
Goldsmith, Andrew John. 1991. Complaints against the Police: The Trend to External Review. Oxford: Clarendon Press.Google Scholar
Goldstein, Herman. 1963. “Police Discretion: The Ideal versus the Real.” Public Administration Review 23 (3): 140148.CrossRefGoogle Scholar
Hafner-Burton, Emilie M. 2008. “Sticks and Stones: Naming and Shaming the Human Rights Enforcement Problem.” International Organization 62 (4): 689716.CrossRefGoogle Scholar
Hawdon, James E., Ryan, John, and Griffin, Sean P.. 2003. “Policing Tactics and Perceptions of Police Legitimacy.” Police Quarterly 6 (4): 469–91.CrossRefGoogle Scholar
Hollyer, James R., Rosendorff, B. Peter, and Vreeland, James Raymond. 2011. “Democracy and Transparency.” The Journal of Politics 73 (4): 11911205.CrossRefGoogle Scholar
Holmes, Malcolm D., Painter, Matthew A., and Smith, Brad W.. 2019. “Race, Place, and Police-Caused Homicide in U.S. Municipalities.” Justice Quarterly 36 (5): 751–86.CrossRefGoogle Scholar
Huber, John D., and Shipan, Charles R.. 2002. Deliberate Discretion: The Institutional Foundations of Bureaucratic Autonomy. Cambridge: Cambridge University Press.CrossRefGoogle Scholar
Jacob, Brian A., and Levitt, Steven D.. 2003. “Rotten Apples: An Investigation of the Prevalence and Predictors of Teacher Cheating.” The Quarterly Journal of Economics 118 (3): 843–77.CrossRefGoogle Scholar
Jacobs, David, and O’Brien, Robert. 1998. “The Determinants of Deadly Force: A Structural Analysis of Police Violence.” American Journal of Sociology 103 (4): 837–62.CrossRefGoogle Scholar
Joint Legislative Audit and Review Commission. 2008. “Report to the Governor and the General Assembly of Virginia: Key Considerations for Marijuana Legalization.Commonwealth of Virginia.Google Scholar
Klinger, David A. 2012. “On the Problems and Promise of Research on Lethal Police Violence: A Research Note.” Homicide Studies 16 (1): 7896.CrossRefGoogle Scholar
Knox, Dean, Lowe, Will, and Mummolo, Jonathan. 2020. “Administrative Records Mask Racially Biased Policing.” American Political Science Review 114 (3): 619–37.CrossRefGoogle Scholar
Koh, Elizabeth. 2018. “Florida House Issues Subpoenas in Probe of Parkland Shooting.” Tampa Bay Times, February 28. https://www.tampabay.com/florida-politics/buzz/2018/02/28/florida-house-issues-subpoenas-in-probe-of-parkland-shooting/.Google Scholar
Kroeger, Mary A. 2022. “Bureaucrats as Lawmakers.” Legislative Studies Quarterly 47 (1): 257–89.CrossRefGoogle Scholar
LaFrance, T. Casey. 2012. “The County Sheriff’s Leadership and Management Decisions in the Local Budget Process Revisited.International Journal of Police Science & Management 14 (2): 154–65.Google Scholar
Landgrave, Michelangelo, and Weller, Nicholas. 2020. “Do More Professionalized Legislatures Discriminate Less? The Role of Staffers in Constituency Service.” American Politics Research 48 (5): 571–78.CrossRefGoogle Scholar
Lax, Jeffrey R., and Phillips, Justin H.. 2012. “The Democratic Deficit in the States.” American Journal of Political Science 56 (1): 148–66.CrossRefGoogle Scholar
Levitt, Steven D. 1997. “Using Electoral Cycles in Police Hiring to Estimate the Effects of Police on Crime.” American Economic Review 87 (3): 270–90.Google Scholar
Lillvis, Denise F., and Mcgrath, Robert J.. 2017. “Directing Discipline: State Medical Board Responsiveness to State Legislatures.” Journal of Health Politics, Policy and Law 42 (1): 123–65.CrossRefGoogle ScholarPubMed
Lipsky, Michael. 1980. Street-Level Bureaucracy: Dilemmas of the Individual in Public Service. New York: Russell Sage Foundation.Google Scholar
Loftis, Matt W., and Kettler, Jaclyn J.. 2015. “Lobbying from inside the System: Why Local Governments Pay for Representation in the U.S. Congress.” Political Research Quarterly 68 (1): 193206.CrossRefGoogle Scholar
Malone, Clare, and Asher, Jeff. 2017. “The First FBI Crime Report Issued under Trump Is Missing a Ton of Info.” Five Thrityeight, October 27. https://fivethirtyeight.com/features/the-first-fbi-crime-report-issued-under-trump-is-missing-a-ton-of-info/.Google Scholar
Maltz, Michael D., and Targonski, Joseph. 2002. “A Note on the Use of County-Level UCR Data.” Journal of Quantitative Criminology 18 (3): 297318.CrossRefGoogle Scholar
McCubbins, Mathew D., and Schwartz, Thomas. 1984. “Congressional Oversight Overlooked: Police Patrols versus Fire Alarms.” American Journal of Political Science 28 (1): 165–79.CrossRefGoogle Scholar
Mughan, Siân, Li, Danyao, and Nicholson-Crotty, Sean. 2020. “When Law Enforcement Pays: Costs and Benefits for Elected versus Appointed Administrators Engaged in Asset Forfeiture.” The American Review of Public Administration 50 (3): 297314.CrossRefGoogle Scholar
Mummolo, Jonathan, and Peterson, Erik. 2018. “Improving the Interpretation of Fixed Effects Regression Results.” Political Science Research and Methods 6 (4): 829–35.CrossRefGoogle Scholar
Payson, Julia A. 2020. “Cities in the Statehouse: How Local Governments Use Lobbyists to Secure State Funding.” The Journal of Politics 82 (2): 403–17.CrossRefGoogle Scholar
Peterson, Erik. 2021. “Paper Cuts: How Reporting Resources Affect Political News Coverage.” American Journal of Political Science 65 (2): 443–59.CrossRefGoogle Scholar
Prenzler, Tim, and Ronken, Carol. 2001. “Models of Police Oversight: A Critique.” Policing and Society: An International Journal 11 (2): 151–80.CrossRefGoogle Scholar
Quinn, Kevin M. 2004. “Bayesian Factor Analysis for Mixed Ordinal and Continuous Responses.” Political Analysis 12 (4): 338–53.CrossRefGoogle Scholar
Rocheleau, Matt. 2019. “How a Trooper’s Alleged Racist Remark Ignited the State Police Overtime Scandal.” Boston Globe, August 17. https://www.bostonglobe.com/metro/2019/08/17/how-one-trooper-alleged-racist-remark-ignited-state-police-overtime-fraud-scandal/xrzydzqhfrfa9rtihwpdhp/story.html.Google Scholar
Romzek, Barbara S., and Dubnick, Melvin J.. 1987. “Accountability in the Public Sector: Lessons from the Challenger Tragedy.” Public Administration Review 47 (3): 227238.CrossRefGoogle Scholar
Rubin, Irene S. 2019. The Politics of Public Budgeting: Getting and Spending, Borrowing and Balancing. Washington, DC: CQ Press.Google Scholar
Soss, Joe, and Weaver, Vesla. 2017. “Police Are Our Government: Politics, Political Science, and the Policing of Race–Class Subjugated Communities.” Annual Review of Political Science 20: 565–91.CrossRefGoogle Scholar
Squire, Peverill. 2007. “Measuring State Legislative Professionalism: The Squire Index Revisited.” State Politics & Policy Quarterly 7 (2): 211–27.CrossRefGoogle Scholar
Squire, Peverill, and Hamm, Keith E.. 2005. 101 Chambers: Congress, State Legislatures, and the Future of Legislative Studies. Athens: Ohio State University Press.Google Scholar
Stolberg, Sheryl Gay. 2020. “Trump Administration Strips C.D.C. of Control of Coronavirus Data.” New York Times, July 14. https://www.nytimes.com/2020/07/14/us/politics/trump-cdc-coronavirus.html.Google Scholar
Sumner, Jane Lawrence, Farris, Emily M., and Holman, Mirya R.. 2020. “Crowdsourcing Reliable Local Data.” Political Analysis 28 (2): 244–62.CrossRefGoogle Scholar
Sunshine, Jason, and Tyler, Tom R.. 2003. “The Role of Procedural Justice and Legitimacy in Shaping Public Support for Policing.” Law & Society Review 37 (3): 513–48.CrossRefGoogle Scholar
Tax Policy Center. 2017. Tax Policy Center Briefing Book. Washington, DC: Urban-Brookings Tax Policy Center. https://www.taxpolicycenter.org/briefing-book.Google Scholar
Thompson, James D. [1967] 2017. Organizations in Action: Social Science Bases of Administrative Theory. New York: Routledge.CrossRefGoogle Scholar
Todak, Natalie, and James, Lois. 2018. “A Systematic Social Observation Study of Police De-Escalation Tactics.” Police Quarterly 21 (4): 509–43.CrossRefGoogle Scholar
Vakilifathi, Mona. 2019. “Constraining Bureaucrats Today Knowing You’ll Be Gone Tomorrow: The Effect of Legislative Term Limits on Statutory Discretion.” Policy Studies Journal 47 (4): 9781001.CrossRefGoogle Scholar
Vannoni, Matia, Ash, Elliott, and Morelli, Massimo. 2021. “Measuring Discretion and Delegation in Legislative Texts: Methods and Application to U.S. States.” Political Analysis 29 (1): 4357.CrossRefGoogle Scholar
Weisburd, David, and Braga, Anthony A.. 2019. Police Innovation: Contrasting Perspectives. Cambridge: Cambridge University Press.CrossRefGoogle Scholar
White, Ariel. 2019. “Family Matters? Voting Behavior in Households with Criminal Justice Contact.” American Political Science Review 113 (2): 607–13.CrossRefGoogle Scholar
Williams, Ida J. 2020. “Crime in Maryland: 2018 Uniform Crime Report.” Baltimore: Maryland State Police, Central Records Division.Google Scholar
Yeung, Bernice, Greenblatt, Mark, Fahey, Mark, and Harris, Emily. 2018. “When It Comes to Rape, Just Because a Case Is Cleared Doesn’t Mean It’s Solved.” Pro Publica. November 15. https://www.propublica.org/article/when-it-comes-to-rape-just-because-a-case-is-cleared-does-not-mean-solved.Google Scholar
Zimring, Franklin E. 2017. When Police Kill. Cambridge, MA: Harvard University Press.CrossRefGoogle Scholar
Figure 0

Table 1. Legislative Capacity on UCR Participation

Figure 1

Figure 1. Police Killings by Source (2013–2016)

Figure 2

Figure 2. Comparing 2015 Police Killings per 100,000 Residents across the United States

Figure 3

Table 2. Police Killings Unreported to UCR

Figure 4

Table 3. Difference-in-Differences Analysis of State Investigation Laws

Supplementary material: Link

Cook and Fortunato Dataset

Link
Supplementary material: PDF

Cook and Fortunato supplementary material

Cook and Fortunato supplementary material

Download Cook and Fortunato supplementary material(PDF)
PDF 314.9 KB