The Future of AI Is in the States: The Case of Autonomous Vehicle Policies

The myriad applications of artificial intelligence (AI) by the private and public sectors have exploded in the public consciousness in the postpandemic period. However, researchers and businesses have been working on AI technology applications for decades, and in many ways, governments are rushing to catch up. This article presents an argument that the future of AI policy in the United States will be driven in large part by current and future state-level policy experiments. This argument is presented by drawing on scholarship surrounding federalism, regulatory fragmentation, and the effects of fragmentation on business and social equity. The article then presents the case of autonomous vehicle policy in the states to illustrate the degree of current fragmentation and considers the effects of layering new AI policies on top of existing rules surrounding privacy, licensing, and more. Following this consideration of existing research and its application of AI policy, the article presents a research agenda for leveraging state differences to study the effects of AI policy and develop a cohesive framework for governing AI.

wrestle with how to effectively regulate rapidly emerging and disruptive technologies, and their potential social side effects, without squashing innovation.
Within the context of the United States, we argue that the future of this struggle is very much in the hands of the states.American federalism has been described as an environment in which the states act as laboratories.Driven by competition and problem-solving, states learn from each other, copy each other, and can push their ideas up to the federal government. 6Often, however, this situation results in significant regulatory fragmentation for businesses.And growing political polarization and gridlock in the national government means that even significant pressure from the states for federal action runs into strong federal stasis. 7While the National Highway Traffic Safety Administration (NHTSA) has issued guidance and technical assistance, 8 states still have the power to go their own way on the myriad policy choices that need to be made when enabling and regulating AI applications.Fragmentation of regulatory authority across agencies and levels of government presents significant costs for industries. 9n the other hand, it also offers firms leverage in extracting favorable policy arrangements from states that are competing for their business. 10Technology companies have shifted from positions of opposing government regulation to supporting regulation that strikes a balance between laissez-faire and prohibitive approaches. 11But they have also lobbied states for industry-friendly policies, like protections against customer suits for privacy violations. 12t is within this environment that the future of American AI policy will be worked out.Given the prominence of the United States in the global economy, and in AI patenting, 13 policy lessons learned in the states could be transferred cross-nationally.Of course, with the European Union (EU) emerging as a distinctive leader in AI governance, policy transfer may very well occur in the other direction, at least until the United States catches up.The EU's AI Act, while not yet finally adopted, is being touted as the first regulatory framework for AI and a model for countries across the world.The United Nations, the Organisation for Economic Co-operation and Development, the G7, and the G20 have all either established or debated principles of AI governance. 14ithin the United States, the states have their own economies, cultures, and political contexts that will shape their approaches to regulating and promoting AI technology.We argue that the policy innovation and diffusion that is already occurring among the states, and will continue to occur, has significant implications for businesses engaged in AI development.Our aim in this article is to make this broader argument, but to flesh it out with a case study of state autonomous vehicle policy.We lay the broader argument with reviews of three distinct yet related literatures.We begin more broadly with the nature of American federalism, policy fragmentation, and innovation.We then examine research on the effects of this fragmentation on businesses and on social equity.Having established this theoretical foundation for understanding AI fragmentation, we present the case of state autonomous vehicle policies.This AI application has seen substantial and varied state activity over the last decade.Thus, it is a useful example of how states are moving in many different directions with respect to AI business regulations and social equity concerns.We then consider how new AI policies are being built on top of existing policy regimes and the implications of policy layering for effective implementation.Finally, we bring the existing literature and the specific challenges of AI technology together to present a research agenda for studying AI policy fragmentation, how it affects business activity, and how scholars can leverage state variation in establishing and evaluating an effective AI governance framework in the United States.

Federalism and regulatory fragmentation in the states
The horizontal and vertical relationships between governments in the American federal system have evolved substantially since the founding.A degree of rigid clarity regarding the respective powers and policy monopolies of the federal government and the states has given way to waves of competitive, cooperative, fragmented, and now increasingly polarized relationships.Policy innovation and diffusion has long been regarded, however, as a significant benefit of the decentralization of policymaking authority in the United States. 15States can take up new policy ideas, try them out, learn from each other, and address failures in their initial experiments. 16In theory, this allows the states to "kick the tires," so to speak, before policy is nationalized by the federal government.
The diffusion of innovations framework is used by policy process scholars to understand why some governments adopt policy innovations and others do not. 17Problem-solving through policy learning is only one driver of the diffusion of policy innovations. 18States are also competitive with each other.They compete using available policy tools to attract benefits, like population, economic development, and tax revenue, as well as to deter unwanted populations. 19Granted, both individual decisions to move and state policy choices in a competitive environment are far more complicated than those envisioned by Tiebout competition. 20However, state legislators rhetorically point to competitor states, particularly contiguous neighbors, when pushing policy innovations.This rhetoric often comes from a defensive posture recognizing that some benefit is lost by population and business movement to a state with a more advantageous policy environment. 21tates as policy laboratories are affected not only by bottom-up identification of problems and horizontal patterns of policy learning and interstate competition.The federal government has used competitive programs like Race to the Top in education policy to foster competition, innovation, and diffusion among the states. 22Additionally, the federal government can incentivize the top-down diffusion of policy innovations through coercive sticks and carrots, as well as through its own issue attention. 23It has also formally devolved policymaking responsibility in a variety of areas to the states. 24ver the last two decades, states have been pushed to act in a variety of policy domains-even on clear federal government issues like immigration-because of gridlock in national political institutions resulting from growing political polarization. 25While this is not a formal devolution of power, states are filling the void left by the reduction of substantive federal policymaking activity. 26This is no less the case for AI policy, as the federal government has established principles for its own use of AI and departments have developed recommended rules (e.g., NHTSA), but there is no uniform policy regime nationally.This means that states can serve as laboratories for exploring the best methods for using and regulating AI.
While this dynamic environment allows for substantial innovation at the state and local levels, it comes with costs.Policy innovation and diffusion can be slow and incomplete when left to the states. 27iven the increasing friction within the federal policymaking system, it is not a given that even successful policy experiments will be nationalized. 28There are also significant costs to businesses that must navigate different rules because of regulatory fragmentation.We turn now to considering these costs and how businesses adapt.

Fragmented regulations and ventures' market and nonmarket strategies
When newly introduced controversial technologies emerge, such as AI, regulators tend to be uncertain about their potential benefits or risks.As a result, regulatory experiments are launched at various levels of government: city, county, state, and federal.Studies have demonstrated that fragmented regulations can increase the costs associated with commercializing technologies in two ways.First, they increase the cognitive burden placed on ventures, as firms must become familiar with all the fragmented regulations prior to starting their business.In fact, reducing regulatory variations helps increase the number of entrepreneurs. 29Second, fragmented regulations increase the costs of compliance for firms that wish to do business in multiple areas, as they must take additional measures to meet the requirements of different regulations. 30ragmented regulations can have a considerable impact on the ecosystem surrounding new technologies, as investors, media, consumers, and suppliers may be reluctant to support or legitimize entrepreneurs when regulations are uncertain. 31Faced with fragmented regulations, ventures have different responses.Ventures can decrease the scope of their innovations.Or they will try to pivot and adjust their businesses according to the different regulatory schemes.Such fragmented regulations can be a significant drawback for investors.For example, when the United Kingdom tried to follow a different set of rules from the EU on Sustainable Finance Disclosure Regulation, investors from JPMorgan Asset Management voiced their concern that the different regulations in European markets would make it difficult for ventures to develop. 32Similarly, as regulations on autonomous vehicles are fragmented across states in the United States, investors are concerned about the scaling up of newly established ventures. 33Yet regional variations of regulations also provide more innovation grounds for ventures, which may attract more equity financing.
Despite the potential downsides of fragmented regulations, ventures can also utilize nonmarket strategies to help coordinate these regulations to make them converge rather than diverge in the system.First, entrepreneurs that lobby or collaborate with regulators can push for preemption from higher levels of government to reduce conflict among local regulations. 34In a converging regulatory structure, legally authorized institutional settings, regulatory scope, legal instruments, and procedural requirements converge into a consistent system. 35Such a system can be gradually institutionalized as the regulation templates for new technologies or business models.For example, the Department of Transportation Act of 1966 created the Federal Railroad Administration, which helped consolidate railway regulations across different states to foster railroad development.Second, entrepreneurs can also pick locations with favorable regulations to develop and then push for other places to change regulations accordingly.This will create a race-to-the-bottom/top effect. 36In fact, entrepreneurial businesses, as opposed to mature firms, view regulations and policy more favorably as it provides some control over the business environment. 37The fragmentation of regulations in the United States not only creates challenges for firms, it also creates inequalities for citizens.

Fragmentation effects on equity
Utilization of AI and the fragmentation of regulatory policies present several opportunities and challenges for public administrators.Automating decision-making poses opportunities for the public sector and can increase efficiency and responsiveness while also raising concerns over equity and the efficacy of democracy. 38In such instances, power is shared between bureaucrats, technocrats, and contractors with certain expertise during the creation of AI metrics.This, however, removes humanity from government transactions and limits adaptability to changes in contexts, social desirability, or the national agenda. 39Further, the implementation of AI policies across states leads to increased inequities as states do not always implement policies equally, even when the same laws are in place. 40This section discusses equity aspects in the creation of AI metrics and expands on fragmentation during implementation of AI policies that lead to inequity.

Equity in the creation of AI metrics
There are several justice and equity considerations in the creation of AI metrics.These include the digital divide (or unequal access to technology); algorithmic bias and values; differences in systems based on cultural and societal norms and values; fair decision-making mechanisms; and diversity, equity, and inclusion (DEI) practices. 41Additional questions remain over the issue of responsibility in terms of machine choices.Automating weapons and vehicles, for instance, may seem desirable or inevitable but creates critical questions on policy and ethics. 42These vary in relevance depending on the domains under consideration-such as in public health, social welfare, policing and criminal justice, military operations, and transportation, to name a few.Discrimination and equity concerns can play out differently in various contexts and when various social groups are involved. 43or example, the military may use AI for efficiency in war response, or more specifically in geospatial intelligence to leverage algorithms for use in identifying targets in drone strikes.AI is becoming paramount for battlefield efficiency, strategy, and tactics.The war in Ukraine is a timely example, using geospatial intelligence to consider satellite images, locate photos, and other information from a variety of sources and governments.This war is unique in the willingness of intelligence entities around the world to assist Ukraine, providing AI software for analyzing the war and its movements. 44However, if a mistake or war crime were to happen because of an algorithm that led to catastrophic consequences, difficult questions would arise over how AI presupposed human decisions and where responsibility should fall. 45Additionally, AI is used for loitering and striking targets on current battlefields, a step before autonomous targeting.While the Pentagon's Joint Artificial Intelligence Center works to avoid AI accidents, particularly in crisis dynamics, escalation, and strategic stability around AI military use, immense justice and equity considerations remain for policymakers in terms of equitable uses of AI, responsibility for decision-making and ethical outcomes. 46ossible remedies for AI equity considerations include external audits, algorithmic judgments across social categories, legal impacts to algorithmic systems, and considerations of certain groups for examining contexts, impacts, and gaining insight, in addition to considerations of DEI practices and management. 47Policies are also needed to ensure that AI will serve the public good and that there are frameworks for responsibility surrounding AI that are satisfactory to stakeholders. 48AI policy can be defined as policy that is "aimed at research, development, and deployment of AI technologies in the economy and society" 49 and involves multiple stakeholders and multistakeholder arrangements, including governments, industries, markets, society, and other organizations. 50AI policymaking is complex, as decisions are multilevel and often situated within policy issues, solutions, and instruments and involve various policy instruments, governance modes, and different social, political, and economic contexts. 51AI policy creation is impacted by design, dynamics and interactions of political regimes, and governance models, and it is impacted by national strategy and laws in place.

Fragmentation and implementation of policies leading to inequity
The United States and many other countries have national strategies for AI that include multistakeholder arrangements, such as vertical and horizontal governmental relationships, and generally aim to improve security and quality of life. 52Composed of six general objectives, the US policy supports international cooperation and soft power mechanisms, and it follows OECD principles on AI. 53 State and local governments are increasingly becoming interested in AI policies and working to fill gaps of limited and general federal guidance. 54State-level policies regarding AI may offer more creativity, flexibility, or agility than policies and strategies provided at the federal level.
Many states have unique approaches that impact citizens differently based on several factors and contexts, from the creation of task forces and special commissions to notices on legal protections and privacy.For instance, California requires organizations to disclose the use of artificially intelligent bots when communicating online with citizens or consumers.The State of Washington is considering requiring governments to use algorithmic accountability reports for automatic decisions. 55Strategies and federal recommendations regarding the adoption of AI is one avenue for understanding AI policy adoption within states.While the literature generally considers the benefits of AI policies, states do not implement policies equally, even when the same laws are adopted. 56Further, an additional equity challenge is often not what information is collected utilizing AI, but how granular information is implemented, which can happen based on context. 57These are larger issues that exist in federalism, which creates inequality between groups and unequal treatment in benefits. 58he devolution of social policy in the United States is instructive to the social equity problems that emerge with policy fragmentation.A tangible example is algorithmic risk assessments in sentencing, where judges and courts use automation to assist human judgment through risk scores and algorithms.While efforts to utilize an AI risk assessment tool were made to reduce incarceration and recidivism in Virginia, for example, results show that utilization of the tool can lead to age and racial disparities and left more questions in terms of issues with implementation, consistency in use, and adequacy of training to utilize such systems. 59Welfare devolution to the states is another example of how inequities have been created as a result of uneven responsiveness of state and local governments. 60Domestic violence policy devolution is another, whereby state and federal factors influence the adoption of policies through political and demographic indicators. 61Because federal domestic violence laws are designed to leave states leeway, cases are seen under state laws, which vary in terms of definition of the crime, the resulting punishment, and the latitude given to officers on how to arrest and to prosecutors on how to classify domestic violence.Research has found clear differences in terms of inequities and responsiveness based on context across the states. 62Similar to domestic violence policies, which have a 50 Raymond and DeNardis (2015).51 Filgueiras (2022).56 Dwivedi et al. (2021); Sidorsky and Schiller (2023).57 Calo (2018).
59 Van Dam (2019).relatively short history, varying support, and inherently unequal policy effects based on many social and demographic indicators, 63 AI must also attract more attention from policymakers and scholars to address the differences across governments at the local, state, and federal level and the way these laws are implemented in order to achieve equity and equality.
Divergent ways of interpreting AI policies horizontally and vertically-meaning across and among government levels and agencies-have happened in other political contexts as well.Af Malmborg and Trondal consider different factors influencing the adoption of AI policies in the EU and the coordination of policies by member states in Nordic countries.They find that there were differences in the adoption of AI policies based on national organizational capacities, thus suggesting that the framing of policies is filtered through organizational structures and governance mechanisms across countries.Because of the different organizational capacities within these countries, there could be weaker implementation of certain policies. 64The implications for national and even transnational roles on policymaking and equity are abundant.
There are other personal and political costs of fragmented AI policies.Any sociotechnical system that utilizes algorithmic processes in decision-making can create discrimination.In fact, Williams, Brooks, and Shmargad find that censoring information like social category data can exacerbate discrimination. 65Regardless, special considerations must be made for equitable considerations of AI policies and fragmentation in states.It is difficult, if not impossible, to ensure equity across the states without federal intervention.However, it is also true that when the federal government acts, it can crowd out desirable state policy innovation as states use their resources to pursue other goals. 66aving considered the broader opportunities and challenges of federalism for innovation, as well as the specific issues of policy fragmentation for business and social equity, we now turn to illustrating this argument with the case of state autonomous vehicle policy.
A case of regulatory fragmentation: Autonomous vehicles Autonomous vehicle (AV) policy fragmentation is a significant challenge presented by the development and deployment of self-driving cars. 67Because of the lack of federal regulations, states have been taking the lead in creating their own policies, resulting in a patchwork of regulations and requirements that vary from state to state.This fragmentation has created a complex regulatory landscape that can be challenging for companies working on autonomous vehicles to navigate. 68AVs are perhaps one of the most visible applications of commercial AI technology.The Insurance Institute for Highway Safety has predicted that 3.5 million self-driving vehicles will be on American roads by 2025, and McKinsey & Company has projected $300 billion to $400 billion in global revenues from AVs by 2035. 69Top producers of personal and commercial AV technology include Tesla, Nvidia, Waymo, and Argo.ai.The industry is much broader than fully autonomous vehicles and includes other driverassistance systems such as cameras, lidar, radar, and sensors. 70o better understand the current state of autonomous vehicle regulation across the United States, we use the National Conference of State Legislatures (NCSL) database on autonomous vehicle laws passed by states from 2010 to 2023 for a descriptive analysis of this fragmentation. 71Doing so offers a clear and concise overview of the current regulatory landscape, helping identify where gaps and inconsistencies exist.Specifically, we illustrate both the business and equity challenges presented by the variation in-and, in many cases, the lack of-state policies.To this end, we focus on five business-63 Ibid.

Business and Politics
related topics that are present in state laws-commercial, insurance and liability, licensing and regulation, operator requirements, and vehicle testing-and one equity concern: privacy of collected vehicle data.Table 1 charts which states have adopted each of these components of AV policy.
There are notable regional patterns in the adoption of these components.Northeastern states have been cautious in their approach to AVs, with more laws focused on safety regulations and testing.Southern states, like Texas and Florida, have been more permissive with testing and deployment.Midwestern states like Michigan and Illinois have led in the areas of research and development.Finally, western states like California and Washington have taken more progressive approaches to regulations that encourage the adoption of AV technology.California is a clear leader in AV policy innovation, which is understandable given that the state has become synonymous with tech innovation.But it does not lead on every component of AV policy, and there are substantial differences in the approaches taken by adopting states.We now consider these differences.

Commercial policy
Autonomous vehicles stand to benefit commercial applications by reducing labor costs and improving efficiency in transportation and logistics operations. 72Commercial laws are policies that regulate business activities within a country.In the United States, each state has its own approach to commercial laws.While among the most widely adopted topics according to the NCSL, the substantial variability in what states have allowed, disallowed, or regulated is evident as we examine the details of these laws.
Regulations supportive of the industry include examples like Alabama's exemption for truck platoons from receiving citations for following too closely. 73The exemption applies to platoons engaged in research for truck platooning technology.Similarly, Arkansas enacted a law that exempts truck platoons under certain conditions.The law defines a driver-assistive truck platooning system as technology that integrates sensor arrays, wireless communications, vehicle controls, and specialized software to coordinate acceleration and braking between two or more vehicles with a human operator in the lead vehicle.
Connecticut offers an example of commercial policy that includes more intrusive regulations.It provides the commissioner of transportation with the right to enter and utilize private property to correct unsafe conditions or restore the interruption of essential railroad or transit services.The commissioner is required to make a reasonable effort to notify the owner of record of such property prior to entering the property.
Other states' commercial provisions have to do with taxation.California enacted specific tax reforms for vehicle testing in San Francisco.Indiana repealed the motor carrier surcharge tax and increased the special fuel tax.It also specified how netted International Fuel Tax Agreement Clearinghouse refunds and receipts are deposited or credited and the method for calculating the commercial vehicle excise tax rate.In sum, each state has a unique approach to commercial laws that reflects the state's economic and political interests.

Insurance and liability policy
Insurance and liability for autonomous vehicles have been significant concerns for lawmakers in multiple states. 74Autonomous vehicle insurance and liability regulations have been enacted in several states, including California, Texas, Florida, Michigan, and New York, with the aim of ensuring that insurance coverage is in place to cover damages and that responsible parties are held liable in case of an accident involving an autonomous vehicle.For example, as early as 2012, California required 72 Brinkley, Daily, and Gilbert (2019)   73 A platoon is defined as a "group of individual commercial trucks traveling in a unified manner at electronically coordinated speeds at following distances that are closer than would be reasonable and prudent without the electronic coordination" (Alabama SB 172, 2018).
Business and Politics manufacturers of autonomous vehicles to obtain a special permit before testing on public roads, and the state requires these vehicles to have liability insurance coverage.States like Texas, Florida, Michigan, and New York also have passed laws establishing minimum insurance and liability coverage requirements for AV testing.
In some states, like Alabama and Arkansas, laws have been more narrowly enacted to address insurance and liability concerns.Alabama established minimum liability insurance coverage requirements for autonomous commercial vehicles operated by an automated driving system and commercial motor vehicles with teleoperation systems.Arkansas established minimum liability insurance coverage requirements for motor carriers of property to ensure that sufficient insurance coverage is in place in case of an accident involving autonomous vehicles.
Other states, like Georgia, Iowa, Utah, Vermont, and West Virginia, have enacted laws that establish broader regulations for autonomous vehicles, including their operation and licensing requirements, while also requiring liability insurance coverage.For example, West Virginia passed the Fully Autonomous Vehicle Act in 2022, which provides requirements for the operation of fully autonomous vehicles without a human driver and with a human driver; provides for licensing, titling, and registration; and provides for the operation of on-demand autonomous vehicle networks and fully autonomous commercial and motor vehicle carriers, while also requiring liability insurance coverage.

Licensing and registration policy
Every state has a procedure for licensing and registering traditional motor vehicles, which serves several important purposes. 75First, it ensures that states possess accurate records about vehicles and their owners. 76Second, it guarantees that vehicles on the road are insured, thereby promoting financial safety for all. 77Third, this process generates revenue for the state, which can be used to support various public services and initiatives. 78Several states have enacted legislation to specifically license and register autonomous vehicles. 79 Wyoming 75 Hubbard (2018).
In New York and North Carolina, for instance, the state's department of transportation is authorized to conduct testing of autonomous vehicle technology, but operators must hold a valid driver's license for the type of vehicle being used.Georgia exempts persons operating an automated motor vehicle with the automated driving system engaged from the requirement to hold a driver's license and provides for registration requirements.Utah defines terms related to autonomous vehicles, allows the operation of a vehicle in the state by an automated driving system, and exempts a vehicle with an engaged automated driving system from licensure.Colorado allows the use of automated driving systems to control motor vehicles in compliance with state and federal laws.It also convened a stakeholder group to make recommendations for further regulations regarding autonomous commercial vehicles.

Operator requirement policy
State operator requirements regulate the use of autonomous and semiautonomous vehicles and establish the conditions and requirements for their operation.Each state has approached their operator requirements differently.Some states, like Alabama and Arkansas, have enacted laws that authorize autonomous commercial vehicles operated by an automated driving system and commercial motor vehicles with teleoperation systems.Other states, like Arizona, have enacted laws that relate specifically to autonomous vehicles, without operator requirements.
States like California have enacted laws that require operators of autonomous vehicles to comply with certain regulations, such as restrictions on the use of wireless communication devices.Autonomous vehicle operators are required to comply with these regulations to ensure the safety of the vehicle's passengers and other road users.Conversely, Florida has made a notable exemption for motor vehicle operators who are operating autonomous vehicles by allowing them to use wireless communication devices, which is otherwise prohibited while driving.This decision recognizes the fact that the operator of an autonomous vehicle may have less need for attention and focus on the road than traditional drivers.
Not yet decided on its regulations, Kansas has established an autonomous vehicle advisory committee to provide guidance on autonomous vehicle policy and development.This committee includes representatives from the transportation industry, academia, and government agencies.Louisiana has taken a more comprehensive approach to regulating autonomous vehicles.It has established a controlling authority for autonomous commercial motor vehicles, which outlines specific requirements for operators, establishes reporting requirements following an accident, and provides guidance on the use of remote drivers and teleoperation systems.These regulations are intended to promote safety and ensure that operators of autonomous vehicles are appropriately trained and qualified.

Vehicle testing policy
Eigheen states have passed legislation authorizing the testing of autonomous vehicles on public roads.The legislation varies from state to state but generally allows autonomous vehicles to be tested under certain conditions, such as having a human driver ready to take control if necessary.States like Arkansas, California, and Colorado have authorized testing by private companies on states roads or highways.However, some states, like Colorado, have stipulations that vehicles being tested must be equipped with technology that allows the human driver to take control at any time.
Several states have enacted their own testing programs.Connecticut and New Hampshire both adopted commissions or tasks forces to study and test AVs.New York has passed several testing laws.The first authorized an autonomous vehicle demonstration project to test autonomous vehicles that do not have a driver in the driver's seat and are not equipped with a steering wheel, brake pedal, or accelerator.The second extended the repeal date of provisions authorizing the department of transportation to conduct testing of technologies that enable drivers to safely operate motor vehicles with less than 100 feet between each vehicle or combination of vehicles.The third repealed a requirement that the department of motor vehicles notify the legislature of receipt of an application seeking approval to operate an autonomous vehicle capable of operating without a driver inside the vehicle on public roads.These laws demonstrate that several states are actively exploring and implementing policies and regulations to enable the safe and effective deployment of autonomous vehicles on their roads.

Vehicle data privacy policy
While each of the five preceding topics relates to the business of autonomous vehicles, the final topic, privacy, captures a facet of equity.As with other data-rich emergent technologies, like smart homes and devices, autonomous vehicles produce vast geo-location data that raise significant privacy concerns. 80Alas, only five states have adopted laws specifically addressing the privacy of collected vehicle data.One potential explanation for the small number of states adopting privacy policy is that the policy is specific to a particular state's needs and circumstances, making it less applicable or necessary in other states. 81For example, some states may prioritize different policy goals related to autonomous vehicles, such as safety regulations or liability laws, over privacy and data protection regulations.Another reason could be the political or cultural differences between states, which could make it more difficult to achieve consensus on certain policies across state lines. 82Policies that require extensive time and resources to adopt and implement are less likely to be replicated by other states, as they may find it difficult to commit the same level of resources and time. 83For example, it may be a challenge for other states to adopt privacy and protection regulations related to autonomous vehicles because the policies may require a significant investment of time, resources, and expertise to develop and implement.Additionally, existing privacy laws can be complex and vary from state to state, which could create confusion and inconsistencies if other states attempt to replicate policies that have been successful in other regions.
California, Georgia, Michigan, Nevada, and Pennsylvania are the five states that have specifically addressed privacy protection for autonomous vehicles.California has regulations in place requiring autonomous vehicle manufacturers to obtain written consent from passengers before collecting or sharing their personal information.Michigan and Nevada have also enacted laws related to the privacy of collected vehicle data, although they do not specifically mention autonomous vehicles.Policies adopted by Georgia and Pennsylvania both address the privacy of collected vehicle data.Overall, while some states have enacted laws related to the use of autonomous vehicles, there is no federal legislation in place that specifically addresses this issue, which may lead to inconsistencies in how different states regulate the use of autonomous vehicles.However, the trend toward regulating the use of autonomous vehicles and protecting privacy in relation to them is likely to continue in the coming years.
These five policy topics clearly illustrate how the regulation of autonomous vehicles varies substantially among states, including requirements for drivers, vehicles, operations, and the reporting of data. 84he diversity of these requirements poses significant challenges for manufacturers now, as well as for fleet operators, human operators, and consumers of AVs in the future. 85Therefore, navigating the complex regulatory landscape is crucial for successfully implementing AV technology. 86On one hand, experimenting with these different rules offers the opportunity for policy lesson drawing and intergovernmental knowledge and policy transfer, though it is not guaranteed. 87On the other hand, a more consistent and streamlined approach to regulating autonomous vehicles would support the growth and development of this new technology. 88ragmentation in new policies being adopted is not the only regulatory concern for autonomous vehicle manufacturers and eventual users.These new policies are layered on top of existing laws that regulate driving, commerce, and privacy.We now consider how administrative layering creates additional political and business challenges. 89e consequences of layering AI policy Often, the regulations discussed regarding autonomous vehicles are about relatively new policies and specifically targeted rules.However, when hardware, software, and communication systems are all considered, there are numerous regulatory issues involving policies that long predate autonomous vehicles or might not be relevant to vehicles at first glance, creating the possibility of burdensome administrative layering that can negatively affect policy implementation. 90These issues include limitations on active sensor use, radio spectrum and power regulations, minimum safety and inspection standards, and required or disallowed equipment.
These issues also exist internationally, especially with differing spectrum standards and software control.For instance, a subsidiary of Baidu, a Chinese company, is testing autonomous vehicles in California, and current proposed regulation about vital technology systems developed outside the United States is likely to have unanticipated consequences. 91Recently, the US Congress considered banning, limiting, or requiring ownership changes for the social media platform TikTok at a national level because of issues with data security and sensitive information being sent to a Chinese firm.Multiple state and local governments have already proposed banning the platform, with Montana adopting the first statewide ban in 2023. 92There are likely to be similar issues with regulations requiring domestic ownership of critical software tools for vehicle brands that are owned or made outside the United States.If these policies are enacted at the state level, it is possible to imagine a future in which vehicles could be banned from operation after purchase because of software and data collection features.As autonomous vehicles age, they are likely to follow the same migration patterns from the nations currently leading autonomous vehicle development to the Global South, and these powerful technological tools may end up in areas with no effective vehicle regulation or autonomous mapping or infrastructure. 93eturning to the domestic conversation, consider current vehicular hardware laws, such as models allowed in the United States that are in violation of more stringent emissions standards in states like California. 94For both commercial and consumer vehicles, there are models that can be registered in some parts of the United States but not others.This precedent violates the idea that once registered, a vehicle can be used throughout the United States.With autonomous vehicles, these hardware issues have the potential to get even more complicated, as hardware changes over the life of vehicle production may force vehicles out of compliance in certain locations in ways that cannot be determined by make and model alone.A current example of this is Tesla's modifications to hardware in "autopilot" systems.Tesla shifted from relying on radar sensors for autonomous motion to visible light cameras and even disabled the radar sensors in existing vehicles. 95This type of deprecation could make an existing vehicle uncompliant because these vehicles are now lacking effective obstacle detection as defined in state standards.Recently, Tesla returned to installing radar in equipped vehicles after a noticeable increase in self-driving accidents. 96n autonomous vehicle is a combination of hardware sensors, external communication devices, and software, and it is likely not possible to know in real time whether all the systems in a vehicle are compliant with state regulations at a specific time and location.With technology companies, there is already an issue enforcing state and local regulation because compliance is too complicated, and these problems will be more severe when they are occurring in physical space instead of on screens. 97For instance, most states require operators of photogrammetry or lidar sensors to be licensed to determine the precision location of items in space.While these regulations are not aimed at autonomous vehicles, the action performed by the operator is the same, and states could use existing regulations to limit the use of autonomous vehicles if they are considered a danger. 98As autonomous vehicles become more common and certain technologies become clearly superior, states may also express preference for hardware standards only used by certain brands, essentially limiting how existing vehicles could function across municipalities.Administrative layering is particularly concerning due to the complexity of the technology and the potential risks associated with autonomous vehicles, including data privacy issues. 99Smoothly functioning autonomous vehicles require collaboration between vehicle manufacturers, software developers, and regulators of transportation systems and other infrastructure.A relationship this complicated may thus require unified federal standards.
It is worth noting that while AVs are the presently most tangible and visible category of AI-enabled devices, they are far from the only category subject to these regulations.Personal and commercial unmanned aerial vehicles are increasingly controlled autonomously, and we are currently experiencing the beginning of widescale adoption of generative AI for everything from chatbots to writing code, class papers, or even legal opinions.100Like autonomous vehicles, sometimes these tools have disastrous results,101 which is followed by heavy-handed regulation that leads to issues with effective enforcement and inconsistent performance across municipalities. 102In the United States, the bizarre inconsistencies in drug and firearm policies serve as good examples of the complications with these regulations.Simply walking across a border can make an action or item a criminal offense, leading to the existence of complicated tools to predict the legality of an action or item by geography. 103dditionally, these tools will run into subject matter regulatory issues for things like debt collection or medical discussions.Hence, this is why we use AVs as a case study but present the concepts of regulatory fragmentation and federalism as issues for AI technology, broadly speaking.

Discussion and conclusion
Federalism can be an incubator for policy innovation and learning.But it can also increase friction, fragmentation, and dis-integrated policies that are layered on top of each other.Like so many policy domains in the United States, we argue that significant policy innovation in AI will be driven by the states.Our review of the literatures on policy fragmentation and innovation, business regulation, and social equity demonstrates that the outcomes of state-level policy experimentation are not determinative.State competition to attract innovative AI firms could be good for entrepreneurs but problematic for national firms that must navigate a web of conflicting rules.Policy experimentation could offer best practices for addressing the myriad privacy and social equity concerns that surround different AI-driven technology applications, but fragmentation could also lead to unequal treatment and outcomes across the states.Our more focused analysis of the current state of autonomous vehicle policy in the states illustrates fragmentation and shows how administrative layering can significantly complicate the development and deployment of emergent AI technology.
Our aim with this article is to draw attention to studying state-level policy.Most research on AI governance is focused on national governments.But political science, public administration, and policy science research offers theory and methods that will be useful for understanding AI policy and its 97 Olson (2022).98   See https://www.asprs.org/news-resources/state-licensure-map-lidar-and-topographic-products.99 Sella-Villa and Hodgson (2023).effects on businesses and citizens.Furthermore, the states and their variation in political, social, economic, and policy contexts have much to offer for significant theoretical development and empirical testing. 104We did not set out to answer these questions, but to raise them to set an agenda for future AI policy and governance research.
The advent of new technologies presents policymakers with a unique set of challenges when dealing with new ventures.Primarily, these challenges involve the decision-making and knowledge-creating processes.On one hand, the utilization of digital technologies, such as AI and platforms, can shift the decision-making power from humans to machines.This shift raises the question of who should be held accountable for the decisions made by AI.Further research is needed to investigate how ventures allocate their responsibilities and how regulators can intervene to affect responsibility allocation after AI is adopted.On the other hand, AI technologies involve more stakeholders in the knowledge creation process, thereby creating potential conflicts of interests between ventures and stakeholders.Further research is needed to explore how ventures can resolve these conflicts and how policymakers can help coordinate the relationship between ventures and stakeholders.
Policymakers and academicians must also continue to address equity concerns in AI policy and what such policies mean for democracy.As governments expand the use of automated decisionmaking mechanisms for various public functions, these systems must be regulated and closely examined for equity, while keeping changing political contexts and the national mood in mind.Future research in this context must consider how fragmentation impacts AI policy implementation, the overall governance of AI, and how the implementation of AI technology by governments and private firms results in disparate outcomes for citizens.
Researchers can also leverage the diversity in policy approaches among the states to better understand the effects of different policy designs and tools on businesses and consumers.In this respect, the potential research questions are endless, as new applications of AI will continue to rapidly emerge.However, examples could include the effects of Tik Tok bans on social media activity, the effects of licensing and liability rules on autonomous vehicle uptake and use, the effects of algorithmic transparency and public engagement in the development of public trust of AI, and much more.
Studying the dynamic process of AI policy innovation across the states is also fruitful grounds for pushing forward theories of the policy process.Policy innovation and diffusion theory is an obvious potential beneficiary.Much is left to be unpacked from Table 1, including regional patterns of policy adoption and the specific drivers of AV policy innovation.AI is a highly technical policy, and thus it can be useful for expanding research on the effects of policy attributes on diffusion. 105This will require additional data collection not only on AV policy, but also on other policies that emerge to address different AI technologies (e.g., generative AI).Data collection on state variation in policy tools, rhetoric and narratives, advocacy coalitions, administrative layering, collaborative governance, institutional grammar, policy feedback effects, and much more can help advance multiple theories. 106nterdisciplinary research on AI policy that brings together business, political science, public administration, and public policy to study variation in state AI policy can thus yield significant dividends for both scholarly understanding and effective practice.

Table 1 .
Autonomous vehicle policy components by state.