State AI Regulation: A Bipartisan Debate on Federal Preemption

Moratorium on State AI Regulation: A Mixed Response from GOP

The recent moratorium on state artificial intelligence (AI) regulation has drawn varied reactions from Republican lawmakers, with some expressing concern while others commend the decision. As the political landscape continues to evolve, the implications of this moratorium could significantly influence AI development across the United States.

Legislative Context

Both the House and Senate versions of the One Big Beautiful Bill Act (OBBBA) include provisions aimed at preempting state regulations on AI. This legislative initiative coincides with ongoing discussions surrounding the Tax Cuts and Jobs Act (TCJA) and its implications for state and local tax (SALT) deductions.

Opponents of the TCJA’s SALT cap argue that it disproportionately affects residents in blue states, which tend to have higher tax burdens. However, the provision in OBBBA that prohibits states from regulating AI has sparked criticism from within the GOP itself. Notably, Congresswoman Marjorie Taylor Greene (R-Ga.) has publicly opposed the AI preemption, insisting it should be removed in the Senate.

The Debate on Federal vs. State Authority

Greene contended that the bill undermines state authority, stating, “This needs to be stripped out in the Senate. We should be reducing federal power and preserving state power.” In contrast, some former tech officials believe that allowing states to regulate AI could lead to a fragmented landscape of conflicting regulations.

Former chief technologist at the Federal Trade Commission, Neil Chilson, responded to Greene’s sentiments by questioning the wisdom of allowing states like California to dictate national AI policy, suggesting that the approach could inadvertently benefit foreign competitors such as China.

State-Level Regulatory Movements

The push for state-level regulation is exemplified by the recent passage of the RAISE Act in New York. This legislation aims to impose new regulations on companies involved in AI, raising concerns among industry stakeholders. A letter from NetChoice, a trade association representing online businesses, warned that the RAISE Act could stifle innovation and economic competitiveness.

In Texas, a state typically viewed as a bastion of conservatism, lawmakers have also sought to regulate AI through the introduction of the Texas Responsible AI Governance Act (TRAIGA). Although initially met with resistance, the bill was eventually reworked and passed as House Bill 149, which focuses on government use of AI rather than broader industry regulations.

Concerns Over a Patchwork of Regulations

Governors from various states have expressed apprehension about the prospect of a fragmented regulatory environment. Governor Ned Lamont of Connecticut highlighted the potential challenges posed by a state-by-state approach, emphasizing the need for cohesive federal legislation. Similarly, Governor Jared Polis of Colorado has advocated for federal preemption to avoid conflicting state regulations.

In Virginia, Governor Glenn Youngkin rejected an AI regulation bill, arguing that a heavy-handed regulatory approach would stifle innovation. He emphasized the importance of enabling creators to thrive rather than imposing burdensome regulations.

Prospects for Federal Preemption

The call for federal preemption of state AI regulations is gaining traction among bipartisan lawmakers. Advocates assert that a unified regulatory framework is essential for maintaining the United States’ competitive edge in AI development. Vance Ginn, president of Ginn Economic Consulting, pointed to the Internet Tax Freedom Act of 1998 as a precedent for a federal moratorium that spurred digital innovation.

As discussions continue, it remains to be seen whether Congress will prioritize a cohesive approach to AI regulation or allow states to dictate their policies, a decision that could have lasting repercussions for the future of technology in America.

More Insights

Harnessing Generative AI for Enhanced Risk and Compliance in 2025

In 2025, the demand for Generative AI in risk and compliance certification is surging as organizations face complex regulatory landscapes and increasing threats. This certification equips...

Building Sustainable Generative AI: Mitigating Carbon Emissions

Generative AI is revolutionizing industries, but it comes with a significant environmental cost due to carbon emissions from extensive compute resources. As the demand for large-scale models grows...

AI Regulation: Balancing Innovation and Oversight

Experts discuss the implications of the recently passed H.R. 1, which would pause state and local regulations on artificial intelligence for ten years. The article examines the benefits and drawbacks...

AI Governance in India: Shaping the Future of Technology

This article examines the evolving landscape of AI governance in India, highlighting both the initiatives aimed at promoting AI adoption and the regulatory frameworks being developed to manage...

AI’s Shadow: Exposing and Addressing Harms Against Women and Girls

AI's rapid advancement presents risks, especially for vulnerable populations targeted by cyber-harassment, hate speech, and impersonation. AI systems can amplify biases and be exploited to harm...

AI Readiness Framework for the Pharmaceutical Industry

This article presents an AI readiness assessment framework tailored for the pharmaceutical industry, emphasizing the importance of aligning AI initiatives with regulatory standards and ethical...

AI as a Strategic Partner in Governance

The UAE has announced that a National Artificial Intelligence System will become a non-voting member of all federal and government company boards, marking a significant shift in governance. This...

New Code of Practice for AI Compliance Set for 2025

The European Commission announced that a code of practice to help companies comply with the EU's artificial intelligence rules may only be implemented by the end of 2025. This delay follows calls from...

New Code of Practice for AI Compliance Set for 2025

The European Commission announced that a code of practice to help companies comply with the EU's artificial intelligence rules may only be implemented by the end of 2025. This delay follows calls from...