Texas Enacts Innovative AI Governance Law

Texas Adopts the Responsible AI Governance Act

On June 22, 2025, Texas Governor Abbott signed HB 149, the “Texas Responsible Artificial Intelligence Governance Act” (TRAIGA), establishing Texas as the third U.S. state to implement a comprehensive artificial intelligence (AI) law, following Colorado and Utah. This legislation aims to create a balanced regulatory framework that encourages innovation while addressing potential risks associated with AI systems.

Key Provisions of the Act

The Act is set to take effect on January 1, 2026, providing companies approximately 18 months to develop compliance programs. Key elements of the law include:

  • Baseline duties for AI developers and deployers
  • Prohibition of AI intended for social scoring or discrimination
  • Creation of a regulatory sandbox for testing innovative AI applications
  • Exclusive enforcement authority vested in the Attorney General (AG)
  • Preemption of local AI regulations

Organizations are encouraged to align their compliance efforts with the forthcoming requirements of the EU AI Act and Colorado’s AI regulations, while also monitoring federal developments that may impact state-level enforcement.

Scope and Definitions

The Act applies to any entity that promotes, advertises, or conducts business in Texas, offers products or services to Texas residents, or develops or deploys an AI system within the state. The definition of an “artificial intelligence system” is broad, encompassing any machine-based system that infers outputs from inputs to influence environments, thereby covering various applications, including generative models and recommendation engines.

Responsibilities are assigned based on roles:

  • Developer: Any entity creating an AI system available in Texas.
  • Deployer: Any entity putting an AI system into service or use in Texas.

Transparency and Consumer Protection

Key duties outlined in the Act include:

  • Transparency: Governmental entities must inform individuals when they are interacting with an AI system, ensuring clarity and accessibility in communication.
  • Behavioral Manipulation: Developers and deployers are prohibited from using AI to incite self-harm or criminal activity.
  • Social Scoring Ban: The Act prohibits the use of AI systems for categorizing individuals in a manner that could lead to unfair treatment.
  • Biometric Data Protection: AI systems cannot identify individuals via biometric data without consent, in compliance with constitutional rights.
  • Constitutional Rights and Discrimination: AI systems must not be developed with the intent to infringe upon constitutional rights or discriminate against protected classes.
  • Child Protection: AI systems must not be developed to produce or distribute harmful content involving minors.

The Act sets a high bar for liability, requiring intent to establish violations related to constitutional rights or discrimination, distinguishing it from other state AI laws.

AI Regulatory Sandbox

The Act introduces a 36-month regulatory sandbox, managed by the Department of Information Resources (DIR) in consultation with the Texas Artificial Intelligence Council. Participants can test innovative AI applications without traditional licensing, provided they submit applications detailing their systems and adhere to reporting requirements. Notably, the AG cannot pursue penalties for violations of waived laws during the testing period.

Enforcement and Safe Harbors

The Texas AG holds exclusive enforcement authority, with a structured approach to violations:

  • Notice-and-Cure: The AG must notify violators and allow 60 days to remedy the situation before legal action can be initiated.
  • Civil Penalties: Fines can reach up to $200,000 per violation, with additional daily penalties for ongoing breaches.
  • Statewide Preemption: The Act nullifies local ordinances regulating AI, aiming to create uniformity.
  • No Private Right of Action: The legislation does not allow individuals to sue under the Act, aligning with Colorado regulations.

Safe harbors are also established, protecting entities from civil penalties if violations are discovered through testing or if they comply with recognized AI risk management frameworks.

Comparison with Other Regulations

The Texas Act is compared with Colorado and the EU AI Act, highlighting differences in effective dates, risk frameworks, and transparency obligations:

Aspect Texas Colorado EU AI Act
Effective Date January 1, 2026 February 1, 2026 February 2, 2025 (staggered implementation)
Risk Framework Duties/prohibitions keyed to specific practices Duties/prohibitions keyed to specific practices Framework designed for “High-risk” AI
Transparency Mandatory AI-interaction notice for government entities Consumer notice for high-risk decisions Mandatory disclosure for most AI interactions
Discrimination Standard “Intent to discriminate” required “Algorithmic discrimination” (impact-focused) Fundamental-rights impact assessment for high-risk AI
Sandbox Yes – 36 months, broad No Member States must establish at least one sandbox by August 2, 2026
Penalties AG enforcement; up to $200,000 AG enforcement; penalties to be determined Tiered penalties; up to €35 million or 7% of global turnover

Future Considerations

As Texas navigates the implications of the Responsible AI Governance Act, it faces the potential challenge of a ten-year federal moratorium on AI regulations. Proposed federal legislation could restrict states from enforcing AI-specific laws, affecting Texas’s ability to implement the Act fully.

Companies operating within Texas are encouraged to prepare for the upcoming regulatory environment by:

  • Inventorying and stratifying AI use cases by risk level
  • Documenting compliance with prohibited practices
  • Establishing testing protocols
  • Aligning governance with national AI risk management frameworks

In summary, the Texas Responsible AI Governance Act represents a significant step in regulating AI technologies while fostering an environment conducive to innovation. As companies adapt to these changes, they must remain vigilant in their compliance efforts.

More Insights

AI Regulations: Comparing the EU’s AI Act with Australia’s Approach

Global companies need to navigate the differing AI regulations in the European Union and Australia, with the EU's AI Act setting stringent requirements based on risk levels, while Australia adopts a...

Quebec’s New AI Guidelines for Higher Education

Quebec has released its AI policy for universities and Cégeps, outlining guidelines for the responsible use of generative AI in higher education. The policy aims to address ethical considerations and...

AI Literacy: The Compliance Imperative for Businesses

As AI adoption accelerates, regulatory expectations are rising, particularly with the EU's AI Act, which mandates that all staff must be AI literate. This article emphasizes the importance of...

Germany’s Approach to Implementing the AI Act

Germany is moving forward with the implementation of the EU AI Act, designating the Federal Network Agency (BNetzA) as the central authority for monitoring compliance and promoting innovation. The...

Global Call for AI Safety Standards by 2026

World leaders and AI pioneers are calling on the United Nations to implement binding global safeguards for artificial intelligence by 2026. This initiative aims to address the growing concerns...

Governance in the Era of AI and Zero Trust

In 2025, AI has transitioned from mere buzz to practical application across various industries, highlighting the urgent need for a robust governance framework aligned with the zero trust economy...

AI Governance Shift: From Regulation to Technical Secretariat

The upcoming governance framework on artificial intelligence in India may introduce a "technical secretariat" to coordinate AI policies across government departments, moving away from the previous...

AI Safety as a Catalyst for Innovation in Global Majority Nations

The commentary discusses the tension between regulating AI for safety and promoting innovation, emphasizing that investments in AI safety and security can foster sustainable development in Global...

ASEAN’s AI Governance: Charting a Distinct Path

ASEAN's approach to AI governance is characterized by a consensus-driven, voluntary, and principles-based framework that allows member states to navigate their unique challenges and capacities...