Colorado’s Struggle with AI Regulation

Inside the Controversy Over Colorado’s AI Law

Colorado’s law aims to prevent AI from being used as a tool of discrimination. It’s intended to apply to “high-risk” AI systems that make or assist in making consequential decisions.

From the onset, the governor sought to have legislators update the law before its implementation to avoid an overly burdensome regulatory framework. However, attempts to revise it during an August special session fell apart. A major sticking point was determining how much liability AI developers and deployers should hold if the AI’s decisions are discriminatory.

Unable to reach an agreement, lawmakers delayed the law’s implementation, planning to revisit the issue during their regular session in January.

Background of the Colorado AI Act

The Colorado AI Act establishes the first comprehensive consumer protections aimed at safeguarding the public from discrimination when AI is used to make decisions regarding health care, employment, or other significant areas. Despite its groundbreaking nature, the law has yet to take effect, with its implementation date continually pushed back.

Concerns arose from companies, including Governor Jared Polis, about the potential burdens the new law could impose, potentially dampening AI innovation. Polis had signed the original bill into law with the expectation that it would be revised prior to its 2026 implementation.

Polis indicated that the law creates a complex compliance regime for AI developers and deployers, particularly because it holds companies accountable for unintentional discrimination, which could hinder business operations in Colorado compared to other states.

Failed Revision Efforts

In August, Polis convened a special session to amend the law. Legislators rushed to craft a revision that would satisfy all stakeholders—tech companies, consumer rights advocates, and those utilizing AI tools. Unfortunately, this effort fell apart over disagreements regarding liability for discriminatory impacts associated with AI tools.

Ultimately, legislators opted to delay the law’s effective date by four months, now set for June 2026. However, they plan to meet again in January to attempt further adjustments.

Key Issues and Stakeholder Concerns

The Colorado AI Act targets automated decision-making systems involved in “consequential decisions” such as hiring, loan approvals, and provision of government services. Its aim is to protect consumers from discrimination or bias, regardless of intent.

For instance, Amazon’s 2018 discovery that its automated recruiting tool favored men over women exemplifies the potential pitfalls of AI systems. The law would have held Amazon accountable for any unintentional discriminatory practices stemming from its AI tool.

Many companies expressed concerns that the AI law was too vague, broad, and burdensome. Efforts to revise the law during the regular session fell short, leading to heightened pressure during the August special session. Various proposals emerged, including a narrower bipartisan bill requiring companies to disclose AI interactions with consumers and clarifying that existing consumer protections apply to AI.

However, as discussions progressed, momentum waned, and a proposed Sunshine Act—which would hold developers and deployers jointly liable for discriminatory AI system outcomes—also collapsed due to a lack of compromise.

Challenges in Reaching Consensus

One major challenge cited by stakeholders was defining what constitutes an “automated decision-making system” that should be governed by the law. Additionally, the timeline for negotiations was deemed insufficient for resolving complex legal matters. Many stakeholders felt that the perspectives of both established innovators and frontline sectors using AI were overlooked during discussions.

Despite these hurdles, some smaller tech firms expressed willingness to assume liability for faulty AI products. In contrast, larger companies resisted liability, complicating negotiations further.

The Path Forward

The delay has garnered mixed reactions. While some tech organizations appreciate the additional time to refine legislation, consumer advocates argue that residents remain unprotected from potential AI harms. As the June 2026 implementation date approaches, the urgency to establish consumer protections intensifies.

The situation exemplifies a pivotal moment in AI regulation; stakeholders must collaborate effectively to forge a path forward. If significant progress is not made soon, the risk of consumers remaining defenseless against the harms of pervasive AI systems increases.

The ongoing discourse around the Colorado AI Act serves as a case study for other states seeking to navigate the complexities of AI regulation and consumer protection.

More Insights

US Rejects UN’s Call for Global AI Governance Framework

U.S. officials rejected the establishment of a global AI governance framework at the United Nations General Assembly, despite broad support from many nations, including China. Michael Kratsios of the...

Agentic AI: Managing the Risks of Autonomous Systems

As companies increasingly adopt agentic AI systems for autonomous decision-making, they face the emerging challenge of agentic AI sprawl, which can lead to security vulnerabilities and operational...

AI as a New Opinion Gatekeeper: Addressing Hidden Biases

As large language models (LLMs) become increasingly integrated into sectors like healthcare and finance, a new study highlights the potential for subtle biases in AI systems to distort public...

AI Accountability: A New Era of Regulation and Compliance

The burgeoning world of Artificial Intelligence (AI) is at a critical juncture as regulatory actions signal a new era of accountability and ethical deployment. Recent events highlight the shift...

Choosing Effective AI Governance Tools for Safer Adoption

As generative AI continues to evolve, so do the associated risks, making AI governance tools essential for managing these challenges. This initiative, in collaboration with Tokio Marine Group, aims to...

UN Initiatives for Trustworthy AI Governance

The United Nations is working to influence global policy on artificial intelligence by establishing an expert panel to develop standards for "safe, secure and trustworthy" AI. This initiative aims to...

Data-Driven Governance: Shaping AI Regulation in Singapore

The conversation between Thomas Roehm from SAS and Frankie Phua from United Overseas Bank at the SAS Innovate On Tour in Singapore explores how data-driven regulation can effectively govern rapidly...

Preparing SMEs for EU AI Compliance Challenges

Small and medium-sized enterprises (SMEs) must navigate the complexities of the EU AI Act, which categorizes many AI applications as "high-risk" and imposes strict compliance requirements. To adapt...

Draft Guidance on Reporting Serious Incidents Under the EU AI Act

On September 26, 2025, the European Commission published draft guidance on serious incident reporting requirements for high-risk AI systems under the EU AI Act. Organizations developing or deploying...