The Risks of Abandoning AI Liability Regulations

Op-ed: Abandoning the AI Liability Directive Brings Unacceptable Risks

Europe’s need to cut red tape is no secret. This issue frequently arises in discussions with businesses, whether they are startups, scale-ups, or established companies. The European Commission has pledged to deliver on this goal, yet doubts persist about the means to achieve it.

The AI Liability Directive (AILD), which the European Commission has decided to abandon, exemplifies this concern. Proponents of this decision argue that additional liability rules could stifle innovation and investment in Europe. However, scrapping the directive could achieve the opposite effect: by leaving companies without clear legal guidelines, the Commission will reduce their incentives to invest.

Legal Uncertainty: A Barrier to AI Innovation in the EU

Investors in Europe are already known for their risk aversion. As AI technologies increasingly interact with both the real and virtual worlds, the risks multiply, and the Commission’s decision adds legal opacity and fragmentation to the mix.

The chains of accountability remain unclear. Who is responsible when risks inevitably materialize—those who develop, deploy, sell, or design? What happens if they share responsibilities among each other? The search for these answers reveals a fragmented legal landscape.

Currently, companies dealing with AI-driven technologies have little idea how innovative the judge facing them might be, nor which of the 27 legal frameworks will confront them.

AILD’s Role in Europe’s Digital Rulebook

Some opponents of the directive argue that there’s no need for further regulation since the AI Act and the new Product Liability Directive (PLD) cover the same ground. This perspective is misguided. Neither the AI Act nor the revised PLD substitutes for the AILD.

The distinction is clear: the AI Act deals with pre-emptive risk management, guiding AI players on how to avoid harm. It does not address who is responsible after harm has occurred. The Product Liability Directive, conversely, covers damages following an incident, but these are different from those addressed by the AILD. The differences between product liability and producer’s liability are well-known and should be recognized by the Commission.

Without AILD, AI Risks Undermining Trust & Safety

AI harms often extend beyond product defects. For instance, what if AI causes damage in a professional context using professional tools? What if the harm arises not from a manufacturing defect but from inadequate user instructions? What if the injury results from a “rogue” AI behavior not rooted in technical fault but in deployment mismanagement?

A growing class of use cases involves programmers using generative AI without apparent defects to create applications that include AI elements. What if such privately used applications cause harm to third parties? Ignoring these scenarios represents not just a legal blind spot but a significant political liability.

The Commission must understand better. By refusing to adopt harmonized AI liability rules, it exposes businesses to a patchwork of national standards and conflicting interpretations, which is detrimental to the acceleration of AI uptake across the continent.

Instead of clarity, we encounter a game of legal roulette. In this case, harmonization does not imply overregulation; it represents smart, targeted, fact-based rules that provide both innovators and consumers with legal certainty.

The current opacity, seeming autonomy, and unpredictability for users complicate the pinpointing of responsibility. The AILD aimed to close these gaps through reasonable, modern tools like disclosure duties and rebuttable presumptions of fault—measures designed for AI’s unique risks.

The Commission’s vague hints about “future legal approaches” offer little comfort. Businesses need legal certainty now, not open-ended promises for the future.

At the heart of this debate lies a broader question: Do we genuinely desire a digital single market in Europe that transcends mere rhetoric? If the answer is affirmative, harmonization is essential and must be grounded in fact. Without it, we risk more fragmentation, not predictability; more confusion, not clarity. With its latest retreat, the Commission isn’t simplifying—it’s surrendering.

More Insights

Building Trust in AI: Strategies for a Secure Future

The Digital Trust Summit 2025 highlighted the urgent need for organizations to embed trust, fairness, and transparency into AI systems from the outset. As AI continues to evolve, strong governance and...

Rethinking Cloud Governance for AI Innovation

As organizations embrace AI innovations, they often overlook the need for updated cloud governance models that can keep pace with rapid advancements. Effective governance should be proactive and...

AI Governance: A Guide for Board Leaders

The Confederation of Indian Industry (CII) has released a guidebook aimed at helping company boards responsibly adopt and govern Artificial Intelligence (AI) technologies. The publication emphasizes...

Harnessing AI for Secure DevSecOps in a Zero-Trust Environment

The article discusses the implications of AI-powered automation in DevSecOps, highlighting the balance between efficiency and the risks associated with reliance on AI in security practices. It...

Establishing India’s First Centre for AI, Law & Regulation

Cyril Amarchand Mangaldas, Cyril Shroff, and O.P. Jindal Global University have announced the establishment of the Cyril Shroff Centre for AI, Law & Regulation, the first dedicated centre in India...

Revolutionizing AI Governance for Local Agencies with a Free Policy Tool

Darwin has launched its AI Policy Wizard, a free and interactive tool designed to assist local governments and public agencies in creating customized AI policies. The tool simplifies the process by...

Building Trust in AI Through Effective Governance

Ulla Coester emphasizes the importance of adaptable governance in building trust in AI, highlighting that unclear threats complicate global confidence in the technology. She advocates for...

Building Trustworthy AI Through Cultural Engagement

This report emphasizes the importance of inclusive AI governance to ensure diverse voices, especially from the Global South, are involved in AI access and development decisions. It highlights the...

AI Compliance: Copyright Challenges in the EU AI Act

The EU AI Act emphasizes the importance of copyright compliance for generative AI models, particularly regarding the use of vast datasets for training. It requires general-purpose AI providers to...