Privacy Groups Urge Against Delays in AI Regulation

Privacy Concerns Surrounding the AI Act

Recent communications from over 50 organizations, including Access Now, Centre for Democracy and Technology Europe (CDT), and the European Consumer Organisation (BEUC), have raised alarms regarding potential delays and reversals in the enforcement of the AI Act. These groups have urged EU Technology Commissioner Henna Virkkunen to refrain from actions that could undermine the key accountability mechanisms established by the Act.

Concerns Over Delays

The organizations express significant concern over the growing pressure to implement a ‘stop the clock’ mechanism that would suspend or delay the AI Act‘s implementation. They argue that the EU’s ‘simplification’ agenda should not serve as a pretext for deregulation, especially in the absence of credible evidence supporting such measures. The letter emphasizes that EU regulations are built on fundamental values and principles, advocating for efforts to simplify regulations to enhance, rather than dismantle, existing legal protections.

Industry Pressure for Delays

This warning follows a request from CEOs of more than 40 European companies, including ASML, Philips, Siemens, and Mistral, who have called for a two-year clock-stop on the AI Act. Their argument centers on the need for reasonable implementation timelines for companies, allowing time for further simplification of the new rules.

Implementation Timeline of the AI Act

The AI Act is designed to regulate artificial intelligence systems based on the risks they pose to society. While the Act came into force in August 2024, its full application is not expected until 2027. The companies have specifically requested a pause on obligations concerning high-risk AI systems, which are set to take effect in August 2026, as well as for general-purpose AI models (GPAI), which are scheduled to be enforced from August 2025.

Code of Practice and Compliance Issues

Additionally, the Code of Practice for GPAI, which aims to assist AI model providers like ChatGPT and Gemini in complying with the AI Act, has yet to be finalized. The draft was prepared by experts appointed by the Commission and is now expected to be released in July. Companies are expressing concerns about insufficient time to comply with regulations before their enforcement starts this August, leading them to seek a grace period for compliance.

Future Directions and Challenges

Thomas Regnier, the Commission spokesperson on digital matters, indicated that discussions are ongoing regarding the timing of the implementation of the Code of Practice, with the end of 2025 being considered. This timeline would precede the enforcement deadlines of the AI Act — 2026 for new models and 2027 for existing models.

The drafting process has faced criticism since it began in September 2024, as major technology companies, publishers, and rights-holders have voiced concerns that the proposed rules may violate the EU’s Copyright laws and hinder innovation.

As the landscape of AI regulation continues to evolve, the balance between fostering innovation and ensuring accountability remains a critical challenge for policymakers.

More Insights

AI Regulations: Comparing the EU’s AI Act with Australia’s Approach

Global companies need to navigate the differing AI regulations in the European Union and Australia, with the EU's AI Act setting stringent requirements based on risk levels, while Australia adopts a...

Quebec’s New AI Guidelines for Higher Education

Quebec has released its AI policy for universities and Cégeps, outlining guidelines for the responsible use of generative AI in higher education. The policy aims to address ethical considerations and...

AI Literacy: The Compliance Imperative for Businesses

As AI adoption accelerates, regulatory expectations are rising, particularly with the EU's AI Act, which mandates that all staff must be AI literate. This article emphasizes the importance of...

Germany’s Approach to Implementing the AI Act

Germany is moving forward with the implementation of the EU AI Act, designating the Federal Network Agency (BNetzA) as the central authority for monitoring compliance and promoting innovation. The...

Global Call for AI Safety Standards by 2026

World leaders and AI pioneers are calling on the United Nations to implement binding global safeguards for artificial intelligence by 2026. This initiative aims to address the growing concerns...

Governance in the Era of AI and Zero Trust

In 2025, AI has transitioned from mere buzz to practical application across various industries, highlighting the urgent need for a robust governance framework aligned with the zero trust economy...

AI Governance Shift: From Regulation to Technical Secretariat

The upcoming governance framework on artificial intelligence in India may introduce a "technical secretariat" to coordinate AI policies across government departments, moving away from the previous...

AI Safety as a Catalyst for Innovation in Global Majority Nations

The commentary discusses the tension between regulating AI for safety and promoting innovation, emphasizing that investments in AI safety and security can foster sustainable development in Global...

ASEAN’s AI Governance: Charting a Distinct Path

ASEAN's approach to AI governance is characterized by a consensus-driven, voluntary, and principles-based framework that allows member states to navigate their unique challenges and capacities...