EU AI Act Set for Review: Potential Changes Ahead

EU’s AI Act Undergoing Potential Revisions

The European Union’s AI Act, hailed as one of the most comprehensive regulatory frameworks governing artificial intelligence, is poised for modifications as part of an upcoming “fitness check” aimed at evaluating existing tech regulations.

Current Landscape of the AI Act

Since its introduction, the AI Act has gradually rolled out a series of obligations for organizations utilizing AI technologies. The Act emphasizes key principles such as ethics, safety, and transparency, setting a precedent for how AI systems should operate within the EU.

Insights from the European Commission

Lucilla Sioli, who oversees the AI and digital industry at the European Commission, recently indicated that the AI Act might be subject to adjustments as part of a broader review of technological regulations. This review is anticipated before the end of the year. Sioli remarked, “It’s a reflection that is taking place, and we will see what it brings about,” highlighting the dynamic nature of regulatory processes.

Balancing Regulation and Innovation

The EU has often been perceived as a heavy-handed regulator, with concerns that stringent rules may stifle technological innovation. In response, the bloc is prioritizing the development of homegrown AI startups capable of competing with well-funded firms based in the United States.

Critics argue that the EU’s rigorous regulations could hinder progress, with some industry leaders advocating for a “DOGE for Europe”—a nod to the deregulated environment in the US, which many believe fosters innovation.

Recent Developments and Future Implications

Earlier this month, a new set of regulations took effect, which includes the outright banning of certain AI applications. Organizations that utilize AI for any of these prohibited purposes—such as emotion recognition, facial databases, and social scoring—face severe penalties, including fines of up to €35 million or 7% of their annual global turnover, whichever amount is higher.

The next critical date for AI enterprises operating within the EU is August 2, when rules surrounding general-purpose AI systems, such as OpenAI’s ChatGPT, will come into effect. Ongoing discussions among lawmakers and industry leaders focus on the precise enforcement of these rules, with practical guidance expected to be released in the spring.

As the landscape of AI regulation evolves, stakeholders in the technology sector will need to remain vigilant and adaptable to the regulatory frameworks that govern their operations.

More Insights

US Rejects UN’s Call for Global AI Governance Framework

U.S. officials rejected the establishment of a global AI governance framework at the United Nations General Assembly, despite broad support from many nations, including China. Michael Kratsios of the...

Agentic AI: Managing the Risks of Autonomous Systems

As companies increasingly adopt agentic AI systems for autonomous decision-making, they face the emerging challenge of agentic AI sprawl, which can lead to security vulnerabilities and operational...

AI as a New Opinion Gatekeeper: Addressing Hidden Biases

As large language models (LLMs) become increasingly integrated into sectors like healthcare and finance, a new study highlights the potential for subtle biases in AI systems to distort public...

AI Accountability: A New Era of Regulation and Compliance

The burgeoning world of Artificial Intelligence (AI) is at a critical juncture as regulatory actions signal a new era of accountability and ethical deployment. Recent events highlight the shift...

Choosing Effective AI Governance Tools for Safer Adoption

As generative AI continues to evolve, so do the associated risks, making AI governance tools essential for managing these challenges. This initiative, in collaboration with Tokio Marine Group, aims to...

UN Initiatives for Trustworthy AI Governance

The United Nations is working to influence global policy on artificial intelligence by establishing an expert panel to develop standards for "safe, secure and trustworthy" AI. This initiative aims to...

Data-Driven Governance: Shaping AI Regulation in Singapore

The conversation between Thomas Roehm from SAS and Frankie Phua from United Overseas Bank at the SAS Innovate On Tour in Singapore explores how data-driven regulation can effectively govern rapidly...

Preparing SMEs for EU AI Compliance Challenges

Small and medium-sized enterprises (SMEs) must navigate the complexities of the EU AI Act, which categorizes many AI applications as "high-risk" and imposes strict compliance requirements. To adapt...

Draft Guidance on Reporting Serious Incidents Under the EU AI Act

On September 26, 2025, the European Commission published draft guidance on serious incident reporting requirements for high-risk AI systems under the EU AI Act. Organizations developing or deploying...