EU’s AI Act: Pioneering Global Standards for Artificial Intelligence

The EU is a Trailblazer, and the AI Act Proves It

On August 2, 2025, the second stage of the EU AI Act came into force, introducing obligations for general purpose models. Initially enacted in February 2025, the AI Act represents the world’s first comprehensive legislation on artificial intelligence, rolling out in multiple stages, with the final implementation phase expected by August 2, 2027.

This landmark regulation is likened to the GDPR for data privacy in the 2010s, establishing a global framework for governing the transformative technology of AI. The urgency for such regulation has been underscored by recent tragedies, including a legal case against OpenAI related to the use of ChatGPT by a teenager, highlighting the risks associated with unregulated AI technologies.

Risk Management Approach

The AI Act adopts a risk management framework aimed at enhancing transparency and accountability for AI systems and their developers. The legislation categorizes AI systems into various risk levels, with the first wave targeting “unacceptable risk” systems that pose clear threats to societal safety. The recent wave, implemented in August 2025, focuses on general purpose AI (GPAI) models, establishing crucial compliance requirements.

Compliance Requirements for GPAI Models

From August 2, 2025, providers of GPAI systems are required to adhere to transparency and copyright obligations when marketing their models within the EU. These obligations extend to organizations operating in the EU, regardless of their location. GPAI models that were available prior to this date must achieve compliance by August 2, 2027.

For the purposes of the law, GPAI models encompass those trained with over 10^23 floating point operations (FLOP) and capable of generating language (text or audio), text-to-image, or text-to-video content. Providers must maintain technical documentation detailing their models, including a comprehensive summary of the training data used.

Systemic Risk Tier

Within the GPAI category, there exists a specialized tier identified as posing systemic risk. Models in this tier require additional compliance measures, including notifying the European Commission when developing such models and ensuring their safety and security. The classification of what constitutes systemic risk may evolve as technology advances. Notably, AI applications related to national security and military purposes are exempt from these regulations, as are open-source systems developed using publicly available code.

GPAI Code of Practice

The European Commission has released a template to assist providers in summarizing the data used for training their models. This GPAI Code of Practice was crafted by independent experts as a voluntary guideline for AI providers to demonstrate compliance with the AI Act. Notable signatories include major tech companies such as Amazon, Google, IBM, and Microsoft. However, some significant players, like Meta, have not signed on.

Criticism and Challenges

The AI Act has faced criticism for its disproportionate impact on startups and small to medium-sized businesses (SMBs), with some experts advocating for exceptions for nascent technologies that have yet to affect the general public significantly. Concerns have been raised that the legislation could stifle progress among European organizations involved in AI development due to its complexity and stringent requirements. Last July, various tech lobbying groups urged the EU to pause the Act’s implementation, arguing that the rollout had been rushed without adequate consideration of potential consequences.

Conclusion

Despite the criticisms, the AI Act has been developed in collaboration with numerous stakeholders across the private sector during a time when businesses are seeking regulatory clarity. It aims to introduce standardized security practices throughout the EU and establish a benchmark for responsible AI governance globally. Following the enactment of the AI Act, other countries, including the US, may find it increasingly challenging to ignore calls for more responsible AI practices, marking a significant milestone in the evolution of technology regulation.

More Insights

Revolutionizing Drone Regulations: The EU AI Act Explained

The EU AI Act represents a significant regulatory framework that aims to address the challenges posed by artificial intelligence technologies in various sectors, including the burgeoning field of...

Revolutionizing Drone Regulations: The EU AI Act Explained

The EU AI Act represents a significant regulatory framework that aims to address the challenges posed by artificial intelligence technologies in various sectors, including the burgeoning field of...

Embracing Responsible AI to Mitigate Legal Risks

Businesses must prioritize responsible AI as a frontline defense against legal, financial, and reputational risks, particularly in understanding data lineage. Ignoring these responsibilities could...

AI Governance: Addressing the Shadow IT Challenge

AI tools are rapidly transforming workplace operations, but much of their adoption is happening without proper oversight, leading to the rise of shadow AI as a security concern. Organizations need to...

EU Delays AI Act Implementation to 2027 Amid Industry Pressure

The EU plans to delay the enforcement of high-risk duties in the AI Act until late 2027, allowing companies more time to comply with the regulations. However, this move has drawn criticism from rights...

White House Challenges GAIN AI Act Amid Nvidia Export Controversy

The White House is pushing back against the bipartisan GAIN AI Act, which aims to prioritize U.S. companies in acquiring advanced AI chips. This resistance reflects a strategic decision to maintain...

Experts Warn of EU AI Act’s Impact on Medtech Innovation

Experts at the 2025 European Digital Technology and Software conference expressed concerns that the EU AI Act could hinder the launch of new medtech products in the European market. They emphasized...

Ethical AI: Transforming Compliance into Innovation

Enterprises are racing to innovate with artificial intelligence, often without the proper compliance measures in place. By embedding privacy and ethics into the development lifecycle, organizations...

AI Hiring Compliance Risks Uncovered

Artificial intelligence is reshaping recruitment, with the percentage of HR leaders using generative AI increasing from 19% to 61% between 2023 and 2025. However, this efficiency comes with legal...