The EU AI Act: Redefining Public Law in the Digital Age

The Hidden Reach of the EU AI Act: Expanding the Scope of EU Public Power

In June 2024, the European Union adopted the AI Act, marking a groundbreaking initiative as the first comprehensive attempt globally to regulate artificial intelligence (AI). As AI systems become increasingly prevalent across various domains, the implications of the AI Act grow more significant.

At the national level, the experimentation with AI systems in the public domain is predominantly observed in areas such as public administration. Notable examples include:

  • The allocation of public benefits
  • The prevention of crime
  • The evaluation of visa or asylum requests

In these scenarios, authorities may utilize AI-generated risk profiles to support decision-making processes. By mandating Member States to align their laws and practices with the AI Act, the regulation is set to have a significant impact on national administrative procedures.

Broader Implications of the AI Act

This document outlines a largely overlooked impact of the AI Act: its potential to influence the operations of public authorities beyond the direct scope of its provisions. As EU legislation, the AI Act activates the application of second order EU law, which encompasses general principles of Union law and the EU Charter of Fundamental Rights (CFR).

This activation has profound implications, as it may facilitate the Europeanisation of national administrative law, effectively shifting the boundary between EU public law and domestic public law in the realm of digital regulation. This analysis bridges two pertinent debates: the transformative influence of the EU Charter of Fundamental Rights and the impact of the AI Act on national security law.

The “Trigger Function” of the EU AI Act

Due to the absence of a general fundamental rights competence at the EU level, the Charter can only constrain Member State actions within the existing competences. According to Article 51(1) of the CFR, the Charter’s provisions are directed at Member States when they are implementing Union law. The scope of this relationship was clarified in the case of Åkerberg Fransson, where the Court of Justice of the European Union determined that the Charter applies to all Member State activities falling within EU law.

However, merely adopting the AI Act does not automatically encompass all national activities under EU law. A direct relationship between a national activity and Union law is required for the Charter’s applicability. The Court considers whether a specific rule of EU law is applicable, independent, and distinct from the fundamental right itself. Thus, a Member State activity only falls under EU law if the AI Act regulates it significantly.

The AI Act as a “Gateway” for EU Principles of Procedural Justice

The AI Act, as a piece of product safety legislation, primarily establishes procedural obligations to ensure the accuracy, fairness, and legality of AI outputs, without mandating specific substantive requirements. Consequently, when AI systems are employed in administrative contexts, procedural aspects are more likely to fall under EU law compared to substantive aspects.

For instance, a public authority utilizing AI-generated risk profiles to assess the likelihood of student loan fraud must comply with specific provisions of the AI Act. The authority is required to conduct a prior fundamental rights impact assessment and ensure human oversight during AI system deployment. These activities clearly fall within the purview of EU law.

Conversely, the decisions based on the AI-created risk profiles—such as deciding to conduct home inspections—are less clear-cut. While the AI Act influences these decisions, there are limited provisions that directly govern them. Article 86 of the AI Act mandates that affected individuals receive meaningful explanations regarding the role of the AI system in decision-making processes. This procedural requirement aligns with EU principles of good administration, including the duty to reason and the right to be heard.

Significance and Implications

The practical significance of these provisions is challenging to predict, as it largely depends on the existing obligations for public authorities under national law. In cases where EU law mandates a duty to reason that exceeds national law requirements, authorities must comply with EU standards when AI systems are involved. This may also introduce remedies that do not exist within national law.

However, where national laws already meet EU requirements, the substantive impact might be less noticeable. Regardless of the practical implications, the applicability of EU principles of procedural justice, as opposed to national standards, alters the relationship between EU and national public law.

As highlighted by Advocate General Saugmandsgaard Øe, once applicable, the Charter restricts the regulatory and policy options available to Member States, thereby enhancing the EU’s capacity to define the boundaries of permissible actions.

In conclusion, the AI Act not only establishes extensive obligations for public authorities but also triggers the application of EU principles of procedural justice, contributing to the Europeanisation of national administrative law and redefining the relationship between EU public law and domestic public law in the digital regulation landscape.

More Insights

The Perils of ‘Good Enough’ AI in Compliance

In today's fast-paced world, the allure of 'good enough' AI in compliance can lead to significant legal risks when speed compromises accuracy. Leaders must ensure that AI tools provide explainable...

European Commission Unveils AI Code of Practice for General-Purpose Models

On July 10, 2025, the European Commission published the final version of the General-Purpose AI Code of Practice, which aims to provide a framework for compliance with certain provisions of the EU AI...

EU Introduces New Code to Streamline AI Compliance

The European Union has introduced a voluntary code of practice to assist companies in complying with the upcoming AI Act, which will regulate AI usage across its member states. This code addresses...

Reforming AI Procurement for Government Accountability

This article discusses the importance of procurement processes in the adoption of AI technologies by local governments, highlighting how loopholes can lead to a lack of oversight. It emphasizes the...

Pillar Security Launches Comprehensive AI Security Framework

Pillar Security has developed an AI security framework called the Secure AI Lifecycle Framework (SAIL), aimed at enhancing the industry's approach to AI security through strategy and governance. The...

Tokio Marine Unveils Comprehensive AI Governance Framework

Tokio Marine Holdings has established a formal AI governance framework to guide its global operations in developing and using artificial intelligence. The policy emphasizes transparency, human...

Shadow AI: The Urgent Need for Governance Solutions

Generative AI (GenAI) is rapidly becoming integral to business operations, often without proper oversight or approval, leading to what is termed as Shadow AI. Companies must establish clear governance...

Fragmented Futures: The Battle for AI Regulation

The article discusses the complexities of regulating artificial intelligence (AI) as various countries adopt different approaches to governance, resulting in a fragmented landscape. It explores how...

Fragmented Futures: The Battle for AI Regulation

The article discusses the complexities of regulating artificial intelligence (AI) as various countries adopt different approaches to governance, resulting in a fragmented landscape. It explores how...