AI Regulation: Paving the Way for Trustworthy Technology

Legislation on AI

The AI legislation represents the world’s first comprehensive legal framework addressing the risks associated with artificial intelligence (AI) and positions Europe as a leading player on the global stage.

This AI legislation establishes clear requirements and obligations for developers and deployers of AI concerning specific uses of the technology. Simultaneously, the regulation aims to reduce the administrative and financial burdens on businesses, particularly small and medium enterprises (SMEs).

Part of a broader set of measures to promote trustworthy AI, the legislation also encompasses initiatives supporting innovation in the AI sector. Collectively, these measures ensure the safety and fundamental rights of individuals and businesses regarding AI, while also enhancing adoption, investment, and innovation across the EU.

Why Do We Need AI Rules?

The AI legislation ensures that Europeans can trust the benefits AI has to offer. While most AI systems pose limited or no risk and can help address numerous societal challenges, some applications create risks that must be managed to avoid undesirable outcomes.

For instance, the decision-making process of an AI system can often be opaque, making it hard to assess whether an individual has been unfairly disadvantaged in situations like hiring or public service applications.

Existing legislation offers some protection but is insufficient to tackle the specific challenges posed by AI systems.

Key Objectives of the New Rules:

  • Address specific risks created by AI applications.
  • Prohibit AI practices that present unacceptable risks.
  • Establish a list of high-risk applications.
  • Set clear requirements for high-risk AI systems.
  • Define specific obligations for deployers and providers of high-risk AI applications.
  • Require compliance assessments before the deployment or marketing of AI systems.
  • Implement enforcement measures post-market for AI systems.
  • Establish a governance structure at the European and national levels.

A Risk-Based Approach

The regulatory framework categorizes AI systems into four risk levels:

All AI systems deemed a clear threat to safety, livelihoods, and rights of individuals are prohibited. This includes social scoring by governments and toys using voice assistance that encourage dangerous behavior.

High Risk

AI systems categorized as high risk include technology used in:

  • Critical infrastructures (e.g., transportation) that could endanger citizens’ lives and health.
  • Educational and vocational training, determining access to education and career paths (e.g., exam grading).
  • Safety components of products (e.g., AI applications in robotic surgery).
  • Employment and worker management (e.g., resume screening software).
  • Essential public and private services (e.g., credit scoring preventing citizens from obtaining loans).
  • Law enforcement services interfering with fundamental rights (e.g., reliability assessment of evidence).
  • Migration management (e.g., automated visa application assessments).
  • Judicial administration and democratic processes (e.g., AI solutions for legal decision searches).

High-risk AI systems are subjected to strict obligations before they can be marketed:

  • Adequate systems for risk assessment and mitigation.
  • High-quality datasets to minimize risks and discriminatory outcomes.
  • Activity logging to ensure traceability of results.
  • Detailed documentation to provide necessary information for compliance evaluation.
  • Clear and adequate information from the deployer.
  • Appropriate human oversight measures to minimize risks.
  • High levels of robustness, security, and accuracy.

All remote biometric identification systems are considered high risk, subjected to strict requirements. The use of remote biometric identification in public spaces for law enforcement purposes is generally prohibited, with narrowly defined exceptions.

Limited Risk

Limited risk refers to the risks associated with a lack of transparency in AI use. The legislation introduces specific transparency obligations to ensure that individuals are informed when necessary, fostering trust. For example, when interacting with AI systems such as chatbots, individuals must be informed they are engaging with a machine, allowing them to make informed decisions. Additionally, AI-generated content must be identifiable, and AI-generated texts published to inform the public must be labeled as artificially generated.

Minimal or No Risk

The AI legislation permits the free use of AI with minimal risk, including applications such as AI-compatible video games or anti-spam filters. The vast majority of AI systems currently in use within the EU fall into this category.

Operational Framework for High-Risk AI System Providers

Once an AI system is on the market, authorities are responsible for market surveillance, deployers ensure human monitoring, and providers maintain a post-market monitoring system. Providers and deployers are also required to report serious incidents and malfunctions.

A Solution for Reliable Use of Large AI Models

Increasingly, general-purpose AI models become components of AI systems, capable of performing and adapting to countless tasks.

While general-purpose AI models can enable better and more powerful AI solutions, supervising all their capabilities poses challenges. The AI legislation introduces transparency obligations for all general-purpose AI models to enhance understanding of these models and additional risk management requirements for high-performing and impactful models.

A Forward-Looking Legislation

Given the rapidly evolving nature of AI technology, the regulation adopts a future-proof approach, allowing rules to adapt to technological advancements. AI applications should remain reliable even post-market, necessitating ongoing quality and risk management by providers.

Application and Implementation

The European AI Office, established in February 2024 within the Commission, oversees the application and implementation of AI legislation alongside member states. Its goal is to create an environment where AI technologies uphold human dignity, rights, and trust while fostering collaboration, innovation, and AI research among stakeholders. Additionally, it engages in international dialogue on AI governance, recognizing the need for global alignment.

Next Steps

The AI law took effect on August 1 and will be fully applicable two years later, with a few exceptions: bans will take effect after six months, governance rules and obligations for general-purpose AI models will apply after 12 months, and regulations for AI systems integrated into regulated products will commence after 36 months. To facilitate the transition to the new regulatory framework, the Commission launched a voluntary AI Pact, inviting AI developers in Europe and beyond to comply in advance with the primary obligations outlined in the AI legislation.

More Insights

Responsible AI Workflows for Transforming UX Research

The article discusses how AI can transform UX research by improving efficiency and enabling deeper insights, while emphasizing the importance of human oversight to avoid biases and inaccuracies. It...

Revolutionizing Banking with Agentic AI

Agentic AI is transforming the banking sector by automating complex processes, enhancing customer experiences, and ensuring regulatory compliance. However, it also introduces challenges related to...

AI-Driven Compliance: The Future of Scalable Crypto Infrastructure

The explosive growth of the crypto industry has brought about numerous regulatory challenges, making AI-native compliance systems essential for scalability and operational efficiency. These systems...

ASEAN’s Evolving AI Governance Landscape

The Association of Southeast Asian Nations (ASEAN) is making progress toward AI governance through an innovation-friendly approach, but growing AI-related risks highlight the need for more binding...

EU AI Act vs. US AI Action Plan: A Risk Perspective

Dr. Cari Miller discusses the differences between the EU AI Act and the US AI Action Plan, highlighting that the EU framework is much more risk-aware and imposes binding obligations on high-risk AI...

The Hidden Risks of AI Integration in the Workplace

As organizations rush to adopt AI, many are ignoring the critical risks involved, such as compliance and oversight issues. Without proper governance and human management, AI can quickly become a...

Investing in AI Safety: Capitalizing on the Future of Responsible Innovation

The AI safety collaboration imperative is becoming essential as the artificial intelligence revolution reshapes industries and daily life. Investors are encouraged to capitalize on this opportunity by...

AI Innovations in Modern Policing

Law enforcement agencies are increasingly leveraging artificial intelligence to enhance their operations, particularly in predictive policing. The integration of technology offers immense potential...

Kenya’s Pivotal Role in UN’s Groundbreaking AI Governance Agreement

Kenya has achieved a significant diplomatic success by leading the establishment of two landmark institutions for governing artificial intelligence (AI) at the United Nations. The Independent...