Impact of the EU AI Act on Digital Innovation Costs

How the EU’s AI Act Will Shape Digitalization and Innovation Costs

This year, the world’s first comprehensive regulation on artificial intelligence (AI) – the European Union’s Artificial Intelligence Act – came into effect, binding for Latvia as well. Gradually implemented until the summer of 2027, this Act aims to ensure that AI systems are developed and used responsibly.

As the Act’s requirements are clarified at the national level, it is crucial to avoid the mistakes made with the General Data Protection Regulation (GDPR), which resulted in an excessive bureaucratic burden.

Purpose of the AI Act

The primary objective of the AI Act is to promote ethical, safe, transparent, and trustworthy AI usage. While it is essential to mitigate the risks associated with AI, it is equally important to prevent placing all responsibility solely on developers, which could adversely affect competition and the pace of digitalization.

Four Risk Levels Defined by the AI Act

The AI Act classifies AI systems into four distinct risk levels:

  • Unacceptable Risk: AI systems that pose a direct threat to human safety, livelihoods, and rights are completely prohibited. This category includes government-led social scoring systems, similar to those in China, and toys that use voice assistance to promote dangerous behavior.
  • High Risk: Systems utilized in critical infrastructures, surgical applications, exam assessments, hiring procedures, and migration and border control management are classified as high-risk and are subject to stringent requirements.
  • Limited Risk: Systems such as recommendation engines based on user behavior, virtual assistants, and translation systems fall under this category.
  • Minimal Risk: Low-risk systems, including email filters for spam, are considered minimal risk.

With the imposition of penalties for non-compliance, including significant fines, the AI Act will substantially impact IT system development and implementation processes. Understanding the rules regarding permitted and prohibited actions will be crucial for both developers and system clients.

Maintaining Digitalization Advantages

Analysis of AI usage and digitalization in the EU indicates that Latvia is ahead of some Western European countries, like Germany, in implementing and maintaining critical infrastructure systems and making state services accessible. It is imperative not to hinder this progress or repeat the mistakes seen during the GDPR implementation, where misunderstandings complicated various IT sector processes.

Small and medium-sized businesses face particular challenges. Therefore, it is vital to focus on solutions that reduce financial and administrative burdens for these entities.

Companies utilizing AI must balance innovation with legal compliance. The system development process will require initial risk assessments, data traceability procedures, quality control mechanisms, and comprehensive technical documentation throughout the development lifecycle.

Shared Responsibility

Responsibility for compliance with the AI Act’s requirements must be shared between system owners (which may include government and municipal institutions) and developers. If the burden and risk are disproportionately placed on developers, responsible parties may hesitate to participate in various IT projects, potentially diminishing competition and affecting quality.

Moreover, penalties for violations of the Act’s requirements are calculated as a percentage of the company’s total annual revenue from the previous fiscal year, indicating that fines could exceed the total cost of the respective development project.

Conclusion

In response to these challenges, the founding of the Latvian Artificial Intelligence Association (MILA) aims to promote responsible AI use, facilitate cross-border cooperation, and enhance knowledge exchange with government institutions. The Association is positioned to be a strong partner for the government, advocating for proportionality in the AI Act’s implementation process to achieve responsible AI use without stalling digitalization.

More Insights

Revolutionizing Drone Regulations: The EU AI Act Explained

The EU AI Act represents a significant regulatory framework that aims to address the challenges posed by artificial intelligence technologies in various sectors, including the burgeoning field of...

Revolutionizing Drone Regulations: The EU AI Act Explained

The EU AI Act represents a significant regulatory framework that aims to address the challenges posed by artificial intelligence technologies in various sectors, including the burgeoning field of...

Embracing Responsible AI to Mitigate Legal Risks

Businesses must prioritize responsible AI as a frontline defense against legal, financial, and reputational risks, particularly in understanding data lineage. Ignoring these responsibilities could...

AI Governance: Addressing the Shadow IT Challenge

AI tools are rapidly transforming workplace operations, but much of their adoption is happening without proper oversight, leading to the rise of shadow AI as a security concern. Organizations need to...

EU Delays AI Act Implementation to 2027 Amid Industry Pressure

The EU plans to delay the enforcement of high-risk duties in the AI Act until late 2027, allowing companies more time to comply with the regulations. However, this move has drawn criticism from rights...

White House Challenges GAIN AI Act Amid Nvidia Export Controversy

The White House is pushing back against the bipartisan GAIN AI Act, which aims to prioritize U.S. companies in acquiring advanced AI chips. This resistance reflects a strategic decision to maintain...

Experts Warn of EU AI Act’s Impact on Medtech Innovation

Experts at the 2025 European Digital Technology and Software conference expressed concerns that the EU AI Act could hinder the launch of new medtech products in the European market. They emphasized...

Ethical AI: Transforming Compliance into Innovation

Enterprises are racing to innovate with artificial intelligence, often without the proper compliance measures in place. By embedding privacy and ethics into the development lifecycle, organizations...

AI Hiring Compliance Risks Uncovered

Artificial intelligence is reshaping recruitment, with the percentage of HR leaders using generative AI increasing from 19% to 61% between 2023 and 2025. However, this efficiency comes with legal...