Philips’ Insights on the EU AI Act’s Impact on Medical Innovation

Navigating the EU AI Act: Implications for Medical Devices

As artificial intelligence (AI) continues to revolutionize various sectors, the European Union (EU) has implemented the EU AI Act, a regulatory framework aimed at ensuring the safe deployment of AI technologies, particularly in the healthcare sector. This act is critical for manufacturers of AI medical devices, as it categorizes applications into four risk levels, thereby determining the compliance obligations they must meet.

Understanding the EU AI Act

The EU AI Act, which became law last summer, will be fully applicable to high-risk AI systems—including AI applications embedded in medical devices—by August 2027. The act’s primary goal is to safeguard patients while promoting a competitive AI ecosystem across Europe. However, there are concerns that the act might lead to increased certification costs, delays in product launches, and excessive regulatory burdens that could stifle innovation.

Classification of AI Medical Devices

Under the EU AI Act, many AI medical devices are likely to be classified as high-risk. This classification subjects them to stringent compliance obligations, requiring adherence not only to the Medical Devices Regulation (MDR) but also to the AI Act itself. Manufacturers are expected to navigate a complex landscape, as notified bodies will be tasked with assessing products under both regulations, a process that may not be straightforward.

Challenges for Manufacturers

Manufacturers are concerned that the overlapping requirements of the MDR and the AI Act could hinder their ability to innovate. Not all notified bodies currently assessing medical devices may opt to also evaluate AI medical devices, leading to potential compliance bottlenecks.

Preparing for Compliance

Companies in the healthcare technology sector, particularly larger organizations, have begun preparing for the full enforcement of the AI Act. This includes training employees on the new requirements, which is described as a “Herculean task.” The emphasis is on ensuring that trust is built without imposing burdens that could slow the pace of innovation.

Future Outlook

While the EU AI Act is designed to enhance safety and trust in AI technologies, there remains a palpable concern that it could result in a competitive disadvantage for European companies if innovation progresses more rapidly in markets like North America. The hope is that adjustments can be made to the regulations to foster a balanced approach to innovation and compliance.

More Insights

The Perils of ‘Good Enough’ AI in Compliance

In today's fast-paced world, the allure of 'good enough' AI in compliance can lead to significant legal risks when speed compromises accuracy. Leaders must ensure that AI tools provide explainable...

European Commission Unveils AI Code of Practice for General-Purpose Models

On July 10, 2025, the European Commission published the final version of the General-Purpose AI Code of Practice, which aims to provide a framework for compliance with certain provisions of the EU AI...

EU Introduces New Code to Streamline AI Compliance

The European Union has introduced a voluntary code of practice to assist companies in complying with the upcoming AI Act, which will regulate AI usage across its member states. This code addresses...

Reforming AI Procurement for Government Accountability

This article discusses the importance of procurement processes in the adoption of AI technologies by local governments, highlighting how loopholes can lead to a lack of oversight. It emphasizes the...

Pillar Security Launches Comprehensive AI Security Framework

Pillar Security has developed an AI security framework called the Secure AI Lifecycle Framework (SAIL), aimed at enhancing the industry's approach to AI security through strategy and governance. The...

Tokio Marine Unveils Comprehensive AI Governance Framework

Tokio Marine Holdings has established a formal AI governance framework to guide its global operations in developing and using artificial intelligence. The policy emphasizes transparency, human...

Shadow AI: The Urgent Need for Governance Solutions

Generative AI (GenAI) is rapidly becoming integral to business operations, often without proper oversight or approval, leading to what is termed as Shadow AI. Companies must establish clear governance...

Fragmented Futures: The Battle for AI Regulation

The article discusses the complexities of regulating artificial intelligence (AI) as various countries adopt different approaches to governance, resulting in a fragmented landscape. It explores how...

Fragmented Futures: The Battle for AI Regulation

The article discusses the complexities of regulating artificial intelligence (AI) as various countries adopt different approaches to governance, resulting in a fragmented landscape. It explores how...