Philips’ Insights on the EU AI Act’s Impact on Medical Innovation

Navigating the EU AI Act: Implications for Medical Devices

As artificial intelligence (AI) continues to revolutionize various sectors, the European Union (EU) has implemented the EU AI Act, a regulatory framework aimed at ensuring the safe deployment of AI technologies, particularly in the healthcare sector. This act is critical for manufacturers of AI medical devices, as it categorizes applications into four risk levels, thereby determining the compliance obligations they must meet.

Understanding the EU AI Act

The EU AI Act, which became law last summer, will be fully applicable to high-risk AI systems—including AI applications embedded in medical devices—by August 2027. The act’s primary goal is to safeguard patients while promoting a competitive AI ecosystem across Europe. However, there are concerns that the act might lead to increased certification costs, delays in product launches, and excessive regulatory burdens that could stifle innovation.

Classification of AI Medical Devices

Under the EU AI Act, many AI medical devices are likely to be classified as high-risk. This classification subjects them to stringent compliance obligations, requiring adherence not only to the Medical Devices Regulation (MDR) but also to the AI Act itself. Manufacturers are expected to navigate a complex landscape, as notified bodies will be tasked with assessing products under both regulations, a process that may not be straightforward.

Challenges for Manufacturers

Manufacturers are concerned that the overlapping requirements of the MDR and the AI Act could hinder their ability to innovate. Not all notified bodies currently assessing medical devices may opt to also evaluate AI medical devices, leading to potential compliance bottlenecks.

Preparing for Compliance

Companies in the healthcare technology sector, particularly larger organizations, have begun preparing for the full enforcement of the AI Act. This includes training employees on the new requirements, which is described as a “Herculean task.” The emphasis is on ensuring that trust is built without imposing burdens that could slow the pace of innovation.

Future Outlook

While the EU AI Act is designed to enhance safety and trust in AI technologies, there remains a palpable concern that it could result in a competitive disadvantage for European companies if innovation progresses more rapidly in markets like North America. The hope is that adjustments can be made to the regulations to foster a balanced approach to innovation and compliance.

More Insights

AI Regulations: Comparing the EU’s AI Act with Australia’s Approach

Global companies need to navigate the differing AI regulations in the European Union and Australia, with the EU's AI Act setting stringent requirements based on risk levels, while Australia adopts a...

Quebec’s New AI Guidelines for Higher Education

Quebec has released its AI policy for universities and Cégeps, outlining guidelines for the responsible use of generative AI in higher education. The policy aims to address ethical considerations and...

AI Literacy: The Compliance Imperative for Businesses

As AI adoption accelerates, regulatory expectations are rising, particularly with the EU's AI Act, which mandates that all staff must be AI literate. This article emphasizes the importance of...

Germany’s Approach to Implementing the AI Act

Germany is moving forward with the implementation of the EU AI Act, designating the Federal Network Agency (BNetzA) as the central authority for monitoring compliance and promoting innovation. The...

Global Call for AI Safety Standards by 2026

World leaders and AI pioneers are calling on the United Nations to implement binding global safeguards for artificial intelligence by 2026. This initiative aims to address the growing concerns...

Governance in the Era of AI and Zero Trust

In 2025, AI has transitioned from mere buzz to practical application across various industries, highlighting the urgent need for a robust governance framework aligned with the zero trust economy...

AI Governance Shift: From Regulation to Technical Secretariat

The upcoming governance framework on artificial intelligence in India may introduce a "technical secretariat" to coordinate AI policies across government departments, moving away from the previous...

AI Safety as a Catalyst for Innovation in Global Majority Nations

The commentary discusses the tension between regulating AI for safety and promoting innovation, emphasizing that investments in AI safety and security can foster sustainable development in Global...

ASEAN’s AI Governance: Charting a Distinct Path

ASEAN's approach to AI governance is characterized by a consensus-driven, voluntary, and principles-based framework that allows member states to navigate their unique challenges and capacities...