North America’s Struggle with AI Compliance Adoption

US Firms Lag Behind EMEA in AI Adoption for Compliance

Recent research indicates that North American firms are significantly trailing their EMEA counterparts in the integration of Artificial Intelligence (AI) within compliance departments. The study reveals that a substantial 56.3% of compliance departments in North America are hesitant to adopt AI technologies, compared to over 70% of firms in Europe, the Middle East, and Africa.

Current AI Utilization in Compliance Workflows

The report highlights that 33.3% of EMEA respondents and 32% of North American respondents are currently employing AI in their compliance workflows. Interestingly, the rest of the world (RoW) showcases a slightly higher adoption rate at 41.2%, indicating a more proactive approach towards AI technology.

Despite some North American firms beginning to shift their perspective on AI, the data illustrates a marked reluctance to embrace these technologies for compliance. Previous surveys indicated that 65.9% of North American firms expressed no intentions of adopting AI, in contrast to 72.7% of EMEA firms who were ready to embrace AI solutions.

Impact of Regulatory Environments

One of the primary factors influencing AI adoption in compliance is the regulatory framework. The EU AI Act aims to provide clear guidelines for the utilization of AI, thus encouraging firms in EMEA to adopt these technologies. Conversely, the US Securities and Exchange Commission (SEC) has taken a ‘technology-neutral’ stance, which may lead North American firms to reconsider their approach to AI integration.

According to insights from industry experts, advancements in AI technology are rapidly changing the compliance landscape. Don McElligott, Vice-President of Compliance Supervision at Global Relay, states, “Recent technological advances in AI are moving the needle quickly when it comes to adoption, although AI is still generally perceived as difficult, expensive, and only marginally effective for risk detection use cases.”

Communication Management and Compliance Risks

Despite the cautious stance on AI, there is significant usage of AI for managing communications and surveillance, driven by the SEC’s ongoing enforcement initiatives. These efforts aim to address off-channel communications and related misconduct, revealing a complex relationship between compliance and technology.

The study also highlights differing opinions on banning communication channels as a compliance solution. While 50.6% of North American firms consider bans effective, only 31.7% of EMEA firms share this view. This trend indicates a shift towards enabling and monitoring all communication channels rather than imposing bans.

In 2025, 39% of firms are likely to allow all communication channels while monitoring them, a significant rise from 17.4% in 2024. Additionally, there is a decline in the belief that bans are effective, with only 48% of respondents supporting this view in 2025, down from 56% in 2023.

Evolving Compliance Challenges

As the compliance landscape evolves, only 29.5% of respondents cite adherence to compliance policies as a major issue, a significant drop from 65.2% in 2024. Meanwhile, 52.4% view social media channels as a substantial compliance risk.

This comprehensive study demonstrates the significant geographical divide in AI adoption, with North American firms displaying considerable hesitancy compared to their EMEA counterparts. As organizations recognize the value and effectiveness of AI solutions, it will be crucial to monitor changes in attitudes and adoption rates in the coming years.

More Insights

AI Regulations: Comparing the EU’s AI Act with Australia’s Approach

Global companies need to navigate the differing AI regulations in the European Union and Australia, with the EU's AI Act setting stringent requirements based on risk levels, while Australia adopts a...

Quebec’s New AI Guidelines for Higher Education

Quebec has released its AI policy for universities and Cégeps, outlining guidelines for the responsible use of generative AI in higher education. The policy aims to address ethical considerations and...

AI Literacy: The Compliance Imperative for Businesses

As AI adoption accelerates, regulatory expectations are rising, particularly with the EU's AI Act, which mandates that all staff must be AI literate. This article emphasizes the importance of...

Germany’s Approach to Implementing the AI Act

Germany is moving forward with the implementation of the EU AI Act, designating the Federal Network Agency (BNetzA) as the central authority for monitoring compliance and promoting innovation. The...

Global Call for AI Safety Standards by 2026

World leaders and AI pioneers are calling on the United Nations to implement binding global safeguards for artificial intelligence by 2026. This initiative aims to address the growing concerns...

Governance in the Era of AI and Zero Trust

In 2025, AI has transitioned from mere buzz to practical application across various industries, highlighting the urgent need for a robust governance framework aligned with the zero trust economy...

AI Governance Shift: From Regulation to Technical Secretariat

The upcoming governance framework on artificial intelligence in India may introduce a "technical secretariat" to coordinate AI policies across government departments, moving away from the previous...

AI Safety as a Catalyst for Innovation in Global Majority Nations

The commentary discusses the tension between regulating AI for safety and promoting innovation, emphasizing that investments in AI safety and security can foster sustainable development in Global...

ASEAN’s AI Governance: Charting a Distinct Path

ASEAN's approach to AI governance is characterized by a consensus-driven, voluntary, and principles-based framework that allows member states to navigate their unique challenges and capacities...