Agentic AI: Managing the Risks of Autonomous Systems

Agentic AI Sprawl: Risks and Governance Needs for Businesses

In the rapidly evolving world of artificial intelligence, companies are increasingly turning to agentic AI systems—autonomous programs that can make decisions, execute tasks, and interact with other systems without constant human oversight. However, as these technologies proliferate, a new challenge is emerging: agentic AI sprawl. This phenomenon, akin to the unchecked growth of shadow IT in past decades, threatens to create chaotic environments where dozens or even hundreds of AI agents operate independently, potentially leading to security vulnerabilities, compliance issues, and operational inefficiencies.

Recent reports highlight how businesses, eager to harness the power of agentic AI for tasks like supply chain optimization or customer service automation, are deploying these tools without adequate governance frameworks. For instance, a warning has been issued that companies are “sleepwalking” into this sprawl, much like they did with cloud services, resulting in fragmented systems that are hard to monitor and control.

The Hidden Dangers of Unmanaged AI Agents

This sprawl isn’t just a theoretical risk; it’s already manifesting in real-world scenarios. Vendors are flooding the market with tools to create AI agents for various tasks, but without proper management, organizations could face escalating costs and security breaches. IT leaders are urged to think proactively about agent orchestration to avoid a repeat of past IT sprawls.

Industry analysts are sounding alarms about the broader implications. A prediction states that over 40% of agentic AI projects will be canceled by the end of 2027 due to unclear business value, soaring expenses, and insufficient risk controls. This forecast underscores the urgency for companies to reassess their strategies before investments go awry.

Navigating Risks in Critical Sectors

Beyond financial pitfalls, the risks extend to critical infrastructure. Growing concerns among tech insiders highlight how agentic AI could disrupt sectors such as healthcare and transportation if not properly governed. These AI systems could revolutionize incident response but also introduce unpredictability if left unchecked, potentially turning routine operations into high-stakes gambles.

Security experts are particularly worried about cyber threats. Reports detail how cybersecurity firms are introducing new capabilities to secure agentic workspaces, addressing challenges like data protection and collaboration risks. Without such measures, AI agents could become vectors for attacks, including memory poisoning or unauthorized data access.

Strategic Responses from Leading Firms

Major corporations are beginning to respond. Advanced industries are implementing agentic AI to boost efficiency, but they stress the importance of vertical and horizontal use cases to mitigate sprawl. Organizations must evolve their risk programs as they transition to multi-agentic systems. This involves investing in employee training, monitoring systems, and intervention protocols to safely deploy these technologies.

Emerging Trends and Future Outlook

Looking ahead, emerging trends suggest a shift toward better governance. As agentic AI evolves, systems capable of perceiving and reasoning are becoming prevalent. Business leaders are urged to prioritize trends like interoperability and ethical deployment to curb sprawl.

Despite the revolutionary potential of agentic AI for autonomous tasks like threat detection, it introduces risks such as hijacking and data breaches. Solutions are emerging to provide real-time monitoring, signaling a maturing market response.

Building a Resilient AI Ecosystem

To combat sprawl, experts recommend establishing centralized AI governance teams. Companies must adapt to avoid obsolescence. Disruption is mandatory, but with strategic planning, it’s manageable. As agentic AI becomes integral to business operations, proactive management is key. By learning from past IT mistakes, companies can harness AI’s potential while minimizing risks, ensuring that innovation doesn’t come at the cost of control.

More Insights

US Rejects UN’s Call for Global AI Governance Framework

U.S. officials rejected the establishment of a global AI governance framework at the United Nations General Assembly, despite broad support from many nations, including China. Michael Kratsios of the...

Agentic AI: Managing the Risks of Autonomous Systems

As companies increasingly adopt agentic AI systems for autonomous decision-making, they face the emerging challenge of agentic AI sprawl, which can lead to security vulnerabilities and operational...

AI as a New Opinion Gatekeeper: Addressing Hidden Biases

As large language models (LLMs) become increasingly integrated into sectors like healthcare and finance, a new study highlights the potential for subtle biases in AI systems to distort public...

AI Accountability: A New Era of Regulation and Compliance

The burgeoning world of Artificial Intelligence (AI) is at a critical juncture as regulatory actions signal a new era of accountability and ethical deployment. Recent events highlight the shift...

Choosing Effective AI Governance Tools for Safer Adoption

As generative AI continues to evolve, so do the associated risks, making AI governance tools essential for managing these challenges. This initiative, in collaboration with Tokio Marine Group, aims to...

UN Initiatives for Trustworthy AI Governance

The United Nations is working to influence global policy on artificial intelligence by establishing an expert panel to develop standards for "safe, secure and trustworthy" AI. This initiative aims to...

Data-Driven Governance: Shaping AI Regulation in Singapore

The conversation between Thomas Roehm from SAS and Frankie Phua from United Overseas Bank at the SAS Innovate On Tour in Singapore explores how data-driven regulation can effectively govern rapidly...

Preparing SMEs for EU AI Compliance Challenges

Small and medium-sized enterprises (SMEs) must navigate the complexities of the EU AI Act, which categorizes many AI applications as "high-risk" and imposes strict compliance requirements. To adapt...

Draft Guidance on Reporting Serious Incidents Under the EU AI Act

On September 26, 2025, the European Commission published draft guidance on serious incident reporting requirements for high-risk AI systems under the EU AI Act. Organizations developing or deploying...