AI-Driven Cybersecurity: Preparing for Intelligent Threats

When AI Hacks Back: Defending Against the Next Wave of Intelligent Threats

Cybersecurity has traditionally evolved through steady, deliberate steps. Organizations invested in firewalls, endpoint protection, Security Information and Event Management (SIEM), and compliance frameworks to safeguard digital assets. For years, the battle between attackers and defenders remained largely symmetrical: exploit and patch, breach and respond, adapt and counter. However, AI and machine learning (ML) are shifting that balance. Today, anomaly detection, behavioral analytics, and real-time threat scoring are essential tools, accelerating both response and resilience.

But with the rise of generative AI (Gen AI) and with the rapid advances in agentic AI, the playing field is permanently altered. The leading Large Language Models (LLMs) from both proprietary and open-source communities have brought contextual, human-like reasoning into the mainstream. In just over a year, millions—from developers and marketers to cybercriminals—have harnessed Gen AI. This democratization of AI has unleashed a dual-use dilemma: the same technology that drives productivity also arms cyber attackers.

Agentic AI compounds this risk. These AI agents are not just generating content; they are pursuing goals, making decisions, executing tasks across APIs, and interacting with systems and environments autonomously. This new breed of AI has the potential to automate attack chains—scanning for vulnerabilities, crafting phishing vectors, exfiltrating data, and even adapting defensive responses in real time.

The Evolving Threat: From Gen AI Misuse to Malicious Agents

Between 2022 and 2023, cybercriminals began using Gen AI to create polymorphic malware, a code that continuously alters to evade signature-based detection. Around the same time, a new malicious model, WormGPT, emerged on hacker forums. Trained on malware repositories and designed without the ethical safeguards present in LLMs, WormGPT enabled automated, context-aware phishing campaigns.

Threat actors even use Gen AI for impersonations. A notable incident took place in Hong Kong, where scammers used deepfake technology to replicate the likeness and voice of a company’s CFO during a video call. The deception was so convincing that it led a finance employee to transfer over $25 million to the fraudsters.

Compounding the risk is the internal use of Gen AI within enterprises. This has intensified the issue of “Shadow AI,” which often leads to data exposure due to the unsanctioned use of Gen AI tools by employees without the explicit knowledge, approval, or oversight of the organization’s IT or security teams. In 2023, a global electronics firm’s engineers inadvertently leaked confidential source code and meeting transcripts into a public Gen AI tool while debugging and summarizing notes. This incident prompted the company to ban such tools internally and invest in building secure, company-specific LLMs.

Such incidents are no longer rare. Attackers now use AI to build malicious payloads, develop adaptive ransomware, and exploit LLMs through jailbreaking, prompt injection, and prompt hijacking.

Attacks have progressed from prompt-level attacks to workflow injection. Agentic AI agents can now crawl systems, find vulnerabilities, trigger code execution, and adapt to access levels—turning a single prompt vulnerability into an orchestrated attack path.

Rethinking Cyber Defense: AI as a Force Multiplier

These challenges are forcing organizations to rethink not just their tools but their security philosophy. Traditional perimeter defenses and signature-based detection models are inadequate against dynamic, contextual, and increasingly automated attacks. In this environment, security teams must act on multiple fronts.

  • Turning Gen AI and agentic AI into cyber defense allies: Defenders must embrace AI not just as a productivity tool, but as a force multiplier in cyber defense. Leading security platform vendors have integrated LLMs to automate log analysis, summarize alerts, reduce false positives, and prioritize incidents based on business risk.
  • Securing the AI life cycle: As AI systems become integral to enterprise workflows, security must be embedded from development to deployment and day-to-day operations. Enterprises developing internal copilots or customer-facing AI apps must treat these models as sensitive infrastructure.
  • Extending governance from prompts to agents: Gen AI governance often focuses on prompt filtering and hallucination risks. Agentic AI governance adds another layer of intent and execution management.
  • Mitigating risk in the AI supply chain: The model supply chain itself is emerging as a new frontier of cyber risk. In early 2024, an open-source platform for sharing AI models, datasets, and applications became a focal point for such risks when over 100 malicious models were discovered on its platform.

Policy Response Catches Up: Global AI Governance Accelerates

Reflecting the growing recognition of AI’s systemic impact, governments worldwide are moving from exploratory guidelines to enforceable mandates. As noted in various reports, over 1,000 AI-related regulations have been enacted across 69 countries, with the US alone introducing 59 in 2024—marking a 2.5x increase over 2023.

These frameworks are becoming increasingly sector-specific and risk-tiered, emphasizing transparency, safety, and accountability. For instance, the EU AI Act, finalized in August 2024, introduces a tiered risk framework and designates general-purpose AI (GPAI) models as posing systemic risk.

MSSPs Scaling Intelligent AI Across Security Functions

Managed Security Service Providers (MSSPs) are evolving rapidly to integrate intelligent automation into their offerings. Some providers have introduced a multi-agent AI system that orchestrates several specialized agents to simulate cyberattacks, detect anomalies, implement defenses, and even conduct business continuity testing.

These use cases signal a transformative shift where Gen AI and agentic AI are no longer experimental but foundational to next-generation cybersecurity services.

Toward a Cyber-Aware, AI-Native Culture

Perhaps the most important transition is not just technological but cultural. Security teams must stop viewing AI as an external problem to defend against and start treating it as an internal capability to master. The most cyber-resilient organizations will be those that build strong internal model governance, enable human-in-the-loop oversight, and train their workforce on prompt engineering, data-driven reasoning, and agent behavior validation.

AI won’t replace cybersecurity professionals. But those who harness it responsibly, at scale, and with ethical guardrails will replace those who don’t.

More Insights

Rethinking AI Innovation: Beyond Competition to Collaboration

The relentless pursuit of artificial intelligence is reshaping our world, challenging our ethics, and redefining what it means to be human. As the pace of AI innovation accelerates without a clear...

Pakistan’s Ambitious National AI Policy: A Path to Innovation and Job Creation

Pakistan has introduced an ambitious National AI Policy aimed at building a $2.7 billion domestic AI market in five years, focusing on innovation, skills, ethical use, and international collaboration...

Implementing Ethical AI Governance for Long-Term Success

This practical guide emphasizes the critical need for ethical governance in AI deployment, detailing actionable steps for organizations to manage ethical risks and integrate ethical principles into...

Transforming Higher Education with AI: Strategies for Success

Artificial intelligence is transforming higher education by enhancing teaching, learning, and operations, providing personalized support for student success and improving institutional resilience. As...

AI Governance for Sustainable Growth in Africa

Artificial Intelligence (AI) is transforming various sectors in Africa, but responsible governance is essential to mitigate risks such as bias and privacy violations. Ghana's newly launched National...

AI Disruption: Preparing for the Workforce Transformation

The AI economic transformation is underway, with companies like IBM and Salesforce laying off employees in favor of automation. As concerns about job losses mount, policymakers must understand public...

Accountability in the Age of AI Workforces

Digital labor is increasingly prevalent in the workplace, yet there are few established rules governing its use. Executives face the challenge of defining operational guidelines and responsibilities...

Anthropic Launches Petri Tool for Automated AI Safety Audits

Anthropic has launched Petri, an open-source AI safety auditing tool that automates the testing of large language models for risky behaviors. The tool aims to enhance collaboration and standardization...

EU AI Act and GDPR: Finding Common Ground

The EU AI Act is increasingly relevant to legal professionals, drawing parallels with the GDPR in areas such as risk management and accountability. Both regulations emphasize transparency and require...