Cyberthreats

Agentic AI security: AI agent risks and how to defend at machine speed

24 March 2026 · 9 min read

Agentic AI is changing cybersecurity in ways most organisations are not prepared for. On the surface, it looks familiar. Attack automation has existed for years. What is new is the combination of speed, autonomy, and the ability to act inside the systems organisations depend on most.

Employees are increasingly using AI agents alongside the systems they rely on every day: email, collaboration platforms, internal dashboards, CRMs, and ticketing tools. These agents can read messages, draft replies, trigger tasks, and update records across multiple systems. Actions that previously required a person to review each step can now happen automatically.

Most organisations already operate with connected systems, delegated permissions, automated workflows, and employees experimenting with AI to work faster. Agentic AI increases the speed and scale at which these behaviours play out.

For security teams, the challenge is expanding. It is no longer only about whether employees can recognise and report a suspicious message in time. It also includes whether employees understand how agentic AI operates inside their workflows, what those agents can access or trigger across systems, and how attackers could exploit those same workflows.

TL;DR

This article introduces an adaptive defence loop to help security leaders respond to AI-driven attacks as they evolve.Used well, that loop helps organisations:

  1. Spot new attacks sooner – Identify emerging tactics early through employee reports and threat signals.
  2. Train against real threats – Turn real attacks into simulations so teams practise spotting current attacker behaviour.
  3. Reduce human-driven risk by interventions – Strengthen behaviour where employee actions and workflows create the most exposure.
  4. Prove security is improving – Measure reporting speed and workforce risk across the organisation.

Machine-speed automation is no longer ‘coming soon’

A recent Anthropic report on a cyber-espionage campaign describes AI agents performing roughly 80–90% of tactical operations with very limited human intervention. Agentic AI was used across multiple stages of the attack chain, including reconnaissance, vulnerability analysis, and data processing, rather than just drafting or research. Most organisations, by contrast, are still trying to secure those early use cases.

At the same time, a CrowdStrike report shows how narrow response windows already are, including an average breakout time of 62 minutes, with the fastest recorded attack coming in at under 2 minutes.

Put together, these signals point to a broader shift in risk. Security teams are dealing with attacks that move from initial access to action faster than many review processes can handle, effectively at machine-speed.

As Rob Daly, CTO at SoSafe, explains in his analysis of agent-driven attacks: “Most response models assume there’s a bit of breathing room between signal and consequence. Agent-driven execution reduces that gap.”

In that environment, early employee reporting and centralised triage become critical. When attacks move this quickly, the first useful signal may come from the workforce before security teams see it elsewhere. Employees therefore need to be trained to recognise suspicious AI behaviour in the tools and workflows they use every day. Those early signals can give security teams the minutes needed to interrupt automated actions before they spread across systems.

Running crafty agentic AI attacks is becoming easier

Studies on agentic AI in cybersecurity describe agents supporting reconnaissance, vulnerability discovery, and multi-stage intrusions. Unit 42’s agentic AI attack framework shows how insecure design patterns and unsafe tool integrations can accelerate attack execution across agentic workflows.

What is now possible:

  • Generate phishing emails and send them at a massive scale.
  • Scan networks continuously for exposed services, weak configurations, and leaked credentials.
  • Coordinate attacks across email, SMS, and phone.
  • Sustain multi-step campaigns over weeks.
  • Chain multiple agents together, with one output feeding the next, and more. 

You have seen these tactics before. The difference now is scale. Attack attempts increase, campaigns iterate faster, and successful tactics get reused on your people sooner.

“Scaling a cyber campaign used to mean adding more people behind keyboards. Agentic AI appears to remove much of that grind. Attacks that were tedious to run are becoming cheaper to run well.”

– Rob Daly, CTO, SoSafe

When campaigns iterate this quickly, security teams need mechanisms to turn new attack patterns into immediate defensive action. Employees, in turn, must be prepared to recognise how these attacks evolve so they can spot them as they appear, not weeks later.

When helpful AI agents gain too much authority

Many agentic AI security risks start as workflow improvements. An AI agent begins with a narrow task. It proves useful. A permission gets added to remove a handoff. Then another permission follows because the process runs smoother that way. Over time, the agent gains more authority.

The difference is delegation. A copilot assists a person’s action. A delegated agent, on the other hand, acts on its own within the permissions it has been given. That is where permission sprawl becomes a governance problem.

In most environments, agents inherit permission models originally designed for people. If one of those agents is compromised, the impact depends on how much authority it has and how quickly it can act. The activity may still look legitimate because valid credentials are doing the work.

That is why least privilege for agents cannot be an afterthought. It needs to be built into the system from the start. As Rob Daly puts it, it is “a design decision.”

    In practice, control usually slips in predictable ways:

  • Temporary permissions that become permanent.

  • Execution rights expanding into decision rights.

  • View access turning into send, change, or trigger access.

These changes accumulate as workflows become more automated and agents take on more responsibility.

In practice, this means focusing behavioural reinforcement where authority and risk intersect: around users with delegated access, elevated permissions, or the ability to trigger operational workflows. These roles influence how actions propagate across systems. Strengthening safe habits at those points helps prevent unsafe automation practices or “shadow AI” behaviours from spreading through everyday workflows. 

More decisions, less oversight

People do not make better security decisions when they face a constant stream of operational decisions: alerts to review, approvals to grant, or automated tasks to validate. They struggle even more when they lack visibility into what those actions will do across connected systems, or when it is unclear who has the authority to intervene.Agentic systems intensify these conditions. Agents can generate large volumes of actions or decisions in a short time, leaving teams managing throughput instead of exercising real control. At that point, human review becomes the bottleneck.

“When an agent surfaces 200 decisions an hour, the human becomes the bottleneck. There’s a phase where humans who want to use AI but don’t trust it will suffer.”
– Rob Daly, CTO, SoSafe

This is where reaction-based defence starts to fail. By the time someone reviews the output, the workflow may already have moved forward and the decision may already have taken effect.

A defence model built for slower threats will struggle here. Teams need a defence model that can spot new attack tactics early, help people recognise them quickly, strengthen safer behaviour in the workflows where risk appears, and show whether those efforts are reducing exposure over time.

That is where the four stages of adaptive defence come in.

Adaptive defence loop for agentic AI threats

An effective defence model needs to operate as a continuous loop.

Step 1: Detect

The Detect stage helps security teams spot new attack patterns early enough to act on them while they are still fresh.

“If you see a tactic today, your environment should be harder to exploit tomorrow, not next quarter.”

– Rob Daly, CTO, SoSafe

This often starts with something simple: an employee flags a message that feels off. When those reports are centralised in Threat Inbox, security teams get a live view of the attacks reaching the workforce. That makes it easier to spot emerging social engineering tactics, AI-generated phishing, and multi-channel attack patterns early, while there is still time to investigate, validate, and feed those signals back into defence.

Step 2: Mirror

The Mirror stage keeps security training aligned with how attacks actually look today.

Recreate Attack allows security teams to turn real phishing emails or social engineering attempts into simulations within minutes. Instead of relying on generic templates or outdated scenarios, teams can recreate the exact tactics observed in their environment and use them in awareness programmes almost immediately.

This shortens the gap between detecting a tactic and preparing people for it, so employees learn to recognise the attacks they are most likely to encounter, not the ones attackers used months ago.

Step 3: Intervene

The Intervene stage focuses on the people and workflows where authority is highest.

Employees with delegated access, elevated permissions, or the ability to trigger operational workflows have the greatest influence on how actions propagate across systems. Strengthening behaviour in these roles therefore has the biggest impact on reducing risk.

Part of this means helping teams become familiar with AI tools themselves. People who never use these systems often struggle to recognise what they can do, when their behaviour changes, or when something looks wrong. Building that familiarity helps employees move from basic AI literacy to working confidently alongside agents that assist with everyday tasks.

Targeted simulations and behavioural interventions focus on these high-agency users. They help teams recognise unsafe automation practices, question unusual agent behaviour, and avoid “shadow AI” shortcuts that bypass established controls.

Step 4: Measure

The Measure stage shows whether the defence loop is actually improving resilience.

Security teams track indicators such as reporting speed, mean time to respond (MTTR), behavioural engagement, and workforce risk signals using metrics like the Human Security Index. These metrics show whether employees are acting as early detection sensors and whether response capability improves as new tactics appear.

Together, these stages form an adaptive defence loop: detection reveals new tactics, mirroring turns them into training, interventions strengthen behaviour in the right workflows, and measurement shows whether those changes are reducing risk.

Networking the defence loop

Attackers reuse what works. Defenders need to do the same.

Many organisations still learn from incidents in isolation. Signals arrive slowly, lessons spread slowly, and the same tactics succeed elsewhere.

When reporting, simulation, behavioural intervention, and measurement are connected, a single incident becomes a learning signal. The tactic can be identified, translated into reinforcement and simulation, prioritised for similar users or workflows, and tracked for behavioural change.

When those signals are shared across organisations, each incident strengthens the wider defence network. Over time, the ecosystem as a whole becomes harder to exploit.

Are you still defending at human speed?

Agentic AI changes risk in two directions at once. Attacks become faster, cheaper, and easier to scale, while employees face growing decision volume and hand more control to systems acting on their behalf. Risk no longer sits only at the point of attack. It increasingly forms inside the workflows and permissions that shape everyday work.

The organisations that adapt will not just detect attacks faster. They will build defence loops that learn faster. That is what it takes to stay effective when attacks move faster and authority increasingly shifts to machines.

FAQs

An agentic AI attack uses autonomous AI systems to execute parts of a cyberattack without constant human direction. The AI can perform tasks such as reconnaissance, phishing generation, and campaign coordination across systems. The key difference is execution authority. The AI is not only analysing information or drafting content, it is acting within an attack workflow.

Traditional phishing campaigns rely heavily on manual effort and static templates. Agentic attacks can generate messages, analyse responses, and adapt tactics automatically. This allows campaigns to iterate quickly, personalise at scale, and operate across multiple channels such as email, SMS, and voice.

Because the time between access and impact becomes smaller. AI systems can move through attack stages faster than many review and response processes can handle. Defence models designed for human-speed attackers struggle to keep up.

Only if governance changes. When agents surface large volumes of decisions, review fatigue reduces effective oversight. Controls need to move earlier into permission design and workflow boundaries.

Defence focuses on several areas:

  • Least-privilege permissions for AI agents.
  • Real-time threat detection and reporting.
  • Behavioural awareness and AI-literate workflows.
  • Adaptive defence loops that detect new tactics, mirror them into simulations, reinforce behaviour, and measure resilience over time.

Do you want to stay ahead of the cyber game?

Sign up for our newsletter to receive the latest cyber security articles, events, and resources. No spam, only content that truly matters.

Newsletter visual

Experience our products first-hand

Use our online test environment to see how our platform can help you empower your team to continuously avert cyber threats and keep your organization secure.

The Forrester Wave™ Strong Performer 2024: Human Risk Management Solutions

This page is not available in English yet.

Diese Seite ist noch nicht in Ihrer Sprache verfügbar. Sie können auf Englisch fortfahren oder zur deutschen Startseite zurückkehren.

Cette page n’est pas encore disponible dans votre langue. Vous pouvez continuer en anglais ou revenir à la page d’accueil en français.

Deze pagina is nog niet beschikbaar in uw taal. U kunt doorgaan in het Engels of terugkeren naar de Nederlandse startpagina.

Esta página aún no está disponible en español. Puedes continuar en inglés o volver a la página de inicio en español.

Questa pagina non è ancora disponibile nella tua lingua. Puoi continuare in inglese oppure tornare alla home page in italiano.