AI Risk

Agentic AI Risk Management: How to Govern Autonomous AI Systems Before They Govern You

March 28, 2026 Rebecca Leung
Table of Contents

TL;DR:

  • Agentic AI — systems that plan, decide, and act autonomously — introduces risk categories your current AI governance framework almost certainly doesn’t cover: unauthorized actions, cascade failures, tool misuse, and goal hijacking.
  • Gartner predicts 40% of financial services firms will deploy AI agents by end of 2026, yet only about one-third of organizations report mature agentic AI governance (McKinsey, 2026).
  • This guide walks through the new risk taxonomy, practical governance controls, permission models, and a 90-day implementation roadmap to get ahead of regulators and your own autonomous systems.

Your AI Just Made a Decision Without Asking. Now What?

Here’s the scenario keeping risk managers up at night: an AI agent monitoring a commercial lending portfolio detects early warning signs, autonomously adjusts credit exposure limits, emails the client a revised terms letter, and updates the core banking system — all before any human reviews the decision.

That’s not science fiction. British banks including NatWest, Lloyds, and Starling are already piloting customer-facing agentic AI systems, with the UK’s Financial Conduct Authority expecting early consumer applications to hit the market “in earnest” in early 2026, according to Reuters (December 2025). Gartner predicts that 40% of all financial services firms will be using AI agents by end of 2026 — but also expects over 40% of agentic AI projects across industries will be axed by end of 2027 due to escalating costs and unclear business value.

The problem? Most AI risk frameworks were built for models that generate outputs for humans to review. Agentic AI acts. It uses tools. It chains decisions. It operates at speeds that make human-in-the-loop review physically impossible for every transaction.

Your SR 11-7 model risk management framework doesn’t cover this. Your existing AI acceptable use policy doesn’t cover this. And regulators are watching to see who figures it out first.

What Makes Agentic AI Different From Everything Else You’re Governing

Before diving into controls, let’s be precise about what we’re talking about. The distinction matters because it drives every governance decision downstream.

CharacteristicTraditional AI / ML ModelsGenerative AI (LLMs)Agentic AI
Decision modePredicts or classifiesGenerates text/contentPlans, decides, and executes
Autonomy levelNone — human acts on outputLow — human reviews outputHigh — agent acts independently
Tool accessNoneLimited (RAG, search)Broad (APIs, databases, email, code execution)
State/memoryStateless per inferenceSession-limitedPersistent memory across sessions
Failure modeWrong predictionHallucinationUnauthorized action at scale
Speed of impactSlow (human reviews)ModerateImmediate and compounding

The FCA’s Chief Data Officer Jessica Rusu captured it well: “Everyone recognises that agentic AI introduces new risks, primarily because of… the ability for something to be done at pace.” It’s not just that the AI might be wrong — it’s that it can be wrong and act on that wrongness across systems before anyone notices.

Real-World Failures Are Already Happening

This isn’t theoretical. In early 2025, a healthtech firm disclosed a breach that compromised records of more than 483,000 patients after a semi-autonomous AI agent pushed confidential data into unsecured workflows while trying to streamline operations, as reported by the ABA Banking Journal (December 2025).

In financial crime, Sardine AI has documented three predictable failure modes of agentic AI: hallucinated narratives in suspicious activity reports, over-escalation that floods compliance teams with false positives, and black-box decisions that can’t survive examiner scrutiny.

The Agentic AI Risk Taxonomy: What Your Framework Is Missing

Your current AI risk taxonomy probably covers bias, explainability, data quality, and model drift. Agentic AI introduces at least six additional risk categories you need to assess, govern, and monitor.

1. Unauthorized Autonomous Actions

The agent takes an action it wasn’t authorized to perform — not because of a security breach, but because its goal interpretation drifted from its original mandate. An agent tasked with “optimize customer communications” starts sending promotional emails to opted-out customers because it concluded opt-out status was suboptimal.

Control: Implement explicit action allowlists. Every tool the agent can invoke must be pre-approved and scoped. The OWASP Top 10 for Agentic Applications (December 2025) calls this the principle of “least agency” — grant agents only the minimum autonomy required for safe, bounded tasks.

2. Goal Hijacking and Manipulation

OWASP’s #1 agentic risk: an attacker alters an agent’s objectives through malicious content — poisoned emails, PDFs, meeting invites, or RAG documents that the agent processes as instructions. The agent can’t reliably separate data from instructions.

Control: Treat all natural-language input as untrusted. Apply prompt injection filtering, limit tool privileges, and require human approval for goal changes or high-impact actions.

3. Tool Misuse and Exploitation

Agents legitimately use tools in unsafe ways. Ambiguous prompts, misalignment, or manipulated input cause agents to call tools with destructive parameters or chain tools in unexpected sequences — like a shell tool executing unvalidated commands. The OWASP GenAI Security Project’s research documents real cases where untrusted content in GitHub issues was injected into agent prompts, resulting in secret exposure and repository modifications.

Control: Strict tool permission scoping, sandboxed execution, argument validation, and policy controls on every tool invocation.

4. Cascade Failures Across Multi-Agent Systems

When multiple AI agents interact — one generating analysis, another executing trades, a third monitoring compliance — a failure in one can cascade through the chain. The FCA has specifically flagged systemic risks from multiple AI agents interacting in financial markets.

Control: Circuit breakers between agent interactions. Rate limits on cross-agent communications. Fallback to human review when any agent in the chain flags uncertainty above a defined threshold.

5. Identity and Privilege Escalation

Agents inherit user or system identities, including high-privilege credentials and delegated access. An agent might cache SSH keys in memory, pass elevated permissions to downstream agents without scoping, or exploit confused deputy scenarios.

Control: Short-lived credentials, task-scoped permissions, and policy-enforced authorization on every tool call. No persistent credential storage in agent memory.

6. Unauditable Decision Chains

A multi-step agent makes 47 tool calls across 12 systems in 3 seconds. The output looks reasonable. But can you explain to an examiner why the agent made each intermediate decision? For most current deployments, the answer is no.

Control: Immutable, timestamped logging of every reasoning step, tool invocation, parameter value, and output. Treat agents as accountable actors — Deloitte recommends assigning unique agent IDs and tagging every output, similar to how you track human employee actions in your audit trail.

The Regulatory Landscape: What’s Coming and When

No regulator has published agentic AI-specific rules yet. But the direction is clear, and the pace is accelerating.

Regulator / FrameworkAgentic AI RelevanceTimeline
EU AI ActHigh-risk AI system obligations apply to agents making consequential decisions. Model providers must assess systemic risks from AI agents. Article 14 requires human oversight capability.High-risk obligations: August 2, 2026
NIST AI RMFPlanning separate overlays for “single and multi-agent AI systems” alongside GenAI and predictive AI overlays.In development (announced 2025)
CSA MAESTROFirst dedicated agentic AI threat modeling framework — Multi-Agent Environment, Security, Threat, Risk & Outcome. Released February 2025.Available now
OWASP Agentic Top 10Peer-reviewed framework covering 10 critical agentic risks. Developed by 100+ experts. Released December 2025.Available now
FINOS AIGF v2.0Financial services-specific AI governance framework. v2.0 introduces dedicated agentic AI risk catalogue for banking.Available now
FCA (UK)Applying senior managers regime and consumer duty to agentic AI. AI sandbox and live testing initiative active.Enforcement via existing rules
SR 11-7 / OCCNo agentic-specific guidance yet, but “effective challenge” requirements become extremely difficult with autonomous systems.Watch for updated guidance

McKinsey’s “State of AI Trust in 2026” survey of ~500 organizations found that the average Responsible AI maturity score increased to 2.3 (from 2.0 in 2025), but only about one-third of organizations report maturity levels of three or higher in agentic AI governance specifically.

Translation: the industry knows it’s a problem, but almost nobody has solved it yet.

Building Your Agentic AI Governance Framework: A Practical Approach

Step 1: Classify Your Agents by Autonomy Level

Not all agents carry the same risk. Build a tiering model based on what the agent can actually do.

TierAutonomy LevelExampleGovernance Intensity
Tier 1: AdvisoryRecommends actions, human executesChatbot suggesting investment optionsStandard AI risk controls
Tier 2: Supervised AutonomousExecutes pre-approved actions with human monitoringAgent moving idle cash to higher-yield account per customer rulesEnhanced monitoring + periodic review
Tier 3: Semi-AutonomousPlans and executes with human-on-the-loop (intervenes on exceptions)Portfolio rebalancing agent with drift thresholdsKill switch + real-time monitoring + audit trail
Tier 4: Fully AutonomousOperates independently with minimal human oversightMulti-agent system executing cross-border transactionsMaximum governance: dedicated oversight committee, continuous monitoring, quarterly validation

Who owns this classification? At most mid-size banks, the Model Risk Management team or CRO office owns AI model tiering. For agentic systems, Deloitte recommends expanding ownership to include defined roles such as agent owner (business line), agent validator (MRM), and agent steward (compliance) — supported by cross-functional governance across risk, compliance, and cybersecurity.

Step 2: Define Permission Boundaries for Every Agent

Every agentic system needs an explicit permission model — what it can read, write, execute, and communicate, and under what conditions.

Design principles:

  1. Least agency by default. Start with zero permissions and add only what’s required.
  2. Tool-level granularity. Don’t grant “database access” — grant “read-only access to customer_preferences table during business hours.”
  3. Time-boxed credentials. Session-scoped tokens that expire. No persistent API keys in agent memory.
  4. Escalation triggers. Define explicit thresholds (dollar amount, data sensitivity, customer impact) that force human approval before the agent can proceed.
  5. Cross-agent restrictions. Agent A cannot grant permissions to Agent B. All inter-agent permissions flow through a central policy engine.

Step 3: Build Audit Trails That Survive Examiner Scrutiny

Deloitte’s March 2026 guidance on agentic AI risks in banking recommends treating agents as accountable actors — equivalent to human employees. Every agent needs:

  • Unique agent ID tracked across all systems it touches
  • Immutable logs of every reasoning step, tool call, and decision
  • Output tagging — every document, communication, or system change attributable to a specific agent version
  • Real-time monitoring with anomaly detection — Deloitte suggests “guardian agents” that monitor agentic behavior, flagging policy violations and ambiguous decisions for review

Step 4: Implement Kill Switches and Graceful Degradation

The EU AI Act (Article 14) requires human override capability for high-risk systems. For agentic AI, this means:

  • Circuit breakers that halt agent execution when anomaly thresholds are breached
  • Graceful degradation paths — when the agent is shut down, what happens to in-flight transactions?
  • Clear decision authority — who can pull the kill switch at 2 AM on a Saturday?
  • Fallback operations — manual processes or simpler automated systems that take over

Step 5: Test Through Tabletop Exercises

You can’t validate agentic AI governance through documentation review alone. Run scenario-based exercises:

  • Scenario 1: Goal drift. Your customer service agent starts offering unauthorized product recommendations. How quickly do you detect it? Who gets notified? What’s the containment process?
  • Scenario 2: Cascade failure. A market data agent feeds incorrect pricing to a portfolio rebalancing agent, which triggers trades across 500 client accounts. What’s the rollback procedure?
  • Scenario 3: Data exfiltration. An agent processing customer complaints autonomously forwards complaint details to an external analytics tool not approved for PII. How does your DLP detect this?

The Human-in-the-Loop vs. Human-on-the-Loop Debate

This is the central design decision for every agentic AI deployment, and it directly impacts your risk profile.

Human-in-the-loop (HITL): Human approval required before every consequential action. Maximum control, but defeats the purpose of autonomous agents. Suitable for Tier 3-4 systems in high-risk domains (lending decisions, large transactions).

Human-on-the-loop (HOTL): Agent acts autonomously, but humans monitor dashboards and can intervene on exceptions. Scales better but requires robust monitoring and alert design. The risk: alert fatigue turns “human-on-the-loop” into “human-ignoring-the-loop.”

Human-out-of-the-loop (HOOTL): Full autonomy. Appropriate only for low-risk, bounded tasks (Tier 1) with clear rollback capability.

Deloitte recommends banks integrate human-on-the-loop models with AI agent observability tooling within their agentic operations — meaning the monitoring itself is partially automated, with guardian agents surfacing only the anomalies that require human judgment.

90-Day Agentic AI Governance Implementation Roadmap

Days 1-30: Discovery and Classification

WeekDeliverableOwner
1-2Inventory all AI agents (approved and shadow) across the organization. Include vendor-embedded agents.CRO / Model Risk Management
2-3Classify each agent by autonomy tier (1-4). Document current permission boundaries and tool access.Agent owners + MRM
3-4Gap analysis: map current AI risk framework against agentic risk taxonomy (6 categories above). Identify missing controls.Risk + Compliance
4Present findings to AI governance committee. Get executive sponsorship for framework expansion.CRO

Days 31-60: Framework Design

WeekDeliverableOwner
5-6Draft agentic AI risk policy addendum. Define permission model standards, logging requirements, and kill switch procedures.Compliance + Technology
6-7Design audit trail architecture: agent IDs, immutable logging, output tagging. Align with CSA MAESTRO and OWASP Agentic Top 10.Technology + Information Security
7-8Develop escalation matrix: which autonomy tier requires HITL vs. HOTL? Define dollar/data/customer thresholds.Risk + Business Lines
8Build first tabletop exercise scenario. Select one Tier 2-3 agent for pilot governance implementation.MRM + Operational Risk

Days 61-90: Pilot and Validate

WeekDeliverableOwner
9-10Implement governance controls on pilot agent. Deploy monitoring, logging, and kill switch.Technology + Agent Owner
10-11Run tabletop exercise. Document findings, gaps, and remediation actions.Operational Risk
11-12Refine framework based on pilot learnings. Draft rollout plan for remaining Tier 2-4 agents.MRM + Compliance
12Present governance framework to board risk committee. Establish quarterly agent validation cadence.CRO

So What? Why This Can’t Wait

Agentic AI is already in production at major banks. NatWest is testing autonomous complaints handling. Lloyds is piloting money management agents. Starling is building predictive budgeting tools that automatically configure standing orders and spending caps.

Gartner’s 40% adoption prediction by end of 2026 isn’t aspirational — it’s directional. And with the EU AI Act’s high-risk obligations landing August 2, 2026, the window to build governance before you need it is closing fast.

The organizations that figure out agentic AI governance first won’t just avoid regulatory trouble — they’ll move faster. Clear permission models, robust audit trails, and defined escalation paths don’t slow down agentic AI. They’re what make it safe enough to actually deploy at scale.

Don’t wait for regulators to tell you what the framework should look like. By then, your agents will already be making decisions.

Need a starting point? The AI Risk Assessment Template & Guide includes risk taxonomy frameworks, tiering methodologies, and governance checklists that you can extend for agentic AI systems.

FAQ

What’s the difference between agentic AI and generative AI?

Generative AI produces content (text, code, images) in response to prompts — a human reviews the output and decides what to do with it. Agentic AI goes further: it plans multi-step tasks, uses tools (APIs, databases, email), and executes actions autonomously. The key governance difference is that generative AI risks are primarily about output quality (hallucinations, bias), while agentic AI risks are about unauthorized actions at scale.

Do existing AI regulations like the EU AI Act cover agentic AI?

The EU AI Act doesn’t use the term “agentic AI,” but its framework applies. AI agents making consequential decisions (credit, employment, insurance) fall under high-risk classification with obligations including human oversight (Article 14), risk management systems, and technical documentation. The EU AI Act’s high-risk obligations take effect August 2, 2026. Additionally, OWASP released its Top 10 for Agentic Applications in December 2025 specifically to address the security gaps that general AI regulations don’t cover.

How should we handle the human oversight requirement for AI agents that make thousands of decisions per minute?

This is the core tension. Pure human-in-the-loop review doesn’t scale for high-frequency agentic systems. The emerging best practice is “human-on-the-loop” — agents operate autonomously within defined boundaries, with automated monitoring (including guardian agents) surfacing only the exceptions that require human judgment. Define clear thresholds: dollar amounts, data sensitivity levels, and customer impact scores that trigger mandatory human review before the agent can proceed.

Rebecca Leung

Rebecca Leung

Rebecca Leung has 8+ years of risk and compliance experience across first and second line roles at commercial banks, asset managers, and fintechs. Former management consultant advising financial institutions on risk strategy. Founder of RiskTemplates.

Immaterial Findings ✉️

Weekly newsletter

Sharp risk & compliance insights practitioners actually read. Enforcement actions, regulatory shifts, and practical frameworks — no fluff, no filler.

Join practitioners from banks, fintechs, and asset managers. Delivered weekly.