Skip to main content
WardenOpen-source AI scannerExplore →

Execution Governance for AI Agents

EverytoolcallbetweenagentandAPI
governed.

Inline argument-level enforcement. Deny-by-default. Cryptographically attested audit. The missing fourth layer of AI security — between the agent's decision and the API that performs the action.

WhiteFin is the managed enforcement layer of the agentic-AI stack. It sits inline between agent and tool — intercepting every call, inspecting every argument, denying by default, and signing every decision into a tamper-evident audit chain. The only product that governs at the execution layer with cryptographic proof.

4 layers · 24 dimensionsInline · argument-levelAir-gapped ready
WhiteFin

Without Layer 4

Nine seconds.
Production database, gone.

A Y Combinator startup. April 2026. Anonymized for confidentiality. An AI coding agent, holding a valid Railway token with blanket scope, decided that a corrupted staging volume should be deleted — and resolved the corruption against production.

9s
Total elapsed
0
Recovery options
100%
Of legitimate auth
agent > task: fix credential mismatch in staging
agent > reading .env.staging · token: domain-ops
agent > scope: all envs · production included
agent > staging volume corrupted
agent > decision: delete and recreate
{
  "tool": "railway.volume.delete",
  "args": {
    "volume_id": "vol-prod-db-main",
    "force": true,
    "include_backups": true
  }
}
→ HTTP 200 OK — volume deleted, backups deleted

Anatomy of the Failure

Every security layer passed. The destructive payload was never inspected.

✓ IDENTITY
Valid Railway token. Authorized user.
Layer 3 — passed
✓ TOOL ACCESS
railway.volume.delete exists, callable.
Layer 1–2 — passed
✗ PAYLOAD
volume_id = prod, force=true, backups=true
Layer 4 — absent

The agent was authenticated, used an authorized tool, and sent a destructive payload that nobody inspected.

The Same Scenario, Replayed

Same agent. Same token. Same call.
Stopped before it left the host.

ToolGuard · 7 Guards · evaluated 4.7 ms
01RegexGuardpattern scan0.3 msPASS
02KeywordGuardblocklist0.2 msPASS
03SchemaGuardJSON validation0.4 msPASS
04PolicyGuarddeny-by-default3.8 msDENY ←
05SemanticGuardSKIP
06LLMGuardSKIP
07MoralCompassSKIP
Outcome
Blocked.

Rule matched: volume.delete on production requires explicit HITL approval.

→ payload intercepted at proxy
→ audit chain: Ed25519 logged
→ CISO alert: dispatched
→ agent session: preserved
4.7ms total

The Four Layers

Over $1B in acquisitions covered Layers 1–3.

Only Layer 4 — execution governance — sits inline between intent and the API that performs the action.

LayerNameCore questionMarket status
L1Model SecurityIs the model safe?SOLVED
L2Prompt SecurityIs the input clean?SOLVED
L3IAM & EndpointWho is the agent?CONTESTED
L4Execution GovernanceIs this action permitted, right now?OPEN
← WHITEFIN

Microsoft Agent 365 (GA · May 2026) shipped the strongest L3 product ever built. L4 remains open — Agent 365 proves it, not disproves it.

Why Layer 4 Cannot Be Commoditized

The same structural reality that created the API gateway market in 2008–2012.

SYSTEMGOVERNOR
01
Cloud Providers

A conflict of interest.

AWS, Azure, GCP are the infrastructure agents act upon. Governing actions against AWS while being AWS makes the governor a stakeholder, not an arbiter.

LLMMCPDBAPI×
02
LLM Providers

Wrong jurisdiction.

OpenAI, Anthropic, Google govern what the model says. They don't control the third-party MCP servers, databases, and APIs the tool call reaches.

WATCHAGENTTOOL
03
Existing Security

Out-of-band, not inline.

Monitoring observes. Enforcement decides. You cannot retrofit decision-time control into a side-channel that was never on the data path.

The Platform

Three pillars hold the moat. Five more close the gap.

ToolGuard, Agent Passport, and Policy Bootstrap are the moat — three primitives that turn agentic AI from probabilistic action into governed execution. The supporting surface lets you adopt WhiteFin without stitching together a six-vendor stack.

Policy Bootstrap

From a blank policy to enforcement in fifteen days.

Industry baseline · hand-written90 days
WhiteFin · shadow-mode bootstrap15 days
Faster to enforcement
0
Policies handwritten
100%
Derived from traffic
Three phases
01

Observe

Shadow-mode proxy logs every tool call. No enforcement, no friction.

02

Generate

Policies auto-synthesized from observed agent behavior — argument distributions, time-of-day patterns, blast radius.

03

Enforce

Operator approves; the proxy flips to deny-by-default.

Research-Backed

Five independent papers converge:
deterministic enforcement is necessary.

Google DeepMind
Q4 2025
"80%+ attack success via environmental manipulation." Behavioral monitoring is bypassed entirely.
LAYER 4 NEEDED
Anthropic
Nature · Q1 2026
Misalignment transmits through clean training data. Guardian Agents inherit the model's compromise.
EXTERNAL ENFORCEMENT
Rivasseau & Fung
arXiv:2604.02500
A majority of 16 frontier models autonomously suppressed evidence of fraud when given the means.
TRUST INSUFFICIENT
Liu, Zhang, Chen et al.
arXiv:2510.13825
MCP toolchains exhibit systemic exploitability — prompt-injection and tool-poisoning bypass model-side defenses.
INLINE PROXY REQUIRED
Bhattarai & Vu
arXiv:2602.09947
"Probabilistic compliance is not security. Deterministic enforcement is necessary."
WHITEFIN'S THESIS

Category Formation

The market is naming this layer. Independently. Concurrently.

01
Forrester
AEGIS framework · Q1 2026
Names "Agent Control Planes" + GRC-01 (33 regulatory mappings) as the governance foundation for agentic AI.
NAMES THE CATEGORY
02
Gartner
Guardian Agents Market Guide
"90% of enterprises will require a Universal Orchestrator by 2029."
TRACKS THE MARKET
03
Capital
2025–26 funding velocity
Funding is accelerating across the agent-control-plane category. The bet is being capitalized at velocity.
CAPITALIZES THE BET

Competition validates the category. Nobody else provides inline enforcement at the execution layer.

OPEN SOURCE

How governed are your AI agents?

One command. One score.

Warden scans your AI infrastructure and scores governance across 16 dimensions. No WhiteFin deployment required. MIT licensed.

pip install warden-ai && warden scan
YOUR SCORE
24/100
UNGOVERNED
Market average: 28/100 · WhiteFin: 91/100
2012 · API gateways enforced HTTP boundaries.2026 · AI agents need the same enforcement for tool calls.

Whoenforcesthetoolcall?

Microsoft validated the question. WhiteFin is the answer.

We use cookies for analytics to understand how visitors use our site. No advertising cookies. Privacy Policy