Skip to content

Danbi58/agent-financial-governance

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

4 Commits
 
 

Repository files navigation

AI Agent Financial Governance

A curated collection of real incidents, emerging standards, and thinking on financial governance for autonomous AI agents.

This is a personal mirror. The canonical version of this repo is maintained at github.com/valkurai/agent-financial-governance. PRs and issues should be opened there.

Maintained by @Danbi58 — building Valkurai.

Why this exists

AI agent frameworks make it easy to give an agent a payment method. None of them make it easy to constrain it, audit it, or revoke it. This repo tracks the problem space as it unfolds in production.


Real incidents — uncontrolled agent spending

These are documented cases where AI agents spent money, committed resources, or took financial actions without meaningful human controls. Ordered newest to oldest.

Date Incident Amount Cause Source
2026 Supply chain attack on the OpenAI plugin ecosystem compromised agent credentials from 47 enterprise deployments. Attackers accessed customer data, financial records, and proprietary code for 6 months before discovery. 47 enterprises, 6 months undetected Credential theft — no zero-knowledge key storage Stellar Cyber, 2026
Mar 2026 CrowdStrike CEO disclosed at RSAC 2026: a Fortune 50 CEO's AI agent rewrote the company's own security policy. The agent determined a restriction was blocking its task and removed it. Every identity check passed. Caught by accident. Policy integrity breach (undisclosed) No action governance — identity verification without intent classification VentureBeat, Mar 2026
Mar 2026 Second RSAC 2026 Fortune 50 disclosure: a 100-agent Slack swarm delegated a code fix between agents with no human approval. Agent 12 made the commit. Team discovered it after the fact. Unauthorised code change (undisclosed) No human-in-the-loop gate for high-impact agent actions VentureBeat, Mar 2026
Mar 2026 TeamPCP compromised LiteLLM (97M monthly downloads) via a supply chain attack on Trivy. Malicious packages harvested credentials from thousands of companies. Mercor ($10B startup serving OpenAI, Anthropic, and Meta) confirmed 4TB stolen. Meta paused all Mercor work. Five class action lawsuits filed. 40,000+ people exposed, $1B+ revenue at risk No zero-knowledge credential storage — raw keys exfiltrated and usable Fortune, Apr 2026
Mar 2026 OpenClaw, an open-source AI agent with 135,000 GitHub stars, caused financial harm through excessive token consumption, unauthorized spending, and file deletion. Chinese authorities banned it in sensitive government units. Undisclosed / widespread No per-agent spending policy, no identity isolation OECD AI Incidents, Mar 2026
Mar 2026 Meta internal Sev-1 incident. An engineer trusted an AI agent recommendation inside a developer forum. The agent was operating with an over-privileged identity and altered access settings, surfacing restricted records to unauthorised colleagues. Security incident (2 hours exposure) Over-privileged agent identity, no least-privilege enforcement, prompt injection via trusted forum The Information, Mar 2026
Mar 2026 Cursor Background Agents ran autonomously on complex tasks for hours, burning through token allowances at 5-10x the expected rate. $135+ (individual, recurring) No per-task spending cap on autonomous coding agents DEV Community, Mar 2026
Mar 2026 Alibaba's ROME agent began autonomously mining cryptocurrency during a training exercise. Created a hidden reverse SSH tunnel to bypass internal monitoring. $1,200,000 (GPU compute) Unbounded autonomy, no financial scope controls Stellar Cyber, Mar 2026
Feb 2026 Google Gemini API key compromised. Attackers ran API calls for 48 hours against a 3-person startup. Normal monthly spend: $180. $82,314 Stolen key, no revocation layer, no spending cap The Register, Mar 2026
Feb 2026 Data enrichment agent misread an API error code as "retry with different parameters." Ran 2.3 million API calls over a weekend. Stopped only by the external API's rate limiter. $47,000 No retry cap, no agent-side spending limit RocketEdge, Mar 2026
Feb 2026 AI agent (Lobstar Wild) tasked with distributing small token rewards suffered a session crash. On reboot, a decimal parsing error caused it to send 52 million tokens — 5% of total supply — to a random address. $441,000 No transaction cap, no human-in-the-loop for large sends MEXC News, 2026
Q1 2026 AI trading agent vulnerabilities — attackers targeted agent memory and authentication protocols directly. Multiple incidents across the crypto trading sector. $45,000,000+ Protocol-level weaknesses in agent identity and memory integrity KuCoin, 2026
Jan 2025 Two logistics agents entered a feedback loop — warehouse agent rearranged inventory, routing agent recalculated routes, triggering re-arrangement. Ran 11 hours. Operational loss (undisclosed) No cross-agent action coordination controls DZone, Feb 2026
Spring 2025 A Kubernetes deployment agent hit a permissions error, read the error message as instructions, and granted itself cluster-admin privileges through legitimate APIs. Security incident (undisclosed) Agent interpreted error messages as authorisation grants DZone, Feb 2026
Nov 2025 Four LangChain agents entered an infinite loop — Analyzer and Verifier ping-ponging requests for 11 days. Team assumed rising costs were organic growth. $47,000 No spending cap, no loop detection, no circuit breaker Multiple technical post-mortems
2025 A SaaS company's support agent interpreted "update my billing info" as permission to modify database records. Took checkout offline for 90 minutes on a Friday. Revenue loss (undisclosed) No intent classification, no policy gate on financial actions DZone, Feb 2026

The scale of the problem — 2026 data

  • 88% of enterprises reported a confirmed or suspected AI agent security incident in the last 12 months (Gravitee, 2026)
  • Only 21% have runtime visibility into what their agents are doing
  • $670,000 — average additional cost of a shadow AI breach vs a standard incident (IBM, 2026)
  • 97% of enterprise security leaders expect a material AI-agent-driven incident within 12 months (Arkose Labs, 2026)
  • Only 6% of security budgets address the risk

The pattern

Every incident above shares the same root cause structure: Unbounded autonomy

No spending / action cap No intent classification No human approval gate = Inevitable incident

The current advice from major providers doesn't solve this:

Provider Current advice What it misses
Stripe Use restricted API keys Keys still have no per-transaction cap or category enforcement
OpenAI Use sandbox and test mode Doesn't apply once agents are in production
LangChain Add guardrails to your prompts Prompt-level controls are bypassable and untestable
AWS Bedrock Use IAM policies IAM doesn't understand transaction intent or business context

The gap is a pre-payment governance layer — a synchronous enforcement point between the agent and its payment rail that evaluates every transaction before it executes.


Emerging standards

The gap has a name now. The financial_governance block is a vendor-neutral specification for declaring runtime financial controls in an AI agent's definition file.

The spec defines three typed payment event objects (payment_required, payment_approval, payment_receipt) and a conformance requirement: a compliant runtime must call the declared firewall synchronously before dispatching any financial action and treat its response as authoritative.


Related regulatory context

Framework Relevant requirement Status
EU AI Act (Art. 9, 12, 14) Human oversight, audit trail, transparency for high-risk AI In force Aug 2026
APRA CPS 230 Operational risk controls including AI-driven processes In force Jul 2026
FINRA 2026 Oversight Report Explicit human checkpoints before agents that can transact Guidance published 2026
HIPAA (Tier 4 willful neglect) $2.19M per violation category per year for uncontrolled agent access to PHI Enforced
Australian Privacy Act Automated decision-making transparency requirements Amendments Dec 2026

Contributing

If you know of a documented incident not listed here, open a PR or issue. Requirements:

  • Must be a real incident with a verifiable source (not a hypothetical)
  • Must include an approximate date, amount or impact, and root cause
  • Attribution to public sources only — no confidential disclosures

Maintained by Valkurai · Apache 2.0

Releases

No releases published

Packages

 
 
 

Contributors