Compliance Guide • EU AI Act

EU AI Act & AI Agents.

Enforcement begins August 2, 2026. Autonomous AI agents are explicitly classified as high-risk AI systems. Fines reach €30 million. Here is your complete compliance roadmap.

TL;DR — Key Takeaways

  • High-risk AI enforcement under the EU AI Act begins August 2, 2026. Most deployed autonomous agents qualify as high-risk.
  • Articles 9, 11, 12, and 14 are the four pillars every agent deployment must implement: risk management, documentation, logging, and human oversight.
  • Fines for non-compliance reach €30 million or 6% of global turnover — whichever is higher.
  • SupraWall implements Articles 9, 12, and 14 automatically through its policy engine, audit logs, and approval queues.

The August 2026 Deadline

Urgent — Enforcement in Effect

The EU AI Act's high-risk AI provisions enter full enforcement on August 2, 2026. This is not a soft launch — non-compliant organizations face fines from day one. The European AI Office has confirmed no extension mechanism exists for most operators.

If you are deploying autonomous AI agents that make consequential decisions — in customer service, finance, HR, legal, or any high-stakes domain — you are almost certainly operating a high-risk AI system under Annex III of the Act. The compliance gap must be closed before August 2, 2026.

Is Your AI Agent High-Risk?

Annex III of the EU AI Act lists categories that automatically qualify as high-risk. Any autonomous agent operating within these sectors — or performing these functions — must comply with all high-risk AI obligations.

Biometric Identification

High-Risk

AI systems used for real-time or post-remote biometric identification of natural persons in publicly accessible spaces.

Critical Infrastructure

High-Risk

AI managing or influencing the safety components of energy, water, transport, and digital infrastructure systems.

Education & Vocational Training

High-Risk

AI determining access to educational institutions, assessing learning outcomes, or evaluating exam performance.

Employment & HR

High-Risk

AI used in recruitment, candidate screening, promotion decisions, task allocation, or performance monitoring of workers.

Essential Private Services

High-Risk

AI evaluating creditworthiness, determining insurance premiums, or making decisions about access to housing or utilities.

Law Enforcement

High-Risk

AI assessing individual risk scores, analyzing crime patterns, or evaluating reliability of evidence in criminal proceedings.

Quick Self-Assessment

Does your agent do any of the following? If yes to any item, you are likely operating a high-risk AI system.

Make or influence employment decisions (screening, scoring candidates)

Process credit applications or assess creditworthiness

Influence access to essential services (insurance, housing, healthcare)

Operate in critical infrastructure (energy, water, transport)

Process biometric data for identification purposes

Make decisions that could affect a person's legal rights

The 4 Critical Articles for AI Agents

While the EU AI Act has 113 articles and 13 annexes, four articles contain the core technical obligations for autonomous AI agent deployments. Each maps directly to a technical control you must implement.

Art. 9

Risk Management System

In Force Aug 2026

Requires an ongoing risk management process throughout the AI system's lifecycle — identifying foreseeable risks, evaluating them, and implementing mitigation measures. This is not a one-time audit.

Technical Implementation

SupraWall block-rate dashboards + policy violation analytics constitute a qualifying ongoing risk management system.

Art. 11

Technical Documentation

In Force Aug 2026

High-risk AI providers must maintain comprehensive technical documentation covering system design, training data, performance benchmarks, and risk assessment outcomes.

Technical Implementation

Maintain a model card, architecture diagram, and training data provenance document for each deployed agent.

Art. 12

Record-Keeping & Logging

In Force Aug 2026

Automatic logging sufficient to enable post-hoc investigation of system behavior. Logs must capture the level of accuracy, robustness, and cybersecurity properties of the system.

Technical Implementation

SupraWall generates per-tool-call audit logs with timestamp, tool name, arguments, decision, cost, agent ID, and session ID.

Art. 14

Human Oversight

In Force Aug 2026

High-risk AI must be designed to enable effective oversight by natural persons. Humans must be able to understand, monitor, and intervene in the system's operation.

Technical Implementation

SupraWall's REQUIRE_APPROVAL policy + dashboard approval queue + kill switch API directly implements Article 14.

Article 14: Human Oversight in Practice

Article 14 requires that high-risk AI systems be designed to allow natural persons to effectively oversee the system during its operation. For autonomous agents, this has specific technical implications — it is not enough to claim oversight exists. You must demonstrate it with evidence.

What "Meaningful Human Oversight" Requires Technically

Approval Queues

High-stakes agent actions (emails, payments, deletions, API calls to external systems) must pause and route to a human for explicit approval before execution.

Kill Switch

A mechanism to immediately halt all agent operations must be available to authorized personnel at all times. SupraWall provides this via dashboard and API.

Audit Trail

Every agent action must be logged with sufficient detail that a human reviewer can reconstruct exactly what happened, why, and what the outcome was.

Override Capability

Humans must be able to override any agent decision — not just observe it. Your system must support after-the-fact correction of automated decisions.

Article 12: Logging Requirements

Article 12 mandates that high-risk AI systems automatically generate logs that enable post-hoc monitoring and investigation. For AI agents, this means logging at the tool-call level — not just aggregate metrics. The logs must be retained for the period mandated by applicable law (minimum 6 months under most EU sectoral regulations).

# Article 12-compliant log entry (SupraWall format)

{

"timestamp": "2026-03-19T14:23:01.847Z",

"agent_id": "prod-agent-finance-01",

"session_id": "sess_8f2k9...",

"tool": "payment.initiate",

"args": {"amount": 2500, "currency": "EUR", "recipient": "[REDACTED]"}, // PII scrubbed

"decision": "REQUIRE_APPROVAL",

"policy_matched": "payment_over_1000_eur",

"cost_estimate_usd": 0.003,

"human_approved_by": "user_alice@corp.com",

"human_approved_at": "2026-03-19T14:24:15Z"

}

SupraWall generates this log entry automatically for every tool call. Logs are exportable in JSON and CSV formats for compliance submissions.

Article 9: Risk Management System

Article 9 requires an ongoing risk management system — not a one-time assessment. You must continuously identify, evaluate, and mitigate risks throughout your AI agent's operational lifecycle. This is where SupraWall's block-rate dashboards and policy violation analytics constitute a qualifying risk management system.

Block Rate

% of tool calls denied by policy — your primary risk indicator

Track daily

Approval Rate

% of calls requiring human review — flags over-reliance on automation

Track weekly

Policy Violations

Total denied calls by policy type — identifies risk hot spots

Track monthly

Self-Assessment Checklist

Use this 10-item checklist to assess your current compliance posture. Each item maps to a specific EU AI Act obligation.

1

AI system classified as high-risk or not

Annex III assessment

2

Technical documentation created and maintained

Article 11

3

Risk management system established

Article 9

4

Training data governance documented

Article 10

5

Automatic logging implemented for all tool calls

Article 12

6

Human oversight mechanism deployed

Article 14

7

Kill switch / override capability available

Article 14(4)

8

Conformity assessment completed

Article 43

9

EU Declaration of Conformity prepared

Article 47

10

Compliance evidence exportable on demand

Article 12 + 14

How SupraWall Generates Compliance Evidence

EU AI Act compliance is not just about having controls in place — you need to demonstrate those controls to regulators. SupraWall's compliance exports give you audit-ready evidence packages that map directly to Articles 9, 12, and 14.

Human Oversight Evidence (HOE) Export

Article 14

A structured PDF/JSON report showing every instance where human oversight was invoked, who approved or rejected the action, and the outcome. This is your Article 14 evidence document.

Audit Log Download

Article 12

Full tool-call-level logs in machine-readable format for any time period. Includes all fields required by Article 12: timestamp, tool, args, decision, cost, agent ID, session ID.

Compliance Dashboard

Article 9

Real-time risk metrics including block rates, policy violation trends, and budget utilization — constituting the ongoing risk monitoring required by Article 9.

Frequently Asked Questions

Does the EU AI Act apply to AI agents?

Yes. Autonomous AI agents that make consequential decisions — especially in high-risk sectors — fall under the EU AI Act as high-risk AI systems. The enforcement date is August 2, 2026.

Which EU AI Act articles apply to AI agents?

Articles 9 (risk management), 11 (technical documentation), 12 (record-keeping/logging), and 14 (human oversight) are the four most critical for autonomous AI agent deployments.

What penalties does the EU AI Act impose?

Non-compliance with high-risk AI rules can result in fines up to €30 million or 6% of global annual turnover — whichever is higher. There is no grace period once enforcement begins.

How does SupraWall help with EU AI Act compliance?

SupraWall implements Articles 9, 12, and 14 through its policy engine (risk management), automatic audit logs (record-keeping), and human-in-the-loop approval queues (human oversight). Compliance evidence is exportable on demand.

Compliance Cluster

August 2026 is closer than you think

Start Protecting
Your Agents.

Get EU AI Act Articles 9, 12, and 14 implemented in your AI agent stack before the August 2026 enforcement date.