← Blog

EU AI Act and Agent Systems: What You Need to Do Before August 2026

Jon Mandraki

The EU AI Act becomes enforceable on August 2, 2026. If your organization runs agent systems in Europe—or serves European customers—this date matters.

This isn't a distant regulatory threat anymore. It's now. If you're operating agents in healthcare, finance, HR, legal, or critical infrastructure, you're looking at high-risk classification. You have four months to build the governance structures regulators will actually look for.

What the EU AI Act Actually Requires for Agents

The regulation itself doesn't explicitly mention "AI agents," but it applies to them anyway. Here's what gets enforced:

High-risk classification comes from the use case, not the technology. A loan-approval agent is high-risk. A customer service agent that only suggests responses isn't. The regulation lists specific sectors: financial services, healthcare, hiring, criminal justice, border control, critical infrastructure.

Conformity assessment means you prove the system works as documented. Not that it's perfect. That it behaves predictably and that you've tested critical failure modes. Documentation isn't bureaucracy here—it's evidence.

Human oversight is mandatory. Not optional. Not "some human eventually reviews logs." It means checkpoints in the decision flow where a human can intervene, audit, or reject. For a high-risk system, "autonomous" doesn't mean unsupervised.

Data governance requires you to show where training data came from, how it was vetted, and whether it contains bias. You need to track this. Regulators will ask.

Technical documentation that actually exists and is current. If you built the agent six months ago and the behavior has drifted, that's a compliance gap. Documentation is also operational evidence: it shows you designed for failure, not just success.

Which Deployments Get Classified as High-Risk

The safe assumption: if an agent's decision affects a human in a material way, it's high-risk.

  • Financial services: Credit decisions, investment recommendations, fraud detection systems that block transactions
  • Healthcare: Diagnosis support, treatment recommendations, triage systems
  • Employment: Hiring, promotion, termination, scheduling decisions that affect people's livelihoods
  • Legal: Case analysis, sentencing recommendations, evidence assessment
  • Critical infrastructure: Energy grids, water systems, transportation networks where agent decisions could cause harm

If your agents make decisions in these domains, assume high-risk classification and build accordingly. Systems that gather data or run analysis but don't trigger final decisions sit in a gray area. Document the boundaries.

The Regulation's Blind Spot: Operational Evidence

Here's what regulators are actually looking for when they audit you: evidence that you built controls into the system, not that you bolted them on after launch.

They want to see:

  • Proof that decision boundaries were defined at design time, not discovered through testing
  • Audit trails that show what data the agent saw when making specific decisions
  • Test results showing how the agent behaves under edge cases
  • Incident response procedures that existed before incidents happened

This is the "operational evidence" requirement. It doesn't mean you need a 500-page compliance manual. It means you need proof that someone thought about failure modes before deployment.

The regulatory language is precise: demonstrating "the compliance of high-risk AI systems with the requirements set out in this Chapter." That "demonstrating" part is the hard part. Post-hoc documentation and "we're monitoring it now" don't count.

What Your Team Needs to Do Right Now (April 2026)

You have 16 weeks. Here's the path.

April: Classify and Assess

Inventory your agent systems. For each one, determine:

  • What domain is it operating in? (Finance, healthcare, hiring, etc.)
  • What decisions does it make? (Advisory vs. autonomous)
  • Who's affected when it makes a mistake? (End users, employees, customers, infrastructure)

If any answer maps to the high-risk list, you're running a high-risk system. Document this. You'll need to show regulators that you classified it deliberately, not discovered it later.

May: Implement Access Controls and Audit Trails

Every agent decision needs a permanent record:

  • Who triggered the action (user, system, scheduled task)
  • What data the agent consumed
  • What constraints were applied (policies, rules, permissions)
  • What decision was made and why
  • Who approved it (if human oversight is required)

Build this into the execution layer. Audit trails added after the fact are evidence that you didn't design for it.

If your agent system can call multiple tools, log which tools it tried to call, which ones succeeded, and which ones it was denied access to. That denial log is crucial—it shows governance was enforced, not just suggested.

June: Set Up Human-in-the-Loop for High-Risk Decisions

Define which decisions require human intervention:

  • Financial transactions above a threshold
  • Healthcare recommendations that contradict patient history
  • Hiring decisions for candidates in protected classes
  • Critical infrastructure changes

Build explicit checkpoints. "A human could review this" isn't the same as "this system requires human review." Make it part of the workflow.

Some teams build separate approval systems. Others integrate human review into the agent execution pipeline. Either approach works, as long as the human review happens before the decision takes effect, not after.

July: Document System Behavior Boundaries

Write down:

  • What the agent is designed to do
  • What it's not designed to do
  • How it behaves at the edges (malformed input, missing data, contradictions)
  • What failure modes you've tested
  • How the system degrades (fail open or fail closed? Escalate to human?)

A design doc is enough. The point: regulators want to see you thought about this before deployment, not during audit.

How Orloj's Governance Model Maps to EU AI Act Requirements

Running agents on Orloj reduces your compliance surface.

Orloj's manifest-based approach gives you documentation as code. Every agent, tool, permission, workflow lives in YAML. Version control is your compliance trail. "How did the system behave on March 15?" Point to the deployed manifest.

Fail-closed enforcement means unauthorized tool calls don't execute silently. They're rejected and logged. That's evidence governance is real.

Audit trails are built in. Every execution is recorded: who initiated it, what tools were attempted, which had permission, which decisions a human approved. That log is what regulators request.

Access controls are declarative. You define roles, permissions, policies in YAML. They're enforced at runtime. If an agent can't call a tool, the execution layer blocks it.

Approval workflows attach to high-risk actions. Financial transfers wait for human approval. Hiring decisions wait for compliance review. That checkpoint is part of the system and auditable.

Orloj isn't a "compliance tool." It's an orchestration plane. But the operational discipline it requires—declarative config, explicit governance, reliable failure, observable execution—maps to what regulators want.

A Practical Timeline

Week 1-2 (Late April): Inventory your agents. Classify by risk. Document the classification.

Week 3-4 (Early May): Audit current logging. Add infrastructure if missing. Extend logs if they don't capture decision context.

Week 5-8 (May-June): Implement access controls. Limit agent tools. Audit and log all denials.

Week 9-12 (June): Wire human approvals into high-risk workflows. Test approval granted and approval denied paths.

Week 13-16 (July): Document behavior boundaries and system design. Include failure modes, edge cases, degradation behavior. This is your compliance artifact.

Week 17 (August 1): Audit everything. Regulators may ask tomorrow. Can you show decision documentation, audit trails, enforced approval workflows?

This assumes your system already works. If you're building from scratch, start classification now.

What Happens if You Don't

Non-compliance penalties are specific. Up to €35 million or 7% of global annual turnover for serious violations. Up to €15 million or 3% for high-risk violations. These fines matter.

Regulators are already auditing AI deployments. Organizations with governance documentation get lighter scrutiny. Those with design-time controls, not retrofits, have clearer stories.

You don't need to be perfect. You need to be intentional. You need to be able to explain how the system works, why it works that way, and what you've tested. That's what operational evidence means.

August 2, 2026 is the date the regulation becomes enforceable. But audits don't start the day after enforcement begins. They start when someone files a complaint, when journalists ask questions, when regulators do sector-wide sweeps. By then, you either have documentation and audit trails, or you don't.

Build for August Now

None of this requires new technology. Audit logging is standard. Human approvals are workflow. Documentation is work. Access controls are practice.

But it all must be in place by August 2. That's four months. Achievable. Don't wait until July.

Start the inventory this week. Classify your systems. Work backward from August. You'll know what needs to happen when.

The regulation stays. Your agents stay. The gap closes August 2. Close it now.

Related posts