Skip to Content
GuidesEU AI Act Compliance

EU AI Act Compliance Guide

The EU AI Act (Regulation 2024/1689) is the world’s first comprehensive legal framework for artificial intelligence. It entered into force on August 1, 2024, with obligations phased in over three years.

This guide explains how Curate-Me’s gateway, governance chain, and managed runners help you meet the EU AI Act’s requirements for deploying and operating AI agents.

Timeline

DateMilestone
August 1, 2024Regulation enters into force
February 2, 2025Prohibited AI practices take effect
August 2, 2025Governance rules and obligations for general-purpose AI models
August 2, 2026High-risk AI system obligations (most relevant for agent operators)
August 2, 2027Full enforcement, including high-risk systems in Annex I

Who Does This Apply To?

If you use LLMs or AI agents in your products or operations, the EU AI Act likely classifies you as a deployer. Deployers have specific obligations under Articles 26-29:

  • Monitor AI system operation and report malfunctions
  • Keep logs generated by the AI system for an appropriate period
  • Ensure human oversight measures are in place
  • Use the AI system according to the provider’s instructions of use
  • Conduct a Data Protection Impact Assessment when required

If you build and distribute AI-powered products, you may also be a provider with additional obligations under Articles 16-25.

Risk Classification

The EU AI Act uses a four-tier risk framework. Your obligations scale with the risk level.

Unacceptable Risk (Article 5 — Prohibited)

AI practices that are banned outright:

  • Social scoring by governments
  • Real-time biometric identification in public spaces (limited exceptions)
  • Exploitation of vulnerable groups
  • Subliminal manipulation techniques

Curate-Me: Does not provide or enable any prohibited AI practices.

High Risk (Articles 6-27, Annex III)

AI systems used in critical areas: biometric identification, critical infrastructure, education, employment, essential services, law enforcement, migration, and justice.

Runner mapping: Full VM Tools profile. These runners have unrestricted system access and may participate in high-risk decision pipelines. The governance chain enforces all Art. 9-15 requirements automatically.

Limited Risk (Article 50)

AI systems with transparency obligations: chatbots must disclose they are AI, deepfake content must be labelled, emotion recognition systems must notify users.

Runner mapping: Web Automation profile. Desktop streaming provides real-time transparency. Audit trail captures all browser interactions.

Minimal Risk (Recital 15)

AI systems posing negligible risk: spam filters, AI-powered games, inventory management.

Runner mapping: Locked profile. Read-only filesystem, no network access, minimal risk surface.

Article-by-Article Mapping

Curate-Me’s compliance engine evaluates 8 EU AI Act articles and maps them to platform features:

Article 9 — Risk Management System

Requirement: Establish, implement, document, and maintain a risk management system throughout the AI system’s lifecycle.

Platform feature: The 5-step governance chain acts as a continuous risk management system. Every gateway request passes through:

  1. Rate limiting — prevents runaway agent loops
  2. Cost estimation — pre-flight cost check against budgets
  3. PII scanning — blocks sensitive data before it reaches providers
  4. Model allowlist — enforces approved models only
  5. HITL gate — routes high-risk operations to human reviewers
# Governance chain is enabled per-org via governance policies curl -X GET https://api.curate-me.ai/gateway/admin/compliance/score \ -H "Authorization: Bearer $JWT_TOKEN"

Article 11 — Technical Documentation

Requirement: Draw up and maintain technical documentation before the AI system is placed on the market.

Platform feature: Immutable audit trail records every gateway request, governance decision, runner lifecycle event, and cost calculation. Events are stored in MongoDB with no automatic deletion.

Article 12 — Record-Keeping

Requirement: Allow automatic recording of events (logs) over the lifetime of the system.

Platform feature: Time-travel debugging and session output recording capture every agent execution step. Enable via feature flags:

  • RUNNER_TIMELINE_DEBUG — step-by-step replay of agent sessions
  • RUNNER_SESSION_OUTPUT — full session I/O recording

Article 13 — Transparency

Requirement: Design AI systems so their operation is sufficiently transparent.

Platform feature: Multiple transparency layers:

  • Desktop streaming — live VNC viewing of agent actions via Guacamole
  • Agent traces — structured traces of every agent decision
  • Request logging — full request/response capture in the audit trail
  • Cost transparency — real-time cost tracking per request

Article 14 — Human Oversight

Requirement: Design AI systems so they can be effectively overseen by natural persons.

Platform feature: The HITL (Human-in-the-Loop) gate provides three configurable approval queues:

GateTriggerUse Case
Cost gateRequest estimated cost exceeds thresholdPrevent expensive runaway operations
Confidence gateModel confidence below thresholdReview uncertain AI decisions
Content gateSensitive content detectedReview outputs before delivery

Human reviewers can approve, reject, or modify requests before execution. All decisions are logged to the audit trail.

Additional oversight features:

  • Runner emergency stop — one-click session termination
  • Model allowlists — control which models each org can use
  • Feature flags — instant platform-wide kill switches

Article 15 — Accuracy, Robustness, and Cybersecurity

Requirement: Achieve appropriate levels of accuracy, robustness, and cybersecurity.

Platform feature:

  • PII scanning — regex-based detection of API keys, passwords, emails, SSNs, credit cards, and other sensitive patterns
  • Model allowlists — restrict usage to tested and approved models
  • Upstream resilience — retry logic with exponential backoff for provider failures
  • Organization isolation — strict tenant isolation prevents cross-org data access

Article 26 — Deployer Obligations

Requirement: Use high-risk AI systems according to instructions, with appropriate monitoring and cost governance.

Platform feature:

  • Daily budgets — per-org daily spending limits
  • Per-request cost limits — maximum cost per individual API call
  • Real-time cost tracking — Redis accumulator + MongoDB audit log
  • Usage dashboards — visualize spend by model, provider, and time period

Article 96 — Record-Keeping Obligations

Requirement: Keep automatically generated logs for a minimum of 6 months.

Platform feature: Audit trail records are retained indefinitely by default. No TTL indexes or cleanup jobs delete records before the 6-month minimum. The compliance engine verifies retention policy compliance automatically.

Compliance Scoring

Curate-Me’s compliance engine scores your organization against each article on a 0-100 scale. The overall score is the average across all articles.

# Get compliance score curl -X GET https://api.curate-me.ai/gateway/admin/compliance/score \ -H "Authorization: Bearer $JWT_TOKEN" # Get full report with per-article details curl -X GET https://api.curate-me.ai/gateway/admin/compliance/report \ -H "Authorization: Bearer $JWT_TOKEN" # Apply one-click remediation curl -X POST https://api.curate-me.ai/gateway/admin/compliance/remediate/art_9 \ -H "Authorization: Bearer $JWT_TOKEN"

One-Click Remediation

Several compliance gaps can be fixed instantly:

Fix IDWhat It Does
enable_governance_policyCreates a default policy with rate limits, budgets, and PII scanning
enable_pii_scanningEnables PII scanning with blocking action
enable_hitlSets HITL approval threshold to $5.00 per request
enable_daily_budgetSets daily budget to $50.00 and per-request limit to $1.00
enable_desktop_streamingGuidance to enable the RUNNER_DESKTOP_STREAMING feature flag

Exporting Evidence Packs

For audits and regulatory reviews, export compliance evidence:

# JSON evidence pack (machine-readable) curl -X GET "https://api.curate-me.ai/gateway/admin/compliance/export?format=json" \ -H "Authorization: Bearer $JWT_TOKEN" \ -o compliance-evidence.json # CSV audit export (spreadsheet-friendly) curl -X GET "https://api.curate-me.ai/gateway/admin/compliance/export?format=csv" \ -H "Authorization: Bearer $JWT_TOKEN" \ -o compliance-audit.csv

Evidence packs include:

  • Full compliance report with per-article scores
  • Governance policy configuration snapshot
  • Remediation history with timestamps
  • Audit trail event summary

PII Patterns Detected

The gateway PII scanner detects the following patterns before any data reaches LLM providers:

  • API keys (OpenAI, Anthropic, AWS, Google, etc.)
  • Passwords and bearer tokens
  • Email addresses
  • Social Security Numbers (SSNs)
  • Credit card numbers
  • Phone numbers
  • JWT tokens
  • Private keys (RSA, EC, etc.)

Detected PII is blocked in-memory and never persisted or forwarded to upstream providers.

Dashboard

The Compliance Dashboard in the Curate-Me console provides:

  • Overall compliance score with circular progress indicator
  • Per-article breakdown with status badges (compliant, partial, non-compliant)
  • One-click remediation buttons for quick fixes
  • Evidence export in JSON and CSV formats
  • Comprehensive documentation with risk classification, transparency checklists, human oversight mapping, and data governance details

Access it at https://dashboard.curate-me.ai/compliance.

Further Reading