Updated: April 15, 2026

EU AI Act compliance: what engineers need to know before August 2026

By Coverge Team

Most writing about the EU AI Act is aimed at lawyers and policy people. This guide is not. This is for the engineer who has been told "we need to be EU AI Act compliant" and needs to understand what that means in terms of actual systems to build, data to capture, and processes to implement.

The EU AI Act is the first major regulation that specifically governs AI systems. It entered into force on August 1, 2024. The provisions for high-risk AI systems — the ones that affect most production agent pipelines — apply from August 2, 2026. That is roughly 16 months from when you are reading this. If you have not started, you are not too late, but you do not have time to waste.

Search volume for "eu ai act compliance" is at 140 monthly searches with 138% year-over-year growth. The keyword growth tells the story: enterprises are waking up to the deadline and scrambling to understand what they need to do.

This guide, part of our pillar content on AI governance, strips the legal language down to engineering requirements. We will cover which of your AI systems the Act actually applies to, what documentation you need to produce, what technical capabilities you need to build, and how pipeline governance platforms address the compliance burden.

The timeline: what is already in effect and what is coming

The EU AI Act uses a phased rollout. Here is what matters:

Already in effect (since February 2, 2025):

  • Prohibited AI practices are banned. This includes social scoring, real-time biometric surveillance in public spaces (with exceptions), and manipulation techniques.
  • AI literacy obligations. Organizations deploying AI must ensure their staff has sufficient understanding of AI systems.

In effect since August 2, 2025:

  • Governance structure for general-purpose AI (GPAI) models.
  • Obligations for providers of GPAI models, including technical documentation and transparency requirements.

Coming August 2, 2026:

  • Requirements for high-risk AI systems. This is the big one — most of the technical obligations land here.
  • Conformity assessments for high-risk systems.
  • Registration requirements in the EU database.

Coming August 2, 2027:

  • Requirements for high-risk AI systems that are also regulated products (medical devices, vehicles, etc.) covered by specific EU legislation listed in Annex I.

For most engineering teams building AI agent pipelines, August 2026 is the deadline that matters.

Risk classification: does the Act apply to your system?

The Act classifies AI systems into four risk tiers. Your first engineering task is determining where your systems fall.

Unacceptable risk (prohibited)

These are banned outright. If your system does any of these, stop:

  • Social scoring by governments
  • Real-time biometric identification in public spaces (with narrow exceptions for law enforcement)
  • Exploitation of vulnerabilities of specific groups (age, disability)
  • Emotion recognition in workplaces or educational institutions (with limited exceptions)
  • Untargeted scraping for facial recognition databases

Most AI agent teams do not need to worry about this tier, but verify.

High risk

This is where most regulatory obligations apply. An AI system is high-risk if it falls into one of two categories:

Category 1: Safety components of regulated products. If your AI system is a safety component of a product covered by existing EU product safety legislation (medical devices, automotive, aviation, machinery), it is high-risk.

Category 2: Standalone high-risk AI systems in specific domains (Annex III). This is a defined list:

  • Biometrics: remote biometric identification, biometric categorization, emotion recognition
  • Critical infrastructure: safety components of critical digital infrastructure, road traffic, water/gas/heating/electricity supply
  • Education and vocational training: systems that determine access to education, evaluate learning outcomes, or determine the level of education someone will receive
  • Employment: systems used for recruitment, screening, interviewing, evaluating candidates, making promotion/termination decisions, allocating tasks, or monitoring workers
  • Essential services: systems that evaluate credit scores, assess eligibility for public benefits, evaluate insurance risk, or are used in emergency dispatch
  • Law enforcement: systems used for risk assessment, polygraph, evidence evaluation, crime prediction, or profiling
  • Migration, asylum, and border control: systems used for risk assessment, document verification, or application processing
  • Justice and democracy: systems used to assist judicial authorities in researching or interpreting facts and law

If your AI agent pipeline operates in any of these domains and serves EU customers or operates within the EU, it is likely classified as high-risk.

Limited risk

Systems with limited risk have transparency obligations. This primarily means:

  • AI chatbots must disclose that users are interacting with an AI
  • AI-generated content (deepfakes, synthetic text) must be marked as AI-generated
  • Emotion recognition systems must inform the people they are observing

Minimal risk

Everything else. No specific obligations beyond the general AI literacy requirement. Most internal tools, creative assistants, and low-stakes applications fall here.

The critical question for engineers

Ask your legal team: "Are any of our AI systems classified as high-risk under the EU AI Act?" If the answer is yes, or even "maybe," the rest of this guide applies to you. If the answer is "definitely no," you still need to handle the transparency and literacy obligations, but the heavy engineering work does not apply.

If your legal team does not know the answer, that is itself a compliance gap. Classification is the first step.

Technical requirements for high-risk systems

Here is what the Act requires for high-risk AI systems, translated into engineering terms.

Article 9: Risk management system

What the law says: Providers must establish, implement, document, and maintain a risk management system. It must be a continuous iterative process throughout the lifecycle of the AI system.

What you need to build:

  • A documented process for identifying risks associated with your AI pipeline
  • Quantitative measures of those risks (eval metrics, error rates, failure mode frequencies)
  • Risk mitigation measures (guardrails, human oversight, fallback behaviors)
  • Testing to verify that mitigations actually reduce the identified risks
  • Monitoring to detect when risk levels change in production

In practice, this means your eval suite needs to test for known risk scenarios, and your production monitoring needs to track risk indicators over time. Your pipeline governance platform should store the risk assessment alongside the pipeline configuration.

Article 10: Data and data governance

What the law says: Training, validation, and test datasets shall be subject to appropriate data governance practices.

What you need to build:

  • Documentation of the data used to train, fine-tune, or evaluate your agents
  • Data quality checks (statistical properties, gaps, biases)
  • Provenance tracking for training and evaluation data
  • Procedures for dataset updates and version control

For teams using third-party foundation models (GPT-4o, Claude, Gemini), you do not control the training data. But you do control your fine-tuning data, RAG knowledge bases, and eval datasets. Document what you can, and note what you cannot.

Article 11: Technical documentation

What the law says: Technical documentation shall be drawn up before the system is placed on the market or put into service and shall be kept up to date.

What you need to build:

This is a big one. The documentation must include (from Annex IV):

  1. General description: intended purpose, who developed it, version information
  2. Detailed description of elements:
    • How the system interacts with hardware and software
    • The computational resources needed
    • The model architecture and algorithms
    • Design choices including the rationale
    • The data requirements
  3. Monitoring, functioning, and control:
    • How oversight is implemented
    • Human-machine interface design
    • Measures for interpretation of outputs
  4. Risk management information
  5. Changes made through the lifecycle
  6. Performance metrics:
    • Accuracy levels, robustness measures, cybersecurity measures
    • Metrics used for evaluation
    • Test results
  7. Information about training, validation, and test data

In engineering terms: you need a living document that describes your pipeline architecture, model choices, eval metrics, and test results. The proof bundle concept addresses items 5 and 6 directly — every pipeline version produces a bundle with configuration snapshots, eval results, and performance metrics. The documentation is not a one-time deliverable; it must be updated with every significant change to the system.

Article 12: Record-keeping

What the law says: High-risk AI systems shall technically allow for the automatic recording of events (logs) throughout their lifetime.

What you need to build:

  • Automatic logging of every AI system interaction (not manual, not sampled — automatic)
  • Logs must enable monitoring of the system's operation
  • Logs must be kept for a period appropriate to the intended purpose
  • Logs must identify the input data, the relevant reference data, and the decisions/outputs

This is the AI audit trail. Your LLM observability infrastructure needs to capture:

  • Every input to the pipeline
  • Every output from the pipeline
  • Intermediate steps in multi-agent execution
  • Tool calls and their results
  • The pipeline configuration that was active
  • Timestamps and duration
  • Any human oversight decisions

The storage must be immutable (append-only) and retained for an appropriate period. "Appropriate" is not defined in the Act — work with your legal team to determine the right retention period for your use case.

Article 13: Transparency and provision of information

What the law says: High-risk AI systems shall be designed and developed in such a way as to ensure that their operation is sufficiently transparent to enable deployers to interpret the system's output and use it appropriately.

What you need to build:

  • Clear documentation of what the system can and cannot do (its capabilities and limitations)
  • Information about the level of accuracy the system has been tested to achieve
  • Instructions for use that enable the deployer to understand the output
  • Specification of input data requirements and any known situations where the system may not perform well

For multi-agent pipelines, transparency means the deployer (your customer, or your internal team) can understand how the pipeline produces its outputs. Decision lineage — the ability to trace a final output back through every agent step — is the technical implementation of this requirement.

Article 14: Human oversight

What the law says: High-risk AI systems shall be designed and developed in such a way as to allow them to be effectively overseen by natural persons during the period in which the system is in use.

What you need to build:

  • A human-in-the-loop mechanism for high-risk decisions
  • The ability for a human to override or reverse the system's output
  • The ability for a human to decide not to use the system
  • Clear presentation of the system's output to support informed human judgment
  • A "stop button" or equivalent that can interrupt the system's operation

In pipeline engineering terms: human approval gates before deployment — what Coverge calls eval gates — where someone reviews eval results and signs off, human review capabilities in production (flagged interactions can be reviewed and overridden), and the ability to instantly roll back to a previous version if the current version is producing problematic results.

Article 15: Accuracy, robustness, and cybersecurity

What the law says: High-risk AI systems shall be designed and developed in such a way that they achieve an appropriate level of accuracy, robustness, and cybersecurity.

What you need to build:

  • Accuracy metrics for your pipeline, measured and documented
  • Robustness testing (how the system performs on edge cases, adversarial inputs, out-of-distribution data)
  • Cybersecurity measures (protection against prompt injection, data poisoning, model evasion)
  • Resilience to errors, faults, and inconsistencies in input data

This is where your eval suite, guardrails, and security testing converge. Your pre-deploy eval gates test accuracy and robustness. Your runtime guardrails protect against adversarial inputs. Your audit trail provides evidence that these protections are in place and functioning.

Article 17: Quality management system

What the law says: Providers of high-risk AI systems shall put a quality management system in place.

What you need to build:

  • Documented procedures for compliance management
  • Design and development procedures including testing and validation
  • Configuration management and version control
  • Risk management procedures
  • Post-market monitoring system
  • Incident reporting procedures
  • Communication with regulators and competent authorities

This is the meta-requirement — it asks you to have a system for managing all the other requirements. A pipeline governance platform like Coverge provides the infrastructure: version control for pipeline configurations, eval evidence generation, audit trail storage, and change control workflows.

How pipeline governance platforms address compliance

Building all of the above from scratch is possible but expensive. Pipeline governance platforms consolidate the technical requirements into a single infrastructure layer. Here is how the mapping works:

EU AI Act RequirementPipeline Governance Feature
Article 9 (Risk management)Eval suites that test for risk scenarios; production monitoring for risk metrics
Article 10 (Data governance)Dataset versioning; golden dataset management; data provenance tracking
Article 11 (Technical documentation)Pipeline configuration snapshots; auto-generated architecture descriptions; proof bundles
Article 12 (Record-keeping)Immutable audit trail; automatic interaction logging; execution traces
Article 13 (Transparency)Decision lineage; pipeline visualization; output explanations
Article 14 (Human oversight)Human approval gates; review workflows; rollback capability
Article 15 (Accuracy, robustness, cybersecurity)Eval gates; guardrail integration; adversarial testing
Article 17 (Quality management)Version control; change control workflows; monitoring dashboards

Not every team needs a dedicated platform from day one. But the gap between "we have some logging" and "we can demonstrate compliance to an auditor" is significant, and it grows with the complexity of your AI system.

For guidance on building governance infrastructure, see our AI governance engineering guide. For the specific audit trail engineering work, see our AI audit trail guide. And for understanding how compliance platforms differ from traditional compliance tools, see our guide on AI compliance platforms.

Conformity assessment: proving compliance

For high-risk systems, the Act requires a conformity assessment before the system can be placed on the market or put into service. For most AI systems (those not covered by specific EU product legislation), this is a self-assessment — you assess your own compliance against the requirements and document the results.

This self-assessment is not a rubber stamp. You need to produce evidence for each requirement:

  • Risk management: documented risk assessment, mitigation measures, and their effectiveness
  • Data governance: dataset documentation, quality metrics, bias assessment
  • Technical documentation: complete Annex IV documentation
  • Record-keeping: demonstration that your logging system captures all required events
  • Transparency: user documentation, capability descriptions, accuracy disclosures
  • Human oversight: description of oversight mechanisms, demonstration that they function
  • Accuracy and robustness: test results, eval metrics, security testing results

The self-assessment is a point-in-time evaluation, but Article 17 requires ongoing quality management. In practice, this means you perform the initial assessment, then maintain continuous compliance through your pipeline governance infrastructure.

For high-risk systems that are safety components of products covered by specific EU legislation (Annex I), you may need third-party conformity assessment depending on the applicable legislation.

Practical implementation roadmap

Here is a prioritized sequence for engineering teams facing the August 2026 deadline.

Months 1-2: Classification and foundation

  1. Classify your AI systems. Work with legal to determine which systems are high-risk, limited-risk, or minimal-risk. Document the classification rationale.
  2. Implement audit logging. Add structured, immutable logging for every AI system interaction. This is the foundation for Article 12 compliance and supports multiple other requirements.
  3. Start pipeline versioning. Begin tracking pipeline configurations (prompts, models, parameters) as versioned units. Even a simple approach (JSON files in Git) is better than nothing.

Months 3-4: Testing and documentation

  1. Build an eval suite. Create a golden dataset of at least 50 verified test cases. Implement automated eval runs that measure accuracy, safety, and robustness.
  2. Implement pre-deploy eval gates. Require eval runs before any pipeline change deploys to production. Block changes that regress below established thresholds.
  3. Start technical documentation. Begin the Annex IV documentation. Focus on the system description, design rationale, and data documentation first.

Months 5-6: Governance workflows

  1. Implement change control. Require human approval for pipeline changes, with eval evidence as part of the review.
  2. Add human oversight mechanisms. Build review capabilities for production interactions. Implement confidence thresholds that route low-confidence outputs to human review.
  3. Implement guardrails. Add input and output guardrails appropriate to your risk profile.

Months 7-8: Monitoring and continuous compliance

  1. Deploy production monitoring. Evaluate a sample of production traffic continuously. Alert on quality degradation, guardrail trigger rates, and behavioral drift.
  2. Complete technical documentation. Finalize the Annex IV documentation, including test results and performance metrics.
  3. Conduct self-assessment. Perform the conformity assessment, documenting compliance evidence for each requirement.

Months 9+: Maintenance

  1. Maintain continuous compliance. Keep documentation updated with every pipeline change. Monitor production performance. Review and update risk assessments periodically.

This timeline assumes a team that has existing AI systems in production but limited governance infrastructure. If you are starting from scratch with both the AI system and the governance, add time accordingly.

Common misconceptions

"The EU AI Act only applies if we are based in the EU." Wrong. The Act applies to providers who place AI systems on the EU market, regardless of where they are established. If your product has EU users and your AI system makes decisions about them, the Act likely applies.

"We use third-party models, so compliance is the model provider's responsibility." Partially wrong. The model provider has obligations related to the general-purpose AI provisions. But if you build a high-risk AI system using their model, you are the "provider" of the high-risk system and bear the high-risk system obligations. The model provider's compliance does not exempt you from yours.

"We just need to add a disclaimer that says 'AI-generated content.'" That handles the transparency obligation for limited-risk systems. High-risk systems have much more extensive requirements (risk management, documentation, record-keeping, human oversight, testing).

"Compliance is a one-time certification." Wrong. Article 17 requires ongoing quality management. The conformity assessment is a point-in-time event, but compliance is continuous. Your governance infrastructure must support ongoing monitoring, testing, and documentation.

"Small companies are exempt." Not entirely. The Act does include some accommodations for SMEs (reduced conformity assessment fees, regulatory sandboxes), but the technical requirements apply regardless of company size. If you operate a high-risk AI system, you need to comply.

The connection to the EU AI Act official sources

The full text of the EU AI Act is available on EUR-Lex, the official EU law portal. The AI Act Explorer from the Future of Life Institute provides a more navigable version with article-by-article explanations. Both are authoritative sources for the actual legal text.

For the NIST side of things, if you are a US-based company also establishing NIST AI RMF compliance, the NIST AI RMF documentation provides the framework and implementation guidance.

Frequently asked questions

When does the EU AI Act take effect for high-risk AI systems?

The main compliance deadline for high-risk AI systems is August 2, 2026. Prohibited AI practices were already banned as of February 2025, and GPAI model obligations took effect in August 2025. High-risk systems covered by specific EU product legislation (Annex I) have a later deadline of August 2027.

How do I know if my AI system is high-risk under the EU AI Act?

An AI system is high-risk if it is a safety component of a regulated product (medical devices, automotive, aviation, etc.) or if it operates in a domain listed in Annex III of the Act. Annex III domains include biometrics, critical infrastructure, education, employment, essential services (credit scoring, insurance, public benefits), law enforcement, migration, and justice. Work with your legal team to classify your systems.

What documentation does the EU AI Act require for high-risk systems?

Annex IV of the Act requires technical documentation covering: general system description, detailed technical description (architecture, algorithms, data requirements), design rationale, monitoring and control mechanisms, risk management information, lifecycle change history, performance metrics with test results, and information about training and evaluation data. This documentation must be created before the system is deployed and kept up to date throughout its lifecycle.

Does the EU AI Act apply to companies outside the EU?

Yes. The Act applies to any provider that places an AI system on the EU market, regardless of where they are established. It also applies to deployers located within the EU and to providers and deployers located outside the EU if the output of their AI system is used in the EU. If your product serves EU customers and your AI makes decisions about them, the Act likely applies.

How does pipeline governance help with EU AI Act compliance?

Pipeline governance platforms provide the technical infrastructure that maps to the Act's requirements: version control for pipeline configurations (Articles 11, 17), immutable audit logging (Article 12), eval evidence and proof bundles (Articles 15, 17), human approval workflows (Article 14), and production monitoring (Article 9). They turn abstract regulatory requirements into concrete engineering capabilities.

What is the penalty for non-compliance with the EU AI Act?

Fines are tiered by violation severity. Prohibited AI practice violations: up to 35 million euros or 7% of global annual turnover, whichever is higher. High-risk system obligation violations: up to 15 million euros or 3% of global turnover. Providing incorrect information to authorities: up to 7.5 million euros or 1% of global turnover. SMEs and startups face proportionally lower caps.

Can I do the conformity assessment myself?

For most high-risk AI systems (those not covered by specific EU product legislation in Annex I), yes — the Act allows self-assessment. You evaluate your own compliance against the requirements and document the results. However, "self-assessment" does not mean casual — you need to produce concrete evidence for each requirement. For high-risk systems that are safety components of regulated products, you may need third-party assessment depending on the applicable product legislation.