top of page

Article 9 EU AI Act in Practice: How to Build a Risk Management System for High-Risk AI

  • Joe Simms
  • Apr 15
  • 8 min read

You've run the analysis. Your AI system is classified as High-Risk. Now what?

If you've followed our KI-Verordnung Framework guide, you know the four steps to clarity:

1. Scope Check — Does the AI Act apply? ✓

2. Risk Classification — What's the risk level? ✓

3. Role Determination — Provider, Deployer, or both? ✓

4. Obligations — What must you do? ← You are here

For High-Risk AI systems, one of the most critical obligations is Article 9: the Risk Management System (RMS). In our previous article, we explained what Article 9 requires. In this article, we show you how to actually build it — with a practical example, templates, and step-by-step guidance.

Connection to the KI-VO Framework:
Article 9 is part of Step 4 (Obligations) in the compliance framework. It applies specifically to High-Risk AI systems. If your system falls under Annex III or is a safety component of regulated products, this article is for you.

→ Haven't determined your risk level yet? Start with our article: "The AI Act Framework: 4 Steps to Clarity"

The Scenario: AI-Powered Recruitment Screening

Let's work with a concrete example throughout this article.

Meet "TalentMatch AI"

You're the CTO of a growing HR Tech startup. You've built an AI-powered Applicant Tracking System (ATS) called "TalentMatch AI" that:

  • Screens incoming CVs automatically

  • Ranks candidates based on job requirements

  • Flags "best matches" for recruiter review

  • Suggests interview questions based on candidate profiles

You've integrated OpenAI's API for natural language understanding and built custom ranking algorithms on top.

Why Is This High-Risk?

Under the AI Act, AI systems used in employment, workers management, and access to self-employment are explicitly listed as High-Risk in Annex III.

Specifically, your system falls under:

Annex III, Point 4(a):

"AI systems intended to be used for the recruitment or selection of natural persons, in particular to place targeted job advertisements, to analyse and filter job applications, and to evaluate candidates."

This means:

  • Article 9 (Risk Management System) is mandatory

  • You need documentation, testing, and ongoing monitoring

  • Non-compliance can result in significant penalties

Let's build your RMS step by step.

Step 1: Risk Identification

The first step is identifying what could go wrong — both the obvious risks and the ones you might not have considered.

Known Risks for Recruitment AI

Based on existing research and documented cases, recruitment AI systems face several well-known risks:

Risk Category

Specific Risk

Impact

Discrimination

Gender bias in ranking (e.g., penalising career gaps)

Fundamental rights violation

Discrimination

Age bias through proxy variables (graduation year, technologies listed)

Fundamental rights violation

Discrimination

Ethnic bias from name or location data

Fundamental rights violation

Data Privacy

Excessive data retention

GDPR violation

Data Privacy

Processing special category data (disabilities, health)

GDPR + AI Act violation

Transparency

Candidates don't know AI is screening them

Article 50 violation

Accuracy

False rejections of qualified candidates

Individual harm, business impact

Emerging Risks to Consider

These are risks that may not be obvious initially but can emerge over time:

Emerging Risk

How It Might Occur

Model drift

After 6 months, the model starts favouring certain patterns that weren't in the training data

Adversarial gaming

Candidates learn to "game" the system with keyword stuffing

Feedback loops

If only AI-selected candidates get hired, the model reinforces its own biases

Context shift

System trained for tech roles gets used for executive hiring without re-validation

Using TrustTroiAI for Risk Identification

In TrustTroiAI, Finn (our Knowledge & Situation Assistant) helps you identify risks specific to your system:

  • Describe your AI system's purpose

  • Finn prompts you with relevant risk categories based on your context

  • You confirm, add, or modify identified risks

  • All risks are logged in your Risk Register

Step 2: Risk Assessment & Prioritisation

Not all risks are equal. Article 9 requires you to prioritise based on severity and probability.

The Risk Assessment Matrix

For each identified risk, assess:

Severity: How harmful is the impact if this risk materialises? (Low / Medium / High / Critical)

Probability: How likely is this to occur? (Rare / Possible / Likely / Almost Certain)

Shows severity (y-axis) vs. probability (x-axis) with specific risks from the ATS example placed in quadrants

Example placements:

- Critical/Likely: Gender bias in ranking

- High/Possible: Model drift

- Medium/Likely: Keyword gaming

- High/Rare: Data breach

Prioritisation for TalentMatch AI

Based on our assessment:

Priority

Risk

Severity

Probability

Action

1

Gender/age bias in ranking

Critical

Likely

Immediate mitigation required

2

Lack of transparency to candidates

High

Almost Certain

Address before deployment

3

Model drift over time

High

Possible

Build monitoring from day 1

4

Excessive data retention

Medium

Likely

Implement data lifecycle policy

5

Adversarial gaming

Medium

Possible

Monitor and adjust

The principle: address the most severe and probable risks first.

Step 3: Mitigation Measures

Now we apply the mitigation hierarchy established by Article 9. Remember: elimination through design takes precedence over instructions.

Applying the Hierarchy to TalentMatch AI

Level 1: Eliminate by Design

What we can remove entirely:

  • Remove candidate names, photos, and addresses from initial screening input

  • Remove graduation year (age proxy) from ranking algorithm

  • Don't process special category data at all

Level 2: Technical Safeguards

What we can detect and prevent:

  • Implement bias detection dashboard tracking selection rates by demographic groups

  • Add demographic parity alerts when disparities exceed thresholds

  • Build confidence scores with automatic human escalation for borderline cases

  • Implement model monitoring for drift detection

Level 3: Organisational Measures

Human oversight and processes:

  • Mandatory human review for all rejections (no fully automated decisions with legal effect)

  • Quarterly bias audits by independent team

  • Clear escalation path for candidate complaints

  • Regular model re-validation schedule

Level 4: User Instructions

Guidance for deployers (companies using your ATS):

  • Clear documentation on intended use and limitations

  • Training requirements for HR staff using the system

  • Guidelines on combining AI recommendations with human judgment

  • Instructions for handling candidate inquiries about AI use

Level 5: Residual Risk Documentation

What remains after all measures:

  • Document that some false positives/negatives will occur

  • Justify acceptance based on: (1) state of the art, (2) human oversight in place, (3) continuous monitoring

  • Inform deployers of residual risks in system documentation

Step 4: Documentation

Documentation isn't bureaucracy — it's your evidence of compliance and your protection in case of audits or incidents.

What to Document

Article 9 requires documentation of:

Document

Content

Update Frequency

Risk Register

All identified risks with assessments

Ongoing (as risks emerge)

Mitigation Plan

Measures for each risk, mapped to hierarchy

After each risk assessment

Residual Risk Justification

Why remaining risks are acceptable

After mitigation implementation

Testing Records

Test procedures, metrics, results

Before deployment + after updates

Monitoring Plan

What you'll track post-deployment

Before deployment

Incident Log

Any issues discovered and how they were addressed

Ongoing

Connection to Article 17 (Quality Management System)

Your RMS documentation feeds into your broader Quality Management System. The QMS provides the organisational structure; the RMS provides the risk-specific content.

Think of it as:

  • QMS = How your organisation ensures quality and compliance (processes, responsibilities, audits)

  • RMS = What specific risks exist and how you address them (content)

Step 5: Post-Market Monitoring Plan

Article 9 requires continuous monitoring after deployment. Here's what to set up:

Metrics to Track for TalentMatch AI

Metric

What It Measures

Alert Threshold

Selection rate by gender

Gender bias

>10% disparity

Selection rate by age group

Age bias

>15% disparity

False rejection rate

Accuracy

>5% (validated sample)

Model confidence distribution

Drift indicator

Shift in distribution

Candidate complaints

User experience

Any mention of unfairness

System override rate

Human oversight effectiveness

<20% or >80% (both concerning)

Feedback Loops

Build mechanisms to learn from deployment:

  • Recruiter feedback on AI recommendations (accurate? helpful?)

  • Hiring outcome tracking (did AI-selected candidates succeed?)

  • Candidate feedback surveys

  • Regular comparison of AI rankings vs. human rankings

Emerging Risk Detection

Schedule quarterly reviews to ask:

  • Are there new patterns in the data we didn't anticipate?

  • Have user behaviours changed (gaming, workarounds)?

  • Has the job market shifted in ways that affect our model?

  • Are there new research findings on recruitment AI risks?

Step 6: Testing & Validation

Before deployment and after significant updates, you need to test against predefined metrics.

Testing Protocol for TalentMatch AI

Test Type

What You Test

Metrics

When

Bias testing

Demographic parity

Selection rates across groups

Pre-deployment + quarterly

Accuracy testing

Prediction quality

False positive/negative rates

Pre-deployment + after updates

Robustness testing

Adversarial inputs

Performance on edge cases

Pre-deployment

Real-world validation

Actual hiring outcomes

Correlation with job success

6 months post-deployment

Documentation for Article 60

If you conduct real-world testing (testing with actual candidates before full deployment), you need to comply with Article 60 requirements:

  • Informed consent from test participants

  • Safety measures and human oversight

  • Clear test boundaries and duration

  • Documentation of results and any issues

The TrustTroiAI Article 9 Risk Management System Templates

Building an RMS from scratch is complex. That's why we've created templates specifically for Article 9 compliance.

What's Included

1. Risk Register Template

  • Pre-structured categories for High-Risk AI

  • Fields for severity, probability, and prioritisation

  • Mapping to fundamental rights impacts

2. Mitigation Plan Template

  • Organised by the 5-level hierarchy

  • Links each measure to specific risks

  • Tracks implementation status

3. RMS Documentation Template

  • Complete structure aligned with Article 9 requirements

  • Guidance text for each section

  • Ready for QMS integration

4. Post-Market Monitoring Checklist

  • Metrics to track by system type

  • Alert threshold recommendations

  • Review schedule templates

Shows the template structure with sections visible (Risk Register, Mitigation Plan, etc.)
TrustTroiAI Article 9 RMS Template — Screenshot

How It Works

1. Start with Scope Check: Confirm your system is High-Risk

2. Guided Risk Identification: Finn prompts you with relevant risk categories

3. Fill the Templates: Enter your specific risks, assessments, and measures

4. Generate Documentation: Export a complete DOCX ready for your records

Shows the professional document output with completed sections, table of contents, and TrustTroiAI formatting
Final RMS Documentation Output — DOCX Preview

Finn: Your RMS Assistant

Have questions while filling out the templates?

  • "Is this risk severity assessment correct for my context?"

  • "What technical safeguards are recommended for recruitment AI?"

  • "How should I document this residual risk?"

Finn knows your specific system and provides context-aware answers.

Bruno: Expert Validation

For High-Risk systems, the stakes are high. When you need certainty:

  • Submit your completed RMS for expert review

  • Get written feedback from qualified compliance experts

  • Have documented validation for audits and regulators

Common Mistakes to Avoid

After helping many teams build their RMS, we've seen patterns in what goes wrong:

Mistake 1: Treating It as a One-Time Exercise

Wrong: "We did our risk assessment before launch. Done."

Right:Risk management is continuous. Schedule quarterly reviews at minimum.

Mistake 2: Ignoring Emerging Risks

Wrong: "We documented the known risks. That's what the law requires."

Right: Article 9 explicitly requires you to consider emerging risks. Build detection mechanisms.

Mistake 3: Skipping Documentation

Wrong: "We mitigated the risks. Why do we need to write it down?"

Right: Without documentation, you have no evidence of compliance. Document everything.

Mistake 4: No Real Testing

Wrong: "It works in our test environment. Ship it."

Right: Article 9 requires testing against predefined metrics. Validate with real-world data.

Mistake 5: RMS in Isolation

Wrong: "Our data scientist manages the RMS. It's separate from our QMS."

Right: RMS and QMS must be structurally linked. Integrate them.

Your RMS Checklist

Before you consider your Article 9 Risk Management System complete, verify:

Risk Identification

□ All known risks documented

□ Emerging risks considered

□ Risks mapped to fundamental rights impacts

Risk Assessment

□ Severity and probability assessed for each risk

□ Prioritisation completed

□ Assessment methodology documented

Mitigation

□ Measures applied following the hierarchy

□ Each risk has assigned mitigation measures

□ Residual risks justified and documented

Documentation

□ Risk Register complete

□ Mitigation Plan documented

□ Testing records available

□ Linked to QMS

Monitoring

□ Post-market monitoring plan in place

□ Metrics and thresholds defined

□ Review schedule established

Visual checklist with checkboxes grouped by category

Conclusion: From Obligation to Advantage

Building a Risk Management System for your High-Risk AI isn't just about compliance. It's about:

  • Building trust with customers who want to know their AI tools are safe

  • Protecting your users — the candidates whose careers are affected by your system

  • Reducing liability by documenting your due diligence

  • Future-proofing as AI regulation becomes the global norm

For TalentMatch AI — and for your High-Risk AI system — Article 9 compliance is achievable. With the right framework, templates, and guidance, you can build an RMS that protects people and positions your company as a responsible AI provider.


Ready to Build Your RMS? 
Start with a free Scope Check to confirm your obligations, then access our Article 9 templates.
→ trusttroiai.com/scope-check

Discover the TrustTroiAI Universe:

Meet Troi, Finn, Bruno, and the other characters that guide you through AI compliance.

Source

[1] Regulation (EU) 2024/1689 — Artificial Intelligence Act

Article 9 (Risk Management System)

Article 17 (Quality Management System)

Article 60 (Testing in Real-World Conditions)

Annex III, Point 4(a) (Employment and Workers Management)


[3] The Academic Guide to AI Act Compliance

hal-05365570v1

[4] CEN-CENELEC JTC 21

Joint Technical Committee on Artificial Intelligence

Developing harmonised standards for AI Act compliance

[5] TrustTroiAI

"The AI Act Framework: 4 Steps to Clarity"

Comments


bottom of page