Methodology

Open methodology for monitoring high-risk AI systems

Guardian uses published statistical methods, documented thresholds, and regulatory references to support monitoring and auditability for high-risk AI systems under the EU AI Act.

What this methodology is

Guardian's methodology is a documented framework for monitoring the signals that matter around high-risk AI systems in production.

It is designed to help compliance, risk, legal, and AI teams understand how signals are measured, how thresholds are set, and how outputs connect to review and follow-up workflows.

Guardian is not a legal verdict and it does not claim that a single score determines compliance. The methodology supports a defensible monitoring and evidence process around one live system. For how that shows up in software, see the product overview; for the regulatory context, see EU AI Act readiness. The same signal families show up in high-scrutiny workflows such as hiring and HR AI and credit, fraud, and underwriting.

Why open methodology matters

For high-risk AI systems, monitoring outputs need to be explainable. If a signal, alert, or score cannot be traced to a documented method, it is difficult to defend in front of a regulator, auditor, internal governance committee, or legal review.

Guardian's approach makes the logic visible: what is being measured, why it matters, what threshold is applied, and what follow-up it is meant to trigger.

That is what makes the outputs more useful operationally and more credible in review.

What Guardian measures

The table is a guide to how we map common monitoring inputs to the EU AI Act; your counsel sets legal obligations in context.

MetricWhat it showsRegulatory link
Demographic parityFairness across cohortsArticle 10 / Article 14
Equalised oddsError-rate equity across groupsArticle 10
Model driftPerformance change over timeArticle 72
Data qualityInput distribution and anomaly signalsArticle 10
Human oversight actionsReview and intervention recordsArticle 14
Incident frequencyRate and nature of flagged eventsArticle 62
Documentation completenessCoverage of required technical recordsArticle 11

How signals connect to operational review

Guardian maps each monitoring signal to the operational and regulatory context it supports.

When a threshold is crossed, the output should not sit in isolation. It should help teams understand what changed, why it matters, who should review it, and what record should be maintained next.

This does not replace legal interpretation. It helps teams connect measurement to action in a way that is easier to govern and easier to defend. A typical first step is the 4-week Readiness Sprint, then day-to-day use in Guardian against the background described on the EU AI Act page.

What becomes easier with a documented methodology

  • Explaining why a signal or alert was generated
  • Showing which metric, threshold, and reference support an output
  • Making monitoring outputs easier for compliance, legal, and risk teams to review
  • Reducing dependence on opaque black-box scoring
  • Building a monitoring and evidence baseline that can be expanded over time

Academic and regulatory grounding

Guardian's methodology is developed with academic oversight from Dr. OJ Akintande of DTU Compute, bringing statistical rigor to fairness, drift, and model-risk monitoring.

Metrics and threshold logic are grounded in published statistical methods and relevant regulatory frameworks, including the EU AI Act, NIST AI RMF, ISO 42001, and peer-reviewed fairness research.

The goal is not to make legal determinations automatically. It is to make monitoring outputs more explicit, reviewable, and defensible.

Team

Nordic AI Integrity

Thomas Noba

Co-founder & CEO

Nordic AI Integrity ApS. Copenhagen, Denmark.

Joris Cappa

Co-founder & COO

Nordic AI Integrity ApS. Copenhagen, Denmark.

Dr. OJ Akintande

Technical Advisor

DTU Compute (Technical University of Denmark). ML fairness and model risk specialist.

Frequently asked questions

Is Guardian’s compliance score a legal determination?
No. Guardian does not treat a score as a legal verdict. A score is only one monitoring signal among others, used to help teams prioritise review and maintain a defensible evidence record.
Why publish the methodology openly?
Because monitoring outputs are more useful when teams can understand and explain them. Open methodology makes it easier to trace outputs back to documented metrics, thresholds, and references.
How are thresholds set?
Thresholds are based on documented statistical methods and calibrated to the monitoring context. The goal is to make review triggers explicit rather than opaque.
Does methodology replace legal review?
No. The methodology supports monitoring and evidence maintenance. Legal interpretation and compliance determinations still require human review in context.

Put the methodology to work on one system

See also the security policy for how we handle data in production.