Catalogue of Tools & Metrics for Trustworthy AI

These tools and metrics are designed to help AI actors develop and use trustworthy AI systems and applications that respect human rights and are fair, transparent, explainable, robust, secure and safe.

Type

Origin

Scope

SUBMIT A TOOL

If you have a tool that you think should be featured in the Catalogue of Tools & Metrics for Trustworthy AI, we would love to hear from you!

Submit

TechnicalUploaded on Mar 20, 2026
garak is an open-source LLM vulnerability scanner developed by NVIDIA that probes large language models for security weaknesses including prompt injection, jailbreaks, hallucination, toxicity, data leakage, and misinformation.

TechnicalUploaded on Mar 20, 2026
OpenEnv is a framework for evaluating AI agents against real systems rather than simulations. It provides a standardised way to connect agents to real tools and workflows while preserving the structure needed for consistent and reliable evaluation.

ProceduralUploaded on Feb 16, 2026
The AI Inherent Risk Scale (AIIRS) is a task-based classification instrument that helps organisations assess the inherent risk of generative AI use. It evaluates tasks against three criteria—epistemic dependence, verifiability, and consequences of error—and assigns a LOW, MEDIUM, or HIGH risk rating using a max-dominant model. AIIRS supports proportionate safeguards, accountable oversight, and governance-aligned decision-making without determining whether AI use is permitted in a given context.

Related lifecycle stage(s)

Operate & monitor

ProceduralUploaded on Mar 20, 2026
Judgment Assurance is a decision-governance discipline that reframes human judgment as a governed institutional asset. It provides a structured framework and practical instruments, including the Underwriting Questionnaire (JA-UQ) and Maturity Model (JAMM-PS), to ensure that consequential AI-mediated decisions are reconstructible and defensible. By defining minimum governance controls for human oversight, it closes the "accountability gap," allowing institutions to define, record, own, and guard the reasoning behind consequential AI-supported outcomes.

ProceduralUploaded on Jan 15, 2026
WasItAI is an image-checker designed to detect AI-generated photos.

Objective(s)


TechnicalProceduralUploaded on Mar 20, 2026
The Approved Intelligence Platform (AIP) provides modular, scenario-based testing workflows to evaluate mission-critical AI systems in defence, public safety, and critical civil use cases. It delivers a comprehensive, end-to-end testing environment based on a proprietary AI trust ontology with measurable AI Solutions Quality Indicators (ASQI) for the testing, evaluation, validation and verification of software solutions with different AI modalities.

TechnicalUploaded on Jan 19, 2026
ASQI Engineer is an open-source framework for testing and assuring AI systems. Built for scale and reliability, it uses containerised test packages, automated assessments, and repeatable workflows to make evaluation transparent and robust. With ASQI Engineer, organisations also run ASQIs that they have created themselves, giving teams full control and confidence in AI quality.

TechnicalProceduralUploaded on Mar 20, 2026
The Resaro AI Solutions Quality Index (ASQI) provides a transparent, use-case-specific measure of AI quality — for applications such as customer chat services, object recognition, deepfake detection, or x-ray anomaly identification.

TechnicalUploaded on Nov 20, 2025
openIMIS is a versatile open source software which supports the administration of health financing and social protection schemes.

TechnicalUploaded on Nov 20, 2025
EuroDaT is a European data trustee with a unique data transaction principle that ensures the secure and legally compliant exchange of data between any parties.

Related lifecycle stage(s)

Collect & process data

ProceduralUploaded on Nov 20, 2025
MISSION KI is developing a voluntary quality standard guideline for artificial intelligence (AI) that strengthens the reliability and trustworthiness of AI applications and systems. It sets a voluntary, evidence-based self-assessment framework for AI providers below the EU AI Act’s high-risk threshold. It defines six quality dimensions (data governance, non-discrimination, transparency, human oversight, reliability, AI-specific cybersecurity) and a stepwise procedure: describe the use case, analyse protection needs, rate requirements via a VCIO catalogue, document tests/evidence, validate findings, issue a report, and monitor validity.

Objective(s)

Related lifecycle stage(s)

Operate & monitorVerify & validate

ProceduralUploaded on Nov 20, 2025
MISSION KI's Compliance Monitor is a tool for monitoring compliance with legal frameworks to facilitate interoperability, data flows and benefit sharing.

EducationalUploaded on Nov 7, 2025
As part of the MISSION KI project, the initiative has developed the innovative data set search engine (Daseen), which for the first time enables cross-source searches for data sets.

TechnicalUploaded on Oct 9, 2025
An open-source framework for large language model evaluations. Inspect can be used for a broad range of evaluations that measure coding, agentic tasks, reasoning, knowledge, behavior, and multi-modal understanding.

Related lifecycle stage(s)

Operate & monitorVerify & validate

EducationalMaltaUploaded on Sep 1, 2025<1 day
This is a complete workshop package for the teaching of practical governance tools when using AI in collaborative teams. It covers the theoretical background, the facilitator notes for each phase, and the student workbook.

TechnicalEducationalUploaded on Oct 1, 2025
An AI-powered speech recognition app that adapts to users' unique speech patterns, facilitating communication for individuals with speech impairments.

Related lifecycle stage(s)

Operate & monitor

EducationalUploaded on Aug 27, 2025
Elements of AI is a free online course, offered in Slovakia by AIslovakIA and Comenius University. Created by the University of Helsinki and Reaktor with EU support, it introduces the basics of artificial intelligence through six interactive modules.

Related lifecycle stage(s)

Operate & monitor

TechnicalProceduralUploaded on Aug 14, 2025>1 year
A legally enforceable AI-user interaction framework that verifies informed consent through multimodal methods, protects user intellectual property via blockchain-based tracking, and ensures lifetime authorship rights with legal safeguards against unauthorized use or AI training reuse.

Related lifecycle stage(s)

Plan & design

TechnicalEducationalUploaded on Aug 27, 2025
A robust multi-agent reinforcement learning framework that evaluates and recommends intervention practices for elementary school children.

TechnicalEducationalUploaded on Aug 27, 2025
AI Screener to enable universal early screening for all children.

Objective(s)

Related lifecycle stage(s)

Plan & design

Partnership on AI

Disclaimer: The tools and metrics featured herein are solely those of the originating authors and are not vetted or endorsed by the OECD or its member countries. The Organisation cannot be held responsible for possible issues resulting from the posting of links to third parties' tools and metrics on this catalogue. More on the methodology can be found at https://oecd.ai/catalogue/faq.