Catalogue of Tools & Metrics for Trustworthy AI

These tools and metrics are designed to help AI actors develop and use trustworthy AI systems and applications that respect human rights and are fair, transparent, explainable, robust, secure and safe.

Type

Fairness

Clear all

Origin

Scope

SUBMIT A TOOL

If you have a tool that you think should be featured in the Catalogue of AI Tools & Metrics, we would love to hear from you!

Submit

TechnicalEducationalUploaded on Oct 1, 2025
An AI-powered speech recognition app that adapts to users' unique speech patterns, facilitating communication for individuals with speech impairments.

Related lifecycle stage(s)

Operate & monitor

EducationalUploaded on Aug 14, 2025
UNICEF’s Accessible Digital Textbooks for All initiative aims to make textbooks accessible for students with disabilities by following Universal Design for Learning (UDL) guidelines and offering flexible delivery formats.

Related lifecycle stage(s)

Deploy

EducationalUploaded on Aug 6, 2025
Efficacy of a Computer-Based Learning Program in Children With Developmental Dyscalculia. What Influences Individual Responsiveness?

EducationalUploaded on Aug 4, 2025
The Early Literacy Interventions Tool uses ChatGPT to answer questions and design learning plans

TechnicalUnited StatesUploaded on Mar 24, 2025
An open-source Python library designed for developers to calculate fairness metrics and assess bias in machine learning models. This library provides a comprehensive set of tools to ensure transparency, accountability, and ethical AI development.

ProceduralUnited KingdomUploaded on Oct 2, 2024
Warden AI provides independent, tech-led AI bias auditing, designed for both HR Tech platforms and enterprises deploying AI solutions in HR. As the adoption of AI in recruitment and HR processes grows, concerns around fairness have intensified. With the advent of regulations such as NYC Local Law 144 and the EU AI Act, organisations are under increasing pressure to demonstrate compliance and fairness.

ProceduralUploaded on Oct 2, 2024
FairNow is an AI governance software tool that simplifies and centralises AI risk management at scale. To build and maintain trust with customers, organisations must conduct thorough risk assessments on their AI models, ensuring compliance, fairness, and security. Risk assessments also ensure organisations know where to prioritise their AI governance efforts, beginning with high-risk models and use cases.

TechnicalInternationalUploaded on Nov 5, 2024
A fast, scalable, and open-source framework for evaluating automated red teaming methods and LLM attacks/defenses. HarmBench has out-of-the-box support for transformers-compatible LLMs, numerous closed-source APIs, and several multimodal models.

Objective(s)


EducationalUnited StatesUploaded on Nov 5, 2024
Community jury is concept where multiple stakeholders impacted by a same technology are given the possibility to learn about a project, discuss with one another and provide feedback.

ProceduralUploaded on Jul 2, 2024
This DIN DKE SPEC defines guidelines for the labelling of training data for QA systems and specifies the characteristics of labels.

ProceduralUploaded on Jul 2, 2024
This standard specifies an architecture and technical requirements for face recognition systems.

Objective(s)


ProceduralUploaded on Jul 2, 2024
This document introduces the effects of population demographics on biometric functions.

ProceduralUploaded on Jul 3, 2024
This document addresses bias in relation to AI systems, especially with regards to AI-aided decision-making.

TechnicalProceduralSpainUploaded on May 21, 2024
LangBiTe is a framework for testing biases in large language models. It includes a library of prompts to test sexism / misogyny, racism, xenophobia, ageism, political bias, lgtbiq+phobia and religious discrimination. Any contributor may add new ethical concerns to assess.

TechnicalProceduralUnited StatesJapanUploaded on Apr 19, 2024
Diagnose bias in LLMs (Large Language Models) from various points of views, allowing users to choose the most appropriate LLM.

Related lifecycle stage(s)

Plan & design

TechnicalChinaUploaded on Apr 2, 2024
A PyTorch implementation of Speech Transformer, an End-to-End ASR with Transformer network on Mandarin Chinese.

Objective(s)


EducationalUploaded on Apr 2, 2024<1 hour
Approaches to disability-centered data, models, systems oversight

ProceduralBrazilUploaded on Mar 14, 2024
Ethical Problem Solving (EPS) is a framework to promote the development of safe and ethical artificial intelligence. EPS is divided into an evaluation stage (performed via Algorithmic Impact Assessment tools) and a recommendation stage (the WHY-SHOULD-HOW method).

TechnicalGermanyUploaded on Dec 15, 2023
A collections of public and free annotated datasets of relationships between entities/nominals (Portuguese and English).

Uploaded on Dec 14, 2023
Our work enables developers and policymakers to anticipate, measure, and address discrimination as language model capabilities and applications continue to expand.

Partnership on AI

Disclaimer: The tools and metrics featured herein are solely those of the originating authors and are not vetted or endorsed by the OECD or its member countries. The Organisation cannot be held responsible for possible issues resulting from the posting of links to third parties' tools and metrics on this catalogue. More on the methodology can be found at https://oecd.ai/catalogue/faq.