Catalogue of Tools & Metrics for Trustworthy AI

These tools and metrics are designed to help AI actors develop and use trustworthy AI systems and applications that respect human rights and are fair, transparent, explainable, robust, secure and safe.

Type

Origin

Scope

SUBMIT A TOOL

If you have a tool that you think should be featured in the Catalogue of AI Tools & Metrics, we would love to hear from you!

SUBMIT

ProceduralUnited KingdomUploaded on Oct 2, 2024
Warden AI provides independent, tech-led AI bias auditing, designed for both HR Tech platforms and enterprises deploying AI solutions in HR. As the adoption of AI in recruitment and HR processes grows, concerns around fairness have intensified. With the advent of regulations such as NYC Local Law 144 and the EU AI Act, organisations are under increasing pressure to demonstrate compliance and fairness.

ProceduralUploaded on Jul 3, 2024
This document addresses bias in relation to AI systems, especially with regards to AI-aided decision-making.

Objective(s)


TechnicalProceduralUnited StatesJapanUploaded on Apr 19, 2024
Diagnose bias in LLMs (Large Language Models) from various points of views, allowing users to choose the most appropriate LLM.

Related lifecycle stage(s)

Plan & design

TechnicalUploaded on Dec 11, 2023
Practical recommendations for mitigating bias in automated speaker recognition, and outline future research directions.

Objective(s)


TechnicalUnited StatesIndiaUploaded on Dec 11, 2023<6 months
Repository of algorithmic bias related metrics and measures to enable researchers and practitioners to leverage their use. The repository is also intended to allow researchers to expand their research to identify more metrics that may be relevant and appropriate to specific context.

Objective(s)

Related lifecycle stage(s)

Verify & validate

TechnicalNetherlandsUploaded on Nov 29, 2023
This bias detection tool identifies potentially unfairly treated groups of similar users by a binary algorithmic classifier. The tool identifies clusters of users that face a higher misclassification rate compared to the rest of the data set. Clustering is an unsupervised ML method, so no data is needed is required on protected attributes of users. The metric by which bias is defined can be manually chosen in advance: False Negative Rate (FNR), False Positive Rate (FPR), or Accuracy (Acc).

Related lifecycle stage(s)

Verify & validate

TechnicalProceduralGermanyUploaded on Sep 7, 2023
Biaslyze is a python package that helps to get started with the analysis of bias within NLP models and offers a concrete entry point for further impact assessments and mitigation measures.

TechnicalUnited StatesUploaded on Jun 8, 2023
[ICML 2021] Towards Understanding and Mitigating Social Biases in Language Models

ProceduralIndiaUploaded on May 23, 2023
The report steers its way to identify the challenges often faced in the process of bias mitigation. The challenges have been categorized into organizational levels, industry-wide levels and societal levels and includes limitations such as lack of domain knowledge and accountability, lack of regulations and actionable guidance, persistence of black box algorithms and outdated education approaches for data and scientists.

Objective(s)


TechnicalUploaded on May 20, 2023
bt4vt is a python library to diagnose performance discrepancies (i.e. bias) in automatic speech processing models.

Objective(s)


TechnicalProceduralUploaded on Apr 4, 2023
This course explains what is meant by bias in the context of machine learning algorithms, how bias can be present in the training data, and how it can be unintentionally introduced in the learning phase. Learners will gain an appreciation of why this is a concern and why it needs to be addressed. It is is divided into two sections: milestones 1-2 address the conceptual background and context, milestones 3-5 explore practical applications.

TechnicalProceduralUploaded on Mar 27, 2023<1 day
A fully independent and impartial audit to ensure compliance with the bias audit requirements for New York City Local Law 144 and upcoming regulations.

TechnicalNetherlandsSwedenUploaded on Mar 27, 2023
Machine learning approach to detect algorithmic bias in all types of binary AI classifiers. No protected user characteristics needed. Identifies unfairly treated groups characterized by mixture of features, detects intersectional bias. Model-agnostic, open-source web application, easy to use for the entire AI auditing community, e.g., journalists, data scientists, policy makers etc. Finalist of Stanford's AI Audit Competition.

TechnicalUploaded on Feb 23, 2022

Open Sourced Bias Testing for Generalized Machine Learning Applications audit-AI is a Python library built on top of pandas and sklearn that implements fairness-aware machine learning algorithms. audit-AI was developed by the Data Science team at pymetrics

Objective(s)


ProceduralUploaded on Feb 23, 2022

HireVue recognizes the impact that their software can have on individuals and on society, and they act upon this responsibility with deep commitment. The following principles guide their thoughts and actions as they develop artificial intelligence (AI) technology and incorporate it into their products and technologies. 1. They are committed to benefiting society 2. They […]


TechnicalProceduralUploaded on Feb 23, 2022

Aequitas is an open source bias and fairness audit toolkit that is an intuitive and easy to use addition to the machine learning workflow, enabling users to seamlessly test models for several bias and fairness metrics in relation to multiple population sub-groups. Aequitas facilitates informed and equitable decisions around developing and deploying algorithmic decision making […]

Objective(s)


TechnicalUnited StatesUploaded on Mar 17, 2022

Using crowd-worker evaluation as well as standard benchmarks, we empirically demonstrate that our algorithms significantly reduce gender bias in embeddings while preserving the its useful properties such as the ability to cluster related concepts and to solve analogy tasks. The resulting embeddings can be used in applications without amplifying gender bias.

Objective(s)


TechnicalIndiaUploaded on Mar 17, 2022

This paper presents RAN-Debias, a gender de-biasing methodology which eliminates the bias present in a word vector and alters the spatial distribution of its neighboring vectors, achieving a bias-free setting while maintaining minimal semantic offset. This paper also proposes a new bias evaluation metric – Gender-based Illicit Proximity Estimate (GIPE), which measures the extent of […]

Objective(s)


TechnicalUploaded on Sep 15, 2022

” ”

Objective(s)


catalogue Logos

Disclaimer: The tools and metrics featured herein are solely those of the originating authors and are not vetted or endorsed by the OECD or its member countries. The Organisation cannot be held responsible for possible issues resulting from the posting of links to third parties' tools and metrics on this catalogue. More on the methodology can be found at https://oecd.ai/catalogue/faq.