Catalogue of Tools & Metrics for Trustworthy AI

These tools and metrics are designed to help AI actors develop and use trustworthy AI systems and applications that respect human rights and are fair, transparent, explainable, robust, secure and safe.

Type

Performance

Clear all

Origin

Scope

SUBMIT A TOOL

If you have a tool that you think should be featured in the Catalogue of AI Tools & Metrics, we would love to hear from you!

SUBMIT
Objective Performance

Uploaded on Apr 23, 2024
This Recommendation specifies use cases and requirements for multimedia communication enabled vehicle systems using artificial intelligence, including overview, use cases, high-layer architecture, service and network requirements, functional requirements, and non-functional requirements.

Objective(s)


TechnicalUnited StatesUploaded on Apr 18, 2024
An end-to-end model risk management platform that automates model documentation and dramatically simplifies AI model validation.

EducationalUploaded on Mar 14, 2024
Teeny-Tiny Castle is a collection of tutorials on how to use tools for AI Ethics and Safety research.

Uploaded on Dec 14, 2023
Our work enables developers and policymakers to anticipate, measure, and address discrimination as language model capabilities and applications continue to expand.

Objective(s)

Related lifecycle stage(s)

Operate & monitorDeployPlan & design

TechnicalTurkeyUploaded on Dec 11, 2023
Framework agnostic sliced/tiled inference + interactive ui + error analysis plots

Objective(s)

Related lifecycle stage(s)

Build & interpret model

TechnicalUploaded on Dec 11, 2023
A hyperparameter optimization framework

Objective(s)


TechnicalUploaded on Dec 11, 2023
Relax! Flux is the ML library that doesn't make you tensor

Objective(s)

Related lifecycle stage(s)

Build & interpret model

TechnicalUnited StatesUploaded on Dec 11, 2023
State-of-the-Art Deep Learning scripts organized by models - easy to train and deploy with reproducible accuracy and performance on enterprise-grade infrastructure.

TechnicalUnited StatesUploaded on Dec 11, 2023
Anomaly detection related books, papers, videos, and toolboxes

Objective(s)

Related lifecycle stage(s)

DeployVerify & validate

TechnicalEducationalUploaded on Nov 29, 2023
This is the First Edition of CAN/CIOSC 101:2019, Ethical design and use of automated decision systems. CAN/CIOSC 101:2019 was prepared by the CIO Strategy Council Technical Committee 2 (TC 2) on the ethical design and use of automated decision systems, comprised of over 100 thought leaders and experts in artificial intelligence, ethics, and related subjects.

Objective(s)

Related lifecycle stage(s)

DeployBuild & interpret modelPlan & design

ProceduralCanadaUploaded on Nov 14, 2023
In undertaking this voluntary commitment, developers and managers of advanced generative systems commit to working to achieve outcomes related to the OECD AI principles.

ProceduralUploaded on Oct 26, 2023
The Playbook provides suggested actions for achieving the outcomes laid out in the AI Risk Management Framework (AI RMF) Core (Tables 1 – 4 in AI RMF 1.0). Suggestions are aligned to each sub-category within the four AI RMF functions (Govern, Map, Measure, Manage).

Objective(s)


ProceduralUploaded on Oct 26, 2023
The AIRC supports all AI actors in the development and deployment of trustworthy and responsible AI technologies. AIRC supports and operationalizes the NIST AI Risk Management Framework (AI RMF 1.0) and accompanying Playbook and will grow with enhancements to enable an interactive, role-based experience providing access to a wide-range of relevant AI resources.

Objective(s)


ProceduralUploaded on Oct 26, 2023
The goal of the AI RMF is to offer a resource to the organizations designing, developing, deploying, or using AI systems to help manage the many risks of AI and promote trustworthy and responsible development and use of AI systems.

Objective(s)


TechnicalUploaded on Oct 26, 2023
Monitaur is a model governance software company that enables you to build repeatable patterns and requirements for model development success. Those policies, templates, and applications are then also aligned with regulatory requirements. We're a full model development lifecycle governance solution, with interfaces specifically for development teams, and then also for risk management teams so that as a system of record, you gain transparency, alignment, and greater execution of success in your investments in AI.

Objective(s)

Related lifecycle stage(s)

Operate & monitorDeployVerify & validate

TechnicalUploaded on Oct 13, 2023
Zero-trust 360 RAIAAS platform for audit and certification of enterprise-wide AI solutions using trust indicators following ethical standards

Objective(s)


TechnicalSwitzerlandUploaded on Oct 13, 2023>1 year
Calvin Risk develops comprehensive, quantitative solutions to assess and manage the risks of AI algorithms in commercial use. The tool helps companies create a framework for transparency, governance and standardization to manage their AI portfolios while ensuring that AI remains safe and compliant with the highest ethical standards and upcoming regulatory requirements.

Objective(s)

Related lifecycle stage(s)

Operate & monitorVerify & validate

Uploaded on Sep 14, 2023
Logically uses a Human in the Loop AI framework called HAMLET (Humans and Machines in the Loop Evaluation and Training) to enable the development of trustworthy and responsible AI technologies.

Uploaded on Sep 14, 2023<1 day
FAIRLY provides an AI governance platform focussed on accelerating the broad use of fair and responsible AI by helping organisations bring safer AI models to market.

ProceduralUploaded on Sep 11, 2023
Meta-learning is a promising strategy for learning to efficiently learn within new tasks, using data gathered from a distribution of tasks.

Objective(s)

Related lifecycle stage(s)

Plan & design

catalogue Logos

Disclaimer: The tools and metrics featured herein are solely those of the originating authors and are not vetted or endorsed by the OECD or its member countries. The Organisation cannot be held responsible for possible issues resulting from the posting of links to third parties' tools and metrics on this catalogue. More on the methodology can be found at https://oecd.ai/catalogue/faq.