Catalogue of Tools & Metrics for Trustworthy AI

These tools and metrics are designed to help AI actors develop and use trustworthy AI systems and applications that respect human rights and are fair, transparent, explainable, robust, secure and safe.

Machine learning models, at the core of AI applications,  typically achieve a high accuracy at the expense of an insufficient explainability. Moreover, according to the proposed regulations,  AI applications based on machine learning must be "trustworthy'', and comply with a set of mandatory requirements, such as Sustainability and Fairness. To date there are no standardised metrics that can ensure an integrated overall assessment of the trustworthiness of AI applications, and provide a summary score of the Sustainability, Accuracy, Fairness and Explainability of an AI application.
To fill the gap, we propose a set of integrated statistical methods, based on the Lorenz Curve,  that can be used to assess and monitor over time whether an AI application is trustworthy, and what are the risks of not being such . Specifically, the methods will measure Sustainability (in terms of robustness with respect to anomalous and cyber inflated data), Accuracy (in terms of predictive accuracy), Fairness (in terms of prediction bias across different population groups) and Explainability (in terms of human understanding and oversight).

About the metric






Target users:


Risk management stage(s):

Modify this metric

catalogue Logos

Disclaimer: The tools and metrics featured herein are solely those of the originating authors and are not vetted or endorsed by the OECD or its member countries. The Organisation cannot be held responsible for possible issues resulting from the posting of links to third parties' tools and metrics on this catalogue. More on the methodology can be found at https://oecd.ai/catalogue/faq.