Catalogue of Tools & Metrics for Trustworthy AI

These tools and metrics are designed to help AI actors develop and use trustworthy AI systems and applications that respect human rights and are fair, transparent, explainable, robust, secure and safe.

Fujitsu LLM Bias Diagnosis



Fujitsu LLM Bias Diagnosis

Fujitsu LLM Bias Diagnosis can evaluate biases across LLMs on different ethical topics. The Toolkit can be used to evaluate LLMs from publicly available LLMs such as those on the HuggingFace Model Hub, or the user can upload their own pre-trained or finetuned LLMs. The evaluation can examine biases in 4 SDGs topics: Climate Action, Gender Equality, Healthcare, and Education, in addition to examining issues related to human wisdom and values.

To evaluate the LLM, Fujitsu LLM Bias Diagnosis uses a curated dataset of test statements. For each statement, one or more words potentially associated with sensitive attributes were masked, and a counterstatement was created. The results of the evaluations are visualized through intuitive Word Clouds, where three exemplar test statements are shown. Bar Plot visualizations, moreover, allow for the comparison of LLMs to identify the LLM that yields the highest probability score of correctly reconstructing the original text, without bias or misinformation.

About the tool


Developing organisation(s):




Country of origin:


Lifecycle stage(s):


Type of approach:



Tags:

  • large language model
  • fairness
  • bias

Modify this tool

Use Cases

There is no use cases for this tool yet.

Would you like to submit a use case for this tool?

If you have used this tool, we would love to know more about your experience.

Add use case
catalogue Logos

Disclaimer: The tools and metrics featured herein are solely those of the originating authors and are not vetted or endorsed by the OECD or its member countries. The Organisation cannot be held responsible for possible issues resulting from the posting of links to third parties' tools and metrics on this catalogue. More on the methodology can be found at https://oecd.ai/catalogue/faq.