Catalogue of Tools & Metrics for Trustworthy AI

These tools and metrics are designed to help AI actors develop and use trustworthy AI systems and applications that respect human rights and are fair, transparent, explainable, robust, secure and safe.

4 citations of this metric

The Hughes Hallucination Evaluation Model (HHEM) Score is a metric designed to detect hallucinations in text generated by AI systems. It outputs a probability score between 0 and 1, where 0 indicates hallucination and 1 indicates factual consistency. The metric is particularly suitable for Large Language Models (LLMs) and Retrieval-Augmented Generation (RAG) systems. Vectara recommends a threshold of 0.5 to classify outputs as factually consistent.

 

Applicable Models

• Large Language Models (LLMs)

• Retrieval-Augmented Generation (RAG) systems

• Summarization models

• Natural Language Inference (NLI) models

 

Background

 

HHEM is built on Microsoft’s DeBERTa-v3-base model, fine-tuned on text summarization datasets after initial training on NLI data. It is optimized for detecting factual inconsistencies and has become a key tool in addressing hallucination detection challenges in generative AI.

 

Formulae

 

HHEM computes scores using a pretrained cross-encoder. The output is a probability score derived from input pairs (ground truth, inference).

Applications

• Hallucination detection in RAG pipelines

• Evaluation of factual consistency in summarization models

• Accuracy enhancement in enterprise LLM deployments

• Real-time inference scoring for generative AI systems

 

Impact

 

HHEM enables low-latency, cost-effective hallucination detection compared to LLM judge methods. Its calibration supports actionable probabilities, allowing users to fine-tune thresholds for specific applications. By offering multilingual support and efficient computation, HHEM contributes to the reliability and trustworthiness of generative AI systems.

References

Vectara. “hallucination_evaluation_model (Revision 7437011).” Published on Hugging Face, 2024. DOI: 10.57967/hf/3240. Accessible at: https://huggingface.co/vectara/hallucination_evaluation_model.

catalogue Logos

Disclaimer: The tools and metrics featured herein are solely those of the originating authors and are not vetted or endorsed by the OECD or its member countries. The Organisation cannot be held responsible for possible issues resulting from the posting of links to third parties' tools and metrics on this catalogue. More on the methodology can be found at https://oecd.ai/catalogue/faq.