Catalogue of Tools & Metrics for Trustworthy AI

These tools and metrics are designed to help AI actors develop and use trustworthy AI systems and applications that respect human rights and are fair, transparent, explainable, robust, secure and safe.

Crosslingual Optimized Metric for Evaluation of Translation (COMET) is a metric for automatic evaluation of machine translation that calculates the similarity between a machine translation output and a reference translation using token or sentence embeddings.

 

COMET supports the Robustness objective by providing a reliable and consistent method for evaluating machine translation outputs. By correlating well with human judgments, it helps ensure that translation systems maintain quality and performance across different languages and conditions, which is a key aspect of robustness. However, its connection to other Trustworthy AI objectives is minimal, as it does not directly address transparency, explainability, or other ethical and governance concerns.

Trustworthy AI Relevance

This metric addresses Robustness, Fairness by quantifying relevant system properties. COMET supports Robustness by providing a reliable and consistent measure of translation quality across different languages and conditions, helping to ensure AI systems maintain performance under diverse and potentially adverse linguistic scenarios. It supports Fairness by enabling the detection and mitigation of biases or quality disparities in machine translation outputs across languages, promoting equitable treatment of different language communities.

Related use cases :

Uploaded on Nov 1, 2022

Neural metrics have achieved impressive correlation with human judgements in the evaluation of machine translation systems, but before we can safely optimise towards such metri...



About the metric





Target users:


Risk management stage(s):

Modify this metric

Partnership on AI

Disclaimer: The tools and metrics featured herein are solely those of the originating authors and are not vetted or endorsed by the OECD or its member countries. The Organisation cannot be held responsible for possible issues resulting from the posting of links to third parties' tools and metrics on this catalogue. More on the methodology can be found at https://oecd.ai/catalogue/faq.