These tools and metrics are designed to help AI actors develop and use trustworthy AI systems and applications that respect human rights and are fair, transparent, explainable, robust, secure and safe.
ROUGE, or Recall-Oriented Understudy for Gisting Evaluation, is a set of metrics and a software package used for evaluating automatic summarization and machine translation software in natural language processing. The metrics compare an automatically produced summary or translation against a reference or a set of references (human-produced) summary or translation.
Note that ROUGE is case insensitive, meaning that upper case letters are treated the same way as lower case letters.
Related use cases :
Better Summarization Evaluation with Word Embeddings for ROUGE
Uploaded on Nov 1, 2022ROUGE is a widely adopted, automatic evaluation measure for text summarization. While it has been shown to correlate well with human judgements, it is biased towards surface le...
BioT5: Enriching Cross-modal Integration in Biology with Chemical Knowledge and Natural Language Associations
Uploaded on Nov 1, 2023ERNIE-GEN: An Enhanced Multi-Flow Pre-training and Fine-tuning Framework for Natural Language Generation
Uploaded on Nov 1, 2023MAST: Multimodal Abstractive Summarization with Trimodal Hierarchical Attention
Uploaded on Nov 1, 2023Momentum Calibration for Text Generation
Uploaded on Nov 1, 2023Multimodal Pretraining for Dense Video Captioning
Uploaded on Nov 1, 2023NITS-VC System for VATEX Video Captioning Challenge 2020
Uploaded on Nov 1, 2023PRIMERA: Pyramid-based Masked Sentence Pre-training for Multi-document Summarization
Uploaded on Nov 1, 2023RefineCap: Concept-Aware Refinement for Image Captioning
Uploaded on Nov 1, 2023Scaling Up Vision-Language Pre-training for Image Captioning
Uploaded on Nov 1, 2023About the metric
You can click on the links to see the associated metrics
Objective(s):
Target sector(s):
Lifecycle stage(s):
Target users: