Skip to content

models Gleu Score Evaluator

github-actions[bot] edited this page Oct 17, 2024 · 5 revisions

Gleu-Score-Evaluator

Overview

Score range Float [0-1]
What is this metric? Measures the degree of overlap between the generated text and both the reference text and source text, balancing between precision and recall.
How does it work? The GLEU score is computed by averaging the precision and recall of n-grams between the generated text and both the reference text and source text. It considers both the overlap of n-grams with the reference (similar to BLEU) and penalizes for over-generation. The score provides a balanced metric, where a value of 1 represents perfect overlap, and 0 represents no overlap.
When to use it? Use the GLEU score when you want a more balanced evaluation of generated text that considers both the precision and recall of n-gram overlap, especially useful in evaluating machine translation or paraphrasing tasks.
What does it need as input? Ground Truth Response, Generated Response

Version: 2

Tags

Preview

View in Studio: https://ml.azure.com/registries/azureml/models/Gleu-Score-Evaluator/version/2

Properties

is-promptflow: True

is-evaluator: True

show-artifact: True

_default-display-file: ./evaluator/_gleu.py

Clone this wiki locally