models Coherence Evaluator - Azure/azureml-assets GitHub Wiki

Coherence-Evaluator

Overview

Score range Integer [1-5]: 1 is the lowest quality and 5 is the highest quality.
What is this metric? Measures how well the language model can produce output that flows smoothly, reads naturally, and resembles human-like language.
How does it work? The coherence metric is calculated by instructing a language model to follow the definition (in the description) and a set of grading rubrics, evaluate the user inputs, and output a score on a 5-point scale (higher means better quality). Learn more about our definition and grading rubrics.
When to use it? The recommended scenario is generative business writing such as summarizing meeting notes, creating marketing materials, and drafting email.
What does it need as input? Query, Response

Version: 4

Tags

hiddenlayerscanned

View in Studio: https://ml.azure.com/registries/azureml/models/Coherence-Evaluator/version/4

Properties

is-promptflow: True

is-evaluator: True

show-artifact: True

_default-display-file: ./coherence.prompty

⚠️ **GitHub.com Fallback** ⚠️