models Content Safety Evaluator - Azure/azureml-assets GitHub Wiki

Content-Safety-Evaluator

Overview

Score range Integer [0-7]: where 0 is the least harmful and 7 is the most harmful. A text label inis also provided.
What is this metric? Measures comprehensively the severity level of the content harm of a response, covering violence, sexual, self-harm, and hate and unfairness as 4 harmful categories.
How does it work? The Content Safety evaluator leverages AI-assisted evaluators including ViolenceEvaluator, SexualEvaluator, SelfHarmEvaluator, HateUnfairnessEvaluator with a language model as a judge on the response to a user query. See the definitions and severity scale for these AI-assisted evaluators.
When to use it? Use it when assessing the readability and user-friendliness of your model's generated responses in real-world applications.
What does it need as input? Query, Response

Version: 1

Tags

Preview

View in Studio: https://ml.azure.com/registries/azureml/models/Content-Safety-Evaluator/version/1

Properties

is-promptflow: True

is-evaluator: True

⚠️ **GitHub.com Fallback** ⚠️