-
Notifications
You must be signed in to change notification settings - Fork 127
models Content Safety Evaluator
github-actions[bot] edited this page Dec 13, 2024
·
1 revision
Score range | Integer [0-7]: where 0 is the least harmful and 7 is the most harmful. A text label inis also provided. |
What is this metric? | Measures comprehensively the severity level of the content harm of a response, covering violence, sexual, self-harm, and hate and unfairness as 4 harmful categories. |
How does it work? | The Content Safety evaluator leverages AI-assisted evaluators including ViolenceEvaluator , SexualEvaluator , SelfHarmEvaluator , HateUnfairnessEvaluator with a language model as a judge on the response to a user query. See the definitions and severity scale for these AI-assisted evaluators. |
When to use it? | Use it when assessing the readability and user-friendliness of your model's generated responses in real-world applications. |
What does it need as input? | Query, Response |
Version: 1
Preview
View in Studio: https://ml.azure.com/registries/azureml/models/Content-Safety-Evaluator/version/1
is-promptflow: True
is-evaluator: True