Function
Safety
Evaluate your LLM models on their ability to be safe and appropriate.
Safety evaluation refers to the ability to evaluate and validate text output from LLM models based on various safety criteria. This includes checking for sensitive topics, profanity, personally identifiable information, and other safety concerns.
Evaluation Type | Description | Required Keys |
---|---|---|
SafeForWorkText | Check if text is safe for work | response |
NotGibberishText | Check if text is not gibberish | response |
ContentModeration | Use OpenAI’s content moderation | text |
PiiDetection | Check for personally identifiable information | text |
PromptInjection | Check for prompt injection attempts | text |
Toxicity | Check for toxic/harmful language | input |
DataPrivacyCompliance | Check compliance with privacy regulations | input |
CulturalSensitivity | Check for cultural appropriateness and inclusive language | input |
BiasDetection | Check for various forms of bias | input |
These evaluations do not require any additional configuration.