Evaluates content safety using OpenAI’s content moderation system to detect and flag potentially harmful, inappropriate, or unsafe content. This evaluation provides binary (Pass/Fail) assessment of text content against established safety guidelines.
Click here to learn how to setup evaluation using the Python SDK.
Input Type | Parameter | Type | Description |
---|---|---|---|
Required Inputs | text | string | The text content to moderate. |
Output | Type | Description |
---|---|---|
Score | bool | Returns a score between 0 and 1. Higher values indicate safer content, lower values indicate potentially inappropriate content. |