Prompt Perplexity measures how well a language model predicts the tokens in a given input prompt. It’s calculated based on the likelihood the model assigns to each token in the prompt.
input
: The prompt text to be evaluated.model
: The language model (e.g., “gpt-4o-mini”).Click here to learn how to setup evaluation using the Python SDK.
Input | Parameter | Type | Description |
---|---|---|---|
Required Inputs | input | string | The prompt text to be evaluated. |
Config Parameters | model | string | The language model (e.g., “gpt-4o-mini”). |
Output | Type | Description |
---|---|---|
Score | float | Percentage score between 0 and 1. |