Function
Custom
Evaluate your LLM models on their ability to generate text based on custom criteria.
Custom validation refers to the ability to evaluate and validate text output from LLM models based on custom criteria. This includes running custom validation code, making external API calls for validation, and verifying text is in the correct language.
Evaluation Type | Description | Required Keys |
---|---|---|
LLMJudge | Uses language models to evaluate content | None |
GradingCriteria | Evaluates responses against custom grading criteria | response |
AgentJudge | Uses AI agents for content evaluation | None |
Evaluation Type | Description | Required Configuration |
---|---|---|
LLMJudge | Uses language models to evaluate content | model : LLM model to use eval_prompt : Evaluation prompt system_prompt : System prompt |
GradingCriteria | Evaluates against grading criteria | grading_criteria : Grading criteria model : LLM model to use |
AgentJudge | Uses AI agents for evaluation | model : LLM model to use eval_prompt : Agent prompt system_prompt : System prompt |