How To
Evaluate Using Agent as a Judge
Uses AI agents to conduct structured evaluations of content, leveraging customisable prompts and system instructions for comprehensive assessment. Unlike rule-based compliance checks, this approach enables nuanced analysis by allowing agents to interpret content contextually based on predefined criteria.
Click here to read the eval definition of Agent as a Judge
a. Using Interface
Configuration Parameters
- Model: Specifies the LLM model used for evaluation.
- Eval Prompt: The main evaluation prompt that defines how the AI should judge the response.
- System Prompt: A higher-level instruction that guides the AI’s evaluation behaviour.