- Incomplete responses – Missing relevant details present in the provided context.
- Context misalignment – The model generating responses that contradict or ignore the context.
- Inefficient information usage – Not fully leveraging the available context to improve accuracy.
1. Chunk Attribution
Evaluates whether a language model references the provided context chunks in its generated response. It checks whether the model acknowledges and utilises the context at all. Click here to read the eval definition of Chunk Attributiona. Using Interface
Required Inputs
- output: The generated response.
- context: The provided contextual information.
Optional Inputs
- input: The original query or instruction.
Output
Returns a Pass/Fail result:- Pass – The response references the provided context.
- Fail – The response does not utilise the context.
b. Using SDK
Export your API key and Secret key into your environment variables.
2. Chunk Utilization
Measures how effectively the model integrates the retrieved context chunks into its response. Unlike Chunk Attribution, which only checks for references, Chunk Utilization assigns a score based on how well the context contributes to a meaningful response.- Higher Score – The model extensively incorporates relevant chunks.
- Lower Score – The model includes minimal or no context.
a. Using Interface
Required Inputs
- output: The model’s generated response.
- context: The provided context chunks.
Optional Inputs
- input: The original query or instruction.
Output
- Score between 0 and 1, where higher values indicate better utilization of context.