Completeness
Evaluates whether the response fully addresses the input query. This evaluation is crucial for ensuring that the generated response is comprehensive and leaves no aspect of the query unanswered.
Evaluation Using Interface
Input:
- Required Inputs:
- input: The original text or query column.
- output: The AI-generated content column.
Output:
- Score: Percentage score between 0 and 100
Interpretation:
- Higher scores: Indicate that the
output
comprehensively addresses the requirements or topics presented in theinput
. - Lower scores: Suggest that the
output
is missing key details or fails to cover significant aspects mentioned in theinput
.
Evaluation Using Python SDK
Click here to learn how to setup evaluation using the Python SDK.
Input Type | Parameter | Type | Description |
---|---|---|---|
Required Inputs | input | string | The original text or query. |
output | string | The AI-generated content. |
Output | Type | Description |
---|---|---|
Score | float | Returns a score between 0 and 1, where higher values indicate more complete content relative to the input. |
What to do when Completeness is Low
Determine which aspects of the query have not been fully addressed and identify any gaps or incomplete sections that require additional information.
Enhancing the response involves adding missing details to ensure it is comprehensive and refining the content to cover all aspects of the query.
To improve completeness in the long term, implementing mechanisms that align responses more closely with query requirements and enhancing the response generation process to prioritise completeness can help ensure more thorough and accurate outputs.
Was this page helpful?