Evaluates whether each user query or statement in a conversation receives an appropriate and complete response from the AI. This metric assesses if the conversation reaches satisfactory conclusions for each user interaction, ensuring that questions are answered and statements are appropriately acknowledged.
Click here to learn how to setup evaluation using the Python SDK.Input:
string
- conversation history between the user and the model provided as query and response pairsfloat
- returns score between 0 and 1