Integrate Future AGI evaluations with Langfuse
evaluator.evaluate()
with the platform="langfuse"
parameter inside an active Langfuse span, the evaluation is executed. The results are then automatically attached as scores to that specific span in your Langfuse dashboard.
levenshtein_similarity
evaluation.
evaluator.configure_evaluations()
:
platform="langfuse"
: This essential parameter directs the evaluation results to be sent to Langfuse and linked with the current active span.custom_eval_name
: This parameter is required and provides a unique, human-readable name for your evaluation instance. This name will appear on the score in the Langfuse UI, helping you distinguish between different evaluations.eval_config
: This dictionary contains the configuration for the evaluation, including the eval_templates
to use and the inputs
for the evaluation.