Detect Hallucination
Identifies if the model fabricated facts or added information that was not present in the input or context
result = evaluator.evaluate(
eval_templates="detect_hallucination",
inputs={
"context": "Honey never spoils because it has low moisture content and high acidity, creating an environment that resists bacteria and microorganisms. Archaeologists have even found pots of honey in ancient Egyptian tombs that are still perfectly edible.",
"output": "Honey doesn’t spoil because its low moisture and high acidity prevent the growth of bacteria and other microbes."
},
model_name="turing_flash"
)
print(result.eval_results[0].output)
print(result.eval_results[0].reason)import { Evaluator, Templates } from "@future-agi/ai-evaluation";
const evaluator = new Evaluator();
const result = await evaluator.evaluate(
"detect_hallucination",
{
context: "Honey never spoils because it has low moisture content and high acidity, creating an environment that resists bacteria and microorganisms. Archaeologists have even found pots of honey in ancient Egyptian tombs that are still perfectly edible.",
output: "Honey doesn’t spoil because its low moisture and high acidity prevent the growth of bacteria and other microbes."
},
{
modelName: "turing_flash",
}
);
console.log(result); | Input | |||
|---|---|---|---|
| Required Input | Type | Description | |
output | string | Output generated by the model | |
context | string | The context provided to the model | |
| Optional Input | |||
input | string | Input provided to the model |
| Output | ||
|---|---|---|
| Field | Description | |
| Result | Returns Passed if no hallucination is detected, Failed if hallucination is detected | |
| Reason | Provides a detailed explanation of the evaluation |
What to do If you get Undesired Results
If the content is evaluated as containing hallucinations (Failed) and you want to improve it:
- Ensure all claims in your output are explicitly supported by the source material
- Avoid extrapolating or generalizing beyond what is stated in the input
- Remove any specific details that aren’t mentioned in the source text
- Use qualifying language (like “may,” “could,” or “suggests”) when necessary
- Stick to paraphrasing rather than adding new information
- Double-check numerical values, dates, and proper nouns against the source
- Consider directly quoting from the source for critical information
Comparing Detect Hallucination with Similar Evals
- Instruction Adherence: Detect Hallucination checks for fabricated information not present in the source, while Instruction Adherence evaluates whether the output follows the instructions provided.
- Groundedness: Detect Hallucination focuses on absence of fabricated content, while Groundedness measures how well the output is supported by the source material.
- Context Adherence: Detect Hallucination identifies made-up information, while Context Adherence evaluates how well the output adheres to the given context.
Was this page helpful?