SDKs

Evaluate LLM outputs, trace AI calls, optimize prompts, and test voice agents. Python, TypeScript, Java, and C# supported.

📝
TL;DR
  • Python: evals, tracing, datasets, prompts, optimization, simulation
  • TypeScript: evals, tracing, datasets, prompts
  • Java / C#: tracing
  • pip install ai-evaluation or npm install @future-agi/ai-evaluation to get started

Future AGI is a set of packages that evaluate LLM outputs, trace calls across your stack, optimize prompts, and load-test voice agents. Install what you need, skip what you don’t.

Language Support

ModulePythonTypeScriptJavaC#
EvaluationsFullFull
TracingFull (45+)Full (40+)Full (25+)Full
DatasetsFullFull
PromptsFullFull
Prompt OptimizationFull
SimulationFull

Quickstart

pip install ai-evaluation

Requires Python 3.10+. This also installs futureagi (datasets, prompts, knowledge bases) automatically.

export FI_API_KEY="your-api-key"
export FI_SECRET_KEY="your-secret-key"
from fi.evals import evaluate

# Local metric — no API key needed
result = evaluate("contains", output="Hello world", keyword="Hello")
print(result.score)    # 1.0
print(result.passed)   # True

# Cloud metric — needs FI_API_KEY and FI_SECRET_KEY
result = evaluate("toxicity", output="Hello world", model="turing_flash")
print(result.score)    # 1.0
print(result.passed)   # True

Want tracing too? Add the instrumentor for your provider:

pip install fi-instrumentation-otel traceai-openai
npm install @future-agi/ai-evaluation
export FI_API_KEY="your-api-key"
export FI_SECRET_KEY="your-secret-key"
import { Evaluator, Tone } from "@future-agi/ai-evaluation";

const evaluator = new Evaluator();

const result = await evaluator.evaluate({
  evalTemplates: [new Tone()],
  inputs: [{
    query: "Write a professional email",
    response: "Dear Sir/Madam, I hope this message finds you well..."
  }],
  modelName: "turing_flash"
});

console.log(result);

Want tracing too?

npm install @traceai/fi-core @traceai/openai

Java support covers tracing only. 25+ instrumentors including Spring AI and LangChain4j.

<!-- Maven — add the JitPack repository -->
<repository>
    <id>jitpack.io</id>
    <url>https://jitpack.io</url>
</repository>

<dependency>
    <groupId>com.github.future-agi.traceAI</groupId>
    <artifactId>traceai-java-openai</artifactId>
    <version>LATEST</version>
</dependency>

See the Tracing docs for setup instructions.

C# support covers tracing only.

dotnet add package fi-instrumentation-otel

See the Tracing docs for setup instructions.

Getting an error? Check these common issues

ModuleNotFoundError: No module named 'fi' — The package is called ai-evaluation, not future-agi or futureagi-sdk:

pip install ai-evaluation

AuthenticationError — Both FI_API_KEY and FI_SECRET_KEY must be set. The API key alone is not enough.

Python version errorai-evaluation requires Python 3.10+. Check with python --version.

Packages

Python

Six packages, each installable independently:

PackageInstallWhat it doesPython
futureagipip install futureagiDatasets, prompt versioning, knowledge bases3.9+
ai-evaluationpip install ai-evaluation76+ local metrics + 100+ cloud templates, guardrails, streaming eval3.10+
fi-instrumentation-otelpip install fi-instrumentation-otelOpenTelemetry tracing for AI apps3.9+
traceai-*pip install traceai-openaiAuto-instrumentation for 45+ frameworks3.9+
agent-optpip install agent-optPrompt optimization (6 algorithms)3.10+
agent-simulatepip install agent-simulateSimulate voice AI agents at scale3.10+
futureagi                    ← standalone base layer
  └── ai-evaluation          ← installs futureagi automatically
        └── agent-opt        ← installs ai-evaluation automatically

fi-instrumentation-otel      ← standalone tracing layer
  ├── traceai-*              ← each installs fi-instrumentation-otel
  └── agent-simulate         ← installs fi-instrumentation-otel

Tip

You don’t need to install dependencies manually. pip install ai-evaluation gives you futureagi too. pip install traceai-openai gives you fi-instrumentation-otel too.

TypeScript

PackageInstallWhat it does
@future-agi/sdknpm install @future-agi/sdkDatasets, prompt versioning, knowledge bases
@future-agi/ai-evaluationnpm install @future-agi/ai-evaluationEval metrics and guardrails
@traceai/fi-corenpm install @traceai/fi-coreTracing core
@traceai/openainpm install @traceai/openaiFramework instrumentors (40+)

Java and C#

Tracing only. Java has 25+ instrumentors (Maven via JitPack, group ID com.github.future-agi.traceAI). C# has a single NuGet package (fi-instrumentation-otel). See the Tracing reference for details.

Evaluations — ai-evaluation

76+ local metrics for things like tone, hallucination, bias, and factual accuracy. Also includes guardrails (toxicity, PII, prompt injection) that run in under 10ms.

Available in Python and TypeScript.

Optional extras (Python)
ExtraInstallWhat it adds
NLI modelspip install ai-evaluation[nli]DeBERTa for faithfulness and hallucination detection
Embeddingspip install ai-evaluation[embeddings]Sentence-transformers for semantic similarity
Feedbackpip install ai-evaluation[feedback]ChromaDB-backed feedback collection
Distributedpip install ai-evaluation[celery]Celery + Redis for distributed eval runs
Everythingpip install ai-evaluation[all]All optional dependencies

Tracing — fi-instrumentation-otel + traceai-*

Install the core library plus one instrumentor per framework. LLM calls, retrieval steps, and agent actions get traced and sent to your Future AGI dashboard.

Available in Python, TypeScript, Java, and C#.

from fi_instrumentation import register
from fi_instrumentation.fi_types import ProjectType

trace_provider = register(
    project_name="my-project",
    project_type=ProjectType.OBSERVE,
)

from traceai_openai import OpenAIInstrumentor
OpenAIInstrumentor().instrument(tracer_provider=trace_provider)

# All OpenAI calls are now traced
# Traces appear in your Future AGI dashboard under "my-project"
PackageFramework
traceai-openaiOpenAI
traceai-anthropicAnthropic
traceai-google-genaiGoogle Generative AI
traceai-vertexaiGoogle Vertex AI
traceai-bedrockAWS Bedrock
traceai-mistralaiMistral AI
traceai-groqGroq
traceai-litellmLiteLLM
traceai-cohereCohere
traceai-ollamaOllama
traceai-deepseekDeepSeek
traceai-togetherTogether AI
traceai-fireworksFireworks AI
traceai-cerebrasCerebras
traceai-xaixAI / Grok
traceai-vllmvLLM
traceai-portkeyPortkey
traceai-huggingfaceHuggingFace
PackageFramework
traceai-langchainLangChain / LangGraph
traceai-llamaindexLlamaIndex
traceai-crewaiCrewAI
traceai-openai-agentsOpenAI Agents SDK
traceai-autogenMicrosoft AutoGen
traceai-smolagentsHuggingFace SmolAgents
traceai-google-adkGoogle Agent Dev Kit
traceai-claude-agent-sdkClaude Agent SDK
traceai-pydantic-aiPydantic AI
traceai-strandsAWS Strands Agents
traceai-agnoAgno
traceai-beeaiIBM BeeAI
traceai-haystackHaystack
traceai-dspyDSPy
traceai-guardrailsGuardrails AI
traceai-instructorInstructor
traceai-mcpModel Context Protocol
PackageFramework
traceai-pipecatPipecat
traceai-livekitLiveKit
PackageFramework
traceai-pineconePinecone
traceai-chromadbChromaDB
traceai-qdrantQdrant
traceai-weaviateWeaviate
traceai-milvusMilvus
traceai-lancedbLanceDB
traceai-mongodbMongoDB
traceai-pgvectorpgvector
traceai-redisRedis

Tip

Each instrumentor is lightweight and independent. Only install the ones for frameworks you actually use.

Core SDK — futureagi

Datasets, prompt versioning, and knowledge bases. If you installed ai-evaluation, you already have this.

Available in Python and TypeScript.

Prompt Optimization — agent-opt

Six optimization algorithms: Random Search, Bayesian, ProTeGi, Meta-Prompt, PromptWizard, and GEPA. Each uses eval metrics to score prompt variants and find the best one.

Python only.

pip install agent-opt

Simulation Testing — agent-simulate

Run simulated conversations against your voice AI agents using configurable personas. Captures audio, transcripts, and eval scores.

Python only.

pip install agent-simulate

Next Steps

Was this page helpful?

Questions & Discussion