Prism AI Gateway

A single API that routes requests across 100+ LLM providers with built-in safety, caching, cost tracking, and reliability.

What is Prism?

Prism is Future AGI’s AI Gateway — a proxy layer that sits between your application and LLM providers. Instead of managing separate API keys, rate limits, and error handling for each provider, you connect to Prism once and it handles routing, failover, safety checks, caching, and cost tracking automatically.

You send requests to Prism using the standard OpenAI API format. Prism routes them to the appropriate provider, applies any guardrails and caching rules, and returns a response — with extra metadata headers showing which provider was used, how much the request cost, and whether the cache was hit.

Note

Already using the OpenAI SDK? You can keep your existing code. Point base_url at https://gateway.futureagi.com, swap your API key for a Prism key, and switch between 100+ providers by changing the model name — no other code changes needed.


Get started


What Prism does

  • Route requests across 100+ providers — OpenAI, Anthropic, Google Gemini, AWS Bedrock, Azure, Mistral, Groq, and more. Switch providers by changing the model name. No client code changes.
  • Add safety guardrails — 18+ built-in checks including PII detection, prompt injection prevention, content moderation, and secret detection. Enforce, monitor, or log.
  • Balance load and fail over — Weighted, latency-based, and cost-optimized routing with automatic failover, retries, and circuit breaking.
  • Cache responses — Exact match and semantic caching at the gateway level. Repeated queries return instantly without calling the provider.
  • Track costs and set budgets — Per-request cost in every response header. Budget limits block requests when exceeded. Cost attribution by team, feature, or user.
  • Stream in real time — Full streaming pass-through. Responses stream token-by-token using the standard format, across all providers.
  • Observe everything — Request logs, latency metrics, error tracking, and a full analytics dashboard.

Explore features


Supported providers

Prism connects to cloud providers, API services, and self-hosted models. Providers with different native APIs (Anthropic, Gemini, Bedrock) are automatically translated to the standard OpenAI format, so your code stays the same regardless of which provider handles the request.

ProviderType
OpenAICloud API
AnthropicCloud API
Google GeminiCloud API
AWS BedrockCloud API
Azure OpenAICloud API
CohereCloud API
Groq, Together AI, FireworksCloud API
Mistral AI, DeepInfra, PerplexityCloud API
xAI, OpenRouterCloud API
Ollama, vLLM, LM StudioSelf-hosted

For the full list of endpoints, request headers, and response metadata headers, see Core Concepts.

Was this page helpful?

Questions & Discussion