Overview

Monitor and evaluate LLM applications in production with real-time tracing, session analysis, and alerting.

About

Observability is how you monitor your AI application after it goes live. Once your app is in production, things change: user inputs vary, model behavior shifts, and issues come up that testing never caught. Observability gives you a continuous view of how your application is performing so you can stay on top of it.

It tracks every response your application generates, groups them by session and user, scores them for quality, and alerts you when something goes wrong. Instead of finding out about problems from users, you see them in the dashboard first.

Sessions Overview

How Observability Connects to Other Features

  • Prototype: After you promote a winning version in Prototype, its traces continue flowing into Observe so you can monitor production performance against the same quality criteria. Learn more
  • Evaluation: Observability uses the same built-in eval templates to score production traces automatically. Any eval you configured in Prototype or Datasets runs the same way here. Learn more
  • Alerts: Observability feeds into the alerting system so you are notified when quality, cost, or latency crosses a threshold in production. Learn more

Getting Started with Observability

Was this page helpful?

Questions & Discussion