AgDex / Directory / Langfuse
🔭

Langfuse

LLM Observability & Analytics

Open-source LLM observability that gives engineering teams real visibility into what their AI applications are actually doing — tracing, evaluation, prompt management, and cost analytics in one platform.

By AgDex Editorial · Reviewed & updated April 2026

What is Langfuse?

Langfuse is an open-source LLM observability platform designed to help engineering teams understand, debug, and improve AI applications in production. Founded in 2023 by Max Langenkamp, Clemens Rawert, and Marc Klingen, Langfuse emerged from a simple but important observation: the tooling that engineers rely on for understanding what their traditional software does — logging, tracing, performance monitoring — simply doesn't exist in a form suitable for LLM-based applications.

When you ship a traditional web service, you have APM tools, error trackers, log aggregators, and dashboards that tell you exactly what's happening. When you ship an LLM application, you're largely flying blind — you might know your application threw an error, but understanding why the AI gave a poor response, what context it had, how long each step took, and how much it cost requires purpose-built tooling. Langfuse is that purpose-built tooling.

The platform centers on "traces" — hierarchical records of everything that happened during an LLM interaction. A trace might contain the user's input, the prompts sent to the model, intermediate reasoning steps for an agent, tool calls and their results, and the final response — all timestamped, all annotatable, all searchable. This visibility is what allows teams to move from "users say the AI is bad sometimes" to "here are the specific failure modes and here's how to fix them."

By 2026, Langfuse had become one of the most widely adopted LLM observability platforms, with deep integrations for LangChain, LlamaIndex, OpenAI SDK, Anthropic SDK, and most major LLM frameworks. Its self-hosting option made it particularly attractive for enterprise teams with data residency requirements, while the cloud offering served teams wanting managed infrastructure without operational overhead.

Key Features

1. Comprehensive LLM Tracing

Langfuse's trace view is the heart of the platform. Each trace captures the full execution context of an LLM operation: input/output at every step, token counts, latency at each stage, model parameters used, and any custom metadata you choose to attach. For multi-step agent workflows, traces show the full decision tree — which tools were called, what they returned, how the model used that information. In our testing, this visibility reduced debugging time for complex agent issues from hours to minutes.

2. Prompt Management and Versioning

Langfuse includes a prompt registry where teams can store, version, and deploy prompts separately from application code. This decoupling means you can A/B test prompt variations, roll back bad prompt changes, and track which prompt version was used for each trace — all without a code deployment. For teams iterating heavily on prompt engineering, this is a surprisingly powerful capability that changes the workflow significantly.

3. Evaluation and Scoring Framework

Langfuse provides a flexible evaluation system that supports both human annotation (annotators reviewing traces and assigning quality scores) and LLM-as-judge evaluation (automatically scoring outputs using another LLM). You can define custom evaluation rubrics, run evaluations on subsets of production traffic, and track score distributions over time as your application and prompts evolve. This is critical for measuring whether prompt changes actually improve quality.

4. Cost and Token Analytics

The platform automatically calculates costs per trace based on the models used and token counts, aggregated across users, sessions, and time periods. Teams can identify which users, use cases, or application flows are driving the most cost — critical for LLM applications where cost per request can vary by orders of magnitude based on what the model generates. We've seen teams cut LLM costs by 30-40% simply by using this visibility to identify inefficient flows they hadn't noticed.

5. Dataset Management for Testing

Langfuse lets you build and manage evaluation datasets directly from your production traces — you identify interesting or problematic interactions, add them to a dataset, and then use those datasets to benchmark future prompt versions or model changes. This "production data to test suite" pipeline is exactly the right workflow for LLM application quality assurance.

6. Framework-Agnostic SDK

Integration with Langfuse is available via Python and JavaScript SDKs, as well as native integrations for LangChain (callback handler), LlamaIndex (callback handler), and OpenTelemetry. The OpenTelemetry support is particularly notable — it means observability data from LLM calls can be unified with your existing APM infrastructure rather than requiring a completely separate observability stack.

Pros & Cons

✅ Pros

❌ Cons

Use Cases

1. Debugging Production LLM Application Issues

The most fundamental use case: a user reports the AI gave a wrong or harmful response, and you need to understand why. Without Langfuse (or similar), you're asking users to reproduce the issue while you stare at logs that don't tell you what the model saw. With Langfuse, you can look up the exact trace — every prompt, every context document retrieved, every tool call — and understand precisely what led to the bad output. In our experience, this changes debugging from a painful guessing game to a systematic investigation.

2. Measuring the Impact of Prompt Changes

Engineering teams use Langfuse to instrument a controlled comparison before and after prompt changes. By creating evaluation datasets from representative production traces, tagging traces by prompt version, and comparing quality scores across versions, teams can make evidence-based decisions about whether a prompt iteration actually improved performance — rather than relying on developer intuition.

3. LLM Cost Optimization

Finance teams and engineering leads use the cost analytics to understand and optimize LLM spending. Common findings: specific users or workflows driving disproportionate token usage, system prompts that grew unnoticed over iterations to inefficient lengths, model selection that could be downgraded for simpler subtasks without quality impact. One team we spoke with reduced their monthly OpenAI bill by 45% within a month of instrumenting with Langfuse.

4. Human Annotation and Quality Assurance Programs

Teams running ongoing quality assurance programs for their AI products use Langfuse's annotation interface to assign reviewers to traces, capture structured quality scores, and build institutional knowledge about failure patterns. This is particularly valuable for AI products in regulated industries or high-stakes domains where systematic quality review is required.

Pricing

Langfuse's self-hosted version is free under the MIT license — you provide the infrastructure (PostgreSQL, Redis, application server) and there are no software costs. This is the recommended path for teams with data residency requirements or significant trace volumes.

The cloud offering includes a free Hobby plan with 50,000 observations/month and 30-day data retention — sufficient for development and early production testing. The Team plan starts at $59/month with higher volume limits and extended retention. The Pro plan at $199/month includes SSO, longer retention, and priority support. Enterprise plans with custom data residency and SLA guarantees are available.

One consideration: trace volume scales fast in production. A single user conversation with a RAG pipeline might generate 10-20 observations. Teams should estimate their production traffic carefully before relying on the Hobby plan limits.

Alternatives

ToolBest ForKey Difference
LangSmithTeams heavily invested in LangChainDeeper LangChain integration; closed-source, no self-hosting, can feel vendor-locky
HeliconeSimple API usage monitoringEasier setup for pure API monitoring; less powerful for agent/RAG trace visualization
Arize PhoenixData science teams with ML backgroundStrong ML experimentation features; steeper learning curve, more complex deployment

Our Verdict

Langfuse has earned its place as the default choice for LLM observability among engineering teams building serious AI applications. The combination of open source licensing, self-hosting capability, excellent trace UI, and active development makes it difficult to argue against — particularly given that the free tiers (both self-hosted and cloud Hobby) allow meaningful production use without initial cost commitment.

The honest critique is that the platform is strongest at the "capture and inspect" layer and still maturing at the "evaluate and improve" layer. If your primary need is understanding what your AI application is doing, Langfuse is excellent. If you need a full ML experimentation platform with statistical rigor, you may eventually need to supplement it with other tooling.

For any team shipping an LLM application to real users, adding Langfuse instrumentation should be treated as a non-negotiable baseline — not an optional enhancement. Flying blind in production with AI systems creates risks that are very difficult to recover from once users have experienced failures.

Rating: 4.6/5 — Essential LLM observability tooling; evaluation layer still maturing.