← Back to changelog
February 13, 2026

Evaluate Individual Operations: Faster, More Precise LLM-as-a-Judge

Picture Marlies MayerhoferMarlies
Picture Hassieb PakzadHassieb
Evaluate Individual Operations: Faster, More Precise LLM-as-a-Judge

Observation-level evaluations enable precise operation-specific scoring for production monitoring.

LLM-as-a-Judge evaluations can now be run on individual observations—LLM calls, retrievals, tool executions, or any operation within your traces. Previously, evaluations could only be run on entire traces. This architectural shift delivers dramatically faster execution and operation-level precision for production monitoring.

Example: In a RAG pipeline with 10 operations (retrieval → reranking → generation → citation), evaluate only the final generation for helpfulness and the retrieval step for relevance—not the entire workflow. Each operation gets its own score.

Target: Individual observations
Filter: type = "GENERATION" AND name = "final-response"
Result: One score per matching observation
Speed: Seconds per evaluation

Scores attach to specific observations in the trace tree. Filter by observation type, name, metadata, or trace-level attributes to target exactly what matters. Run different evaluators on different operations simultaneously.

Observation-level evaluation in trace tree

Why This Matters

Operation-level precision Evaluate only what matters. Target final LLM responses, retrieval steps, or specific tool calls—not entire workflows. Reduces evaluation volume and cost by filtering to specific operations.

Compositional evaluation Run different evaluators on different operations simultaneously. Toxicity on LLM outputs, relevance on retrievals, accuracy on generations—all within one trace. Stack observation filters (type, name, metadata) with trace attributes filters (userId, sessionId, tags).

Scalable architecture Built for high-volume workloads. At ingest time, observations are evaluated against filter criteria, added to an evaluation queue, and processed asynchronously. No joins, no complex queries—just fast, reliable evaluation at scale.

Getting Started

Navigate to your project → Evaluation → LLM-as-a-Judge → Set up Evaluator

  1. Select or create an evaluator template
  2. Choose “Live Observations” as your evaluation target
  3. Configure observation filters (type, name, metadata)
  4. Add trace-level filters (userId, sessionId, tags) if needed
  5. Map variables from observation fields (input, output, metadata)
  6. Set sampling percentage to manage evaluation costs

Follow the complete LLM-as-a-Judge setup guide for detailed configuration steps and examples.

Requirements

Existing trace-level evaluators continue to work. For users with existing evaluations, see the upgrade guide to transition to observation-level for faster execution.

Learn More

Was this page helpful?