Langfuse — Open source product analytics for LLM applications
This is a cross-post of our Launch YC (W23) which explains why we're building Langfuse.
TLDR: Langfuse is building open source product analytics (think 'Mixpanel') for LLM apps. We help companies to track and analyze quality, cost and latency across product releases and use cases.
we're Max, Marc and Clemens. We were part of the Winter 23 batch and work on Langfuse, where we help teams make sense of how their LLM applications perform.
LLMs represent a new paradigm in software. Single LLM calls are probabilistic and add substantial latency and cost. Applications use LLMs in new ways via advanced prompting, embedding-based retrieval, chains, and agents with tools. Teams building production-grade LLM applications have new product analytics and monitoring needs:
- Quality of outputs is difficult to measure. Outputs can e.g. be inaccurate, unhelpful, poorly formatted, hallucinated or error.
- Cost of compute is a priority again given high inference costs.
- Latency of responses matters for synchronous use cases.
- Debugging is challenging due to increasingly complex LLM applications (chains, agents, tool usage).
- Understanding user behavior is difficult given open-ended user prompts and conversational interactions.
Langfuse derives actionable insights from production data. Our customers use Langfuse to answer questions such as: "How helpful are my LLM app's outputs? What is my LLM API spend by customer? What do latencies look like across geographies and steps of LLM chains? Did the quality of the application improve in newer versions? What was the impact of switching from zero-shotting GPT4 to using few-shotted Llama calls?"
- Quality is measured through user feedback, model-based scoring and human-in-the-loop scored samples. Quality is assessed over time as well as across prompt versions, LLMs and users.
- Cost and Latency are accurately measured and broken down by user, session, geography, feature, model and prompt version.
- Monitor quality/cost/latency tradeoffs by release to facilitate product and engineering decisions.
- Cluster use cases by employing a classifier to understand what users are doing.
- Break down LLM usage by customer for usage-based billing and profitability analysis.
- Python and Typescript SDKs to easily monitor complex LLM apps
- Frontend SDK to directly capture feedback from users as a quality signal
Langfuse can be self-hosted or used with a generous free tier in our managed cloud version.
Based on the ingested data, Langfuse helps developers debug complex LLM apps in production:
- Inspect LLM app executions in a nested UI for chains, agents and tool usage.
- Segment by user feedback to find the root cause of quality problems.
- Star us on GitHub (opens in a new tab) + follow along on Twitter (opens in a new tab) & LinkedIn (opens in a new tab).
- If you run an LLM app, go ahead and talk to us (opens in a new tab), we'd love to see how we can be helpful.
- Please forward Langfuse to teams building commercial LLM applications.
- Launch YC (opens in a new tab)
- LinkedIn post (opens in a new tab)
- Twitter thread (opens in a new tab)