Docs
Integrations
Mirascope
Tracing

Mirascope Integration

Mirascope (GitHub (opens in a new tab)) is a Python toolkit for building LLM applications.

Developing LLM-applications with Mirascope feels just like writing the Python code you’re already used to. Python Toolkit for LLMs: Mirascope simplifies the development of applications using Large Language Models (LLMs) by providing an intuitive interface similar to standard Python coding practices.

The Mirascope team recorded a video with a short demo of how to use Mirascope with Langfuse.

Mirascope Integration Demo

How to use Mirascope with Langfuse

from mirascope.langfuse import with_langfuse

Mirascope automatically passes the Langfuse observe() decorator to all relevant functions within Mirascope via its with_langfuse decorator.

Example

Call

This is a basic call example but will work with all of Mirascope's call functions, call, stream, call_async, stream_async.

import os
from mirascope.langfuse import with_langfuse
from mirascope.anthropic import AnthropicCall
 
os.environ["LANGFUSE_SECRET_KEY"] = "sk-lf-..."
os.environ["LANGFUSE_PUBLIC_KEY"] = "pk-lf-..."
os.environ["LANGFUSE_HOST"] = "https://cloud.langfuse.com"
 
 
@with_langfuse
class BookRecommender(AnthropicCall):
    prompt_template = "Please recommend some {genre} books"
 
    genre: str
 
 
recommender = BookRecommender(genre="fantasy")
response = recommender.call()  # this will automatically get logged with langfuse
print(response.content)
#> Here are some recommendations for great fantasy book series: ...

This will give you:

  • A trace around the AnthropicCall.call() that captures items like the prompt template, and input/output attributes and more.
  • Human-readable display of the conversation with the agent
  • Details of the response, including the number of tokens used

Mirascope Trace call

Extract

Mirascope's extraction functionality is built on top of Pydantic and offers a convenient extract method on extractor classes to extract structured information from LLM outputs. This method leverages tools (function calling) to reliably extract the required structured data. (Docs (opens in a new tab))

import os
from mirascope.langfuse import with_langfuse
from mirascope.openai import OpenAIExtractor
 
os.environ["LANGFUSE_SECRET_KEY"] = "sk-lf-..."
os.environ["LANGFUSE_PUBLIC_KEY"] = "pk-lf-..."
os.environ["LANGFUSE_HOST"] = "https://cloud.langfuse.com"
 
class TaskDetails(BaseModel):
    description: str
    due_date: str
    priority: Literal["low", "normal", "high"]
 
 
@with_langfuse
class TaskExtractor(OpenAIExtractor[TaskDetails]):
    extract_schema: Type[TaskDetails] = TaskDetails
    prompt_template = """
    Extract the task details from the following task:
    {task}
    """
 
    task: str
 
 
task = "Submit quarterly report by next Friday. Task is high priority."
task_details = TaskExtractor(
    task=task
).extract()  # this will be logged automatically with langfuse
assert isinstance(task_details, TaskDetails)
print(task_details)
# > description='Submit quarterly report' due_date='next Friday' priority='high'

This will give you the same view as you would get from using the OpenAI integration. We will be adding more extraction support for other providers soon.

Mirascope Trace extract

Learn more

Was this page useful?

Questions? We're here to help

Subscribe to updates