IntegrationsFrameworksMicrosoft Agent Framework
This is a Jupyter notebook

Trace the Microsoft Agent Framework with Langfuse

This notebook demonstrates how to integrate Langfuse into your Microsoft Agent Framework workflow to monitor, debug and evaluate your AI agents.

What is the Microsoft Agent Framework?: The Microsoft Agent Framework is an open-source framework for building intelligent agents. It provides a comprehensive set of tools for creating agents that can interact with various services, execute tasks, and handle complex workflows. The framework supports multiple LLM providers including Azure OpenAI and OpenAI, and offers built-in observability through OpenTelemetry.

What is Langfuse?: Langfuse is an open-source observability platform for AI agents. It helps you visualize and monitor LLM calls, tool usage, cost, latency, and more.

1. Install Dependencies

Below we install the agent-framework library (the Microsoft Agent Framework) and langfuse for observability.

%pip install agent-framework langfuse --pre

2. Configure Environment & Langfuse Credentials

Next, set up your Langfuse API keys. You can get these keys by signing up for a free Langfuse Cloud account or by self-hosting Langfuse. These environment variables are essential for the Langfuse client to authenticate and send data to your Langfuse project.

import os
 
# Get keys for your project from the project settings page: https://cloud.langfuse.com
os.environ["LANGFUSE_PUBLIC_KEY"] = "pk-lf-..." 
os.environ["LANGFUSE_SECRET_KEY"] = "sk-lf-..." 
os.environ["LANGFUSE_HOST"] = "https://cloud.langfuse.com" # 🇪🇺 EU region
# os.environ["LANGFUSE_HOST"] = "https://us.cloud.langfuse.com" # 🇺🇸 US region
 
# Your Azure OpenAI credentials
os.environ["AZURE_OPENAI_API_KEY"] = "your-azure-openai-key"
os.environ["AZURE_OPENAI_ENDPOINT"] = "https://your-resource.openai.azure.com/"
os.environ["AZURE_OPENAI_CHAT_DEPLOYMENT_NAME"] = "gpt-5-mini"
os.environ["OPENAI_CHAT_MODEL_ID"] = "gpt-5-mini"

3. Initialize Langfuse Client

Initialize the Langfuse client to verify the connection. get_client() initializes the Langfuse client using the credentials provided in the environment variables.

from langfuse import get_client
 
langfuse = get_client()
 
# Verify connection
if langfuse.auth_check():
    print("Langfuse client is authenticated and ready!")
else:
    print("Authentication failed. Please check your credentials and host.")

4. Enable Observability

The Microsoft Agent Framework includes built-in observability support through OpenTelemetry. Enable it by calling setup_observability() which automatically exports traces to Langfuse.

ℹ️

Note: Set enable_sensitive_data=True to capture full request/response data including function arguments and results.

from agent_framework.observability import setup_observability
 
setup_observability(enable_sensitive_data=True)

5. Hello World Example with Tool

Below we create a weather agent using the Microsoft Agent Framework with Azure OpenAI. The agent has access to a get_weather function tool that it can call to retrieve weather information.

import asyncio
from random import randint
from typing import Annotated
from agent_framework.azure import AzureOpenAIChatClient
from pydantic import Field
 
def get_weather(
    location: Annotated[str, Field(description="The location to get the weather for.")],
) -> str:
    """Get the weather for a given location."""
    conditions = ["sunny", "cloudy", "rainy", "stormy"]
    return f"The weather in {location} is {conditions[randint(0, 3)]} with a high of {randint(10, 30)}°C."
 
async def main():
    # Create an agent with Azure OpenAI
    async with AzureOpenAIChatClient().create_agent(
        instructions="You are a helpful weather agent.",
        tools=get_weather,
    ) as agent:
        query = "What's the weather like in Seattle?"
        print(f"User: {query}")
        result = await agent.run(query)
        print(f"Agent: {result}\n")
 
# Run the agent
await main()

6. Using OpenAI Directly

The Microsoft Agent Framework also supports using OpenAI directly (not through Azure). Simply use OpenAIResponsesClient instead of AzureOpenAIChatClient.

# Required for OpenAI API access
os.environ["OPENAI_API_KEY"]="sk-proj-..."
os.environ["OPENAI_RESPONSES_MODEL_ID"]="gpt-5-mini" 
from agent_framework.openai import OpenAIResponsesClient
 
async def main():
    async with OpenAIResponsesClient().create_agent(
        instructions="You are a helpful assistant.",
        tools=get_weather,
    ) as agent:
        query = "What's the weather in Tokyo?"
        print(f"User: {query}")
        result = await agent.run(query)
        print(f"Agent: {result}\n")
 
await main()

7. See Traces in Langfuse

Example trace in Langfuse

Example Trace in Langfuse

Was this page helpful?