How to trace your application with prompt flow SDK | Azure AI Studio

Important

Some of the features described in this article might only be available in preview. This preview is provided without a service-level agreement, and we don't recommend it for production workloads. Certain features might not be supported or might have constrained capabilities. For more information, see Supplemental Terms of Use for Microsoft Azure Previews.

Tracing is a powerful tool that offers developers an in-depth understanding of the execution process of their generative AI applications such as agents, AutoGen, and retrieval augmented generation (RAG) use cases. It provides a detailed view of the execution flow, including the inputs and outputs of each node within the application. This essential information proves critical while debugging complex applications or optimizing performance.

While more developers are using various frameworks such as Langchain, Semantic Kernel, OpenAI, and various types of agents to create LLM-based applications. Tracing with the prompt flow SDK offers enhanced visibility and simplified troubleshooting for LLM-based applications, effectively supporting development, iteration, and production monitoring. Tracing in AI Studio follows the OpenTelemetry specification, capturing and visualizing the internal execution details of any AI application, enhancing the overall development experience.

Benefits of AI Studio tracing on the enterprise-grade cloud platform

Moreover, we now offer persistent local testing on AI Studio, which is the enterprise-grade cloud platform, significantly enhancing collaboration, persistence, and test history management.

With tracing, you can:

  • Have a cloud-based location to persist and track your historical tests.
  • Easily extract and visualize the test results, comparing the outputs of different test cases.
  • Reuse your previous test assets for later usage, for example, human feedback, data curation, etc.
  • Facilitate better resource utilization in the future.
  • Debug and optimize your application with ease. To get started with debugging LLM application scenarios, refer to Tracing with LLM application
  • Analyze retrieval and generation processes in RAG applications.
  • Observe the multi-agents interactions in the multi-agent scenarios. To get started with tracing in multi-agent scenarios, refer to Tracing with AutoGen.

Log and view traces of your applications

AI Studio provides the tracing capability for logging and managing your LLM applications tests and evaluations, while debugging and observing by drilling down the trace view.

The tracing any application feature today is implemented in the prompt flow open-source package, to enable user to trace LLM call or function, and LLM frameworks like LangChain and AutoGen, regardless of which framework you use, following OpenTelemetry specification.

Enable trace in your application

Code first - Make sure you have annotated your code for tracing in prompt flow!

More details about tracing in prompt flow, please refer to this prompt flow documentation.

Log the trace to AI Studio

Set the trace destination

By default, the trace is logged and viewed in your local environment. To log it in the AI Studio in the cloud, you need to set the trace destination to a specified AI Studio project.

You can refer to the following steps to set the trace destination to AI Studio project.

First, ensure that Azure CLI is installed and logged in:

az login

Next, execute the following command to set the trace destination. Replace <your_subscription_id>, <your_resourcegroup_name>, and <your_studio_project_name> with your specific subscription ID, resource group name, and AI Studio project name:

pf config set trace.destination=azureml://subscriptions/<your_subscription_id>/resourcegroups/<your_resourcegroup_name>/providers/Microsoft.MachineLearningServices/workspaces/<your_studio_project_name>

Note

The West US 3 (westus3) region does not support tracing.

Collections

A collection is a group of associated traces. In AI Studio, these collections along with their internal traces are managed and stored in the Tracing module under the Collections tab.

  1. Go to your project in AI Studio.

  2. From the left pane, select Tracing. You can see the Collections tab. You can only see your own collections in the list of collections. In this example, there aren't any collections yet.

    Screenshot of the button to add a new connection.

The collection tab displays a comprehensive list of all the collections you've created. It shows essential metadata for each collection, including its name, run location, last updated time, and created time.

  • Run location: Indicates whether the application runs locally or in the cloud. The cloud collection is associated with specific prompt flow cloud authoring test history and generated traces. In this case, the collection name is the same as the prompt flow display name.
  • Updated on: Shows the most recent time a new trace was logged to a collection. By default, collections are sorted in descending order based on their updated times.
  • Created on: The time when the collection was initially created.

By selecting a collection's name, you can access a list of all the traces within that collection. Only a subset of traces can be shared with others. Refer to share trace for more information.

When logging a trace, you have the option to specify a collection name to group it with other related traces. You can create multiple collections for better organization of your traces. If a collection name isn't specified when logging a trace, it defaults to the project folder name or to the default collection.

Customize the collections

For better organization of your traces, you can specify a custom collection name when logging a trace.

If you're tracing your own application, you can set the collection name in the start_trace() function in your code:

from promptflow.tracing import start_trace, trace

@trace
def my_function(input: str) -> str:
    output = input + "Hello World!"
    return output

my_function("This is my function")
start_trace(collection="my_custom_collection")

More details about customizing collections, please refer to tracing tutorial and prompt flow command.

View the traces

First, you must complete the previous steps to view the traces in the cloud:

Now, run your python script directly. Upon successful execution, a cloud trace link appears in the output. It might look something like this:

Starting prompt flow service...
...
You can view the traces in cloud from AI Studio: https://ai.azure.com/projecttrace/detail/....

Selecting the URL to navigate to a trace detail page on the cloud portal. This page is similar to the local trace view.

The trace detail view provides a comprehensive and structured overview of the operations within your application.

Understand the trace detail view

In the top right corner of the trace view, you find:

  • Trace name: This is same as the root span name, representing the entry function name of your application.
  • Status: This could either be "completed" or "failed".
  • Total duration: This is total duration time of the test execution. Hover over to view the start and end times.
  • Total tokens: This is the total token cost of the test. Hover over to view the prompt tokens and completed tokens.
  • Created time: The time at which the trace was created.

On the left side, you can see a hierarchical tree structure. This structure shows the sequence of function calls. Each function call's metadata is organized into spans. These spans are linked together in a tree-like structure, illustrating the sequence of execution.

In prompt flow SDK, we defined several span types, including LLM, Function, Embedding, Retrieval, and Flow. And the system automatically creates spans with execution information in designated attributes and events.

Each span allows you to view:

  • Function name: By default, this is the name of the function as defined in your code. However, it can also be a customized span name defined via Open Telemetry.
  • Duration: This represents the duration for which the function ran. Hover over to view the start and end times.
  • Tokens for LLM calls: This is the token cost of the LLM call. Hover over to view the prompt tokens and completed tokens.

By selecting a specific span, you can shift to viewing its associated detailed information on the right side. This information includes input, output, raw JSON, logs, and exceptions, which are essential for observing and debugging your application.

For the LLM span, a clear conversation view is provided. This includes the system prompt, user prompt, and assistant response. This information is especially crucial in multi-agent cases, as it allows you to understand the flow of the conversation and the interaction within the LLM intermediate auto-calling.

You can select the raw JSON tab to view the json data of the span. This format might be more suitable for developers when it comes to debugging and troubleshooting.

Share trace

If you want to share the trace with others who has the project permission, you can select the Share button on the right corner of the trace detail page, then you have the page link copied to share with others.

Note

The shared trace is read-only, and only the people who has the project permission can view it via the link.