Skip to main content

MLflow Tracing UI

GenAI Experiment Overview

The Overview tab in GenAI experiments provides comprehensive analytics and visualizations for your GenAI application traces. This tab is organized into three sub-tabs to help you monitor different aspects of your application.

All tabs include a time range selector and time unit selector to customize the granularity and range of the displayed data.

Usage

The Usage tab displays key metrics about your trace requests over time:

  • Requests: Shows the total number of trace requests, with an average reference line
  • Latency: Visualizes response time distribution to help identify performance bottlenecks
  • Errors: Tracks error rates to quickly spot issues
  • Token Usage & Token Stats: Monitors token consumption across your traces

Experiment Overview Usage Tab

Quality

The Quality tab provides insights into the quality of your GenAI outputs:

  • Quality Summary: Provides overview of scorers result
  • Quality Insights: Displays metrics computed by scorers, with a dedicated chart for each assessment type
  • Charts are dynamically generated based on the assessments available in your traces

Experiment Overview Quality Tab

Tool Calls

The Tool Calls tab provides insights into agent tool usage:

  • Statistics Cards: Shows at-a-glance metrics including total tool calls, average latency, success rate, and failed calls
  • Tool Performance Summary: Provides an overview of how each tool is performing
  • Tool Usage & Latency: Visualizes tool invocation patterns and response times
  • Tool Error Rate: Tracks error rates per tool

Experiment Overview Tool Calls Tab

Traces within MLflow Experiments

After logging your traces, you can view them in the MLflow UI, under the "Traces" tab in the main experiment page. This tab is also available within the individual run pages, if your trace was logged within a run context.

MLflow Tracking UI

This table includes high-level information about the traces, such as the trace ID, the inputs / outputs of the root span, and more. From this page, you can also perform a few actions to manage your traces:

Using the search bar in the UI, you can easily filter your traces based on name, tags, or other metadata. Check out the search docs for details about the query string format.

Browsing span data

In order to browse the span data of an individual trace, simply click on the link in the "Trace ID" or "Trace name" columns to open the trace viewer:

Jupyter Notebook integration

note

The MLflow Tracing Jupyter integration is available in MLflow 2.20 and above

You can also view the trace UI directly within Jupyter notebooks, allowing you to debug your applications without having to tab out of your development environment.

Jupyter Trace UI

This feature requires using an MLflow Tracking Server, as this is where the UI assets are fetched from. To get started, simply ensure that the MLflow Tracking URI is set to your tracking server (e.g. mlflow.set_tracking_uri("http://localhost:5000")).

By default, the trace UI will automatically be displayed for the following events:

  1. When the cell code generates a trace (e.g. via automatic tracing, or by running a manually traced function)
  2. When mlflow.search_traces() is called
  3. When a mlflow.entities.Trace() object is displayed (e.g. via IPython's display function, or when it is the last value returned in a cell)

To disable the display, simply call mlflow.tracing.disable_notebook_display(), and rerun the cell containing the UI. To enable it again, call mlflow.tracing.enable_notebook_display().