Monitor Apache Spark capacity consumption

The purpose of this article is to offer guidance for admins who want to monitor activities in the capacities they manage. By utilizing the Apache Spark capacity consumption reports available in the Microsoft Fabric Capacity Metrics app, admins can gain insights into the billable Spark capacity consumption for items, including Lakehouse, Notebook, and Apache Spark job definitions. Some Spark capacity consumption activities aren't reported in the app.

Spark capacity consumption reported

The following operations from lakehouses, notebooks, and Spark job definitions are treated as billable activities.

Operation name Item Comments
Lakehouse operations Lakehouse Users preview table in the Lakehouse explorer.
Lakehouse table load Lakehouse Users load delta table in the Lakehouse explorer.
Notebook run Synapse Notebook Synapse Notebook runs manually by users.
Notebook HC run Synapse Notebook Synapse Notebook runs under the high concurrency Apache Spark session.
Notebook scheduled run Synapse Notebook Synapse Notebook runs triggered by notebook scheduled events.
Notebook pipeline run Synapse Notebook Synapse Notebook runs triggered by pipeline.
Notebook VS Code run Synapse Notebook Synapse Notebook runs in VS Code.
Spark job run Spark Job Definition Spark batch job runs initiated by user submission.
Spark job scheduled run Spark Job Definition Synapse batch job runs triggered by notebook scheduled events.
Spark job pipeline run Spark Job Definition Synapse batch job runs triggered by pipeline.
Spark job VS Code run Spark Job Definition Synapse Spark job definition submitted from VS Code.

Spark capacity consumption that isn't reported

There are some Spark capacity consumption activities that aren't reported in the metrics app. These activities include system Spark jobs for library management and certain system Spark jobs for Spark Live pool or live sessions.

  • Library management - The capacity consumption associated with library management at the workspace level isn't reported in the metrics app.

  • System Spark jobs - Spark capacity consumption that isn't associated with a notebook, a Spark job definition, or a lakehouse, isn't included in the capacity reporting.

Capacity consumption reports

All Spark related operations are classified as background operations. Capacity consumption from Spark is displayed under a notebook, a Spark job definition, or a lakehouse, and is aggregated by operation name and item.

Screenshot showing items report.

Background operations report

Background operations are displayed for a specific timepoint. In the report's table, each row refers to a user operation. Review the User column to identify who performed a specific operation. If you need more information about a specific operation, you can use its Operation ID to look it up in the Microsoft Fabric monitoring hub.

Screenshot showing background operations report.