Del via


Run your Azure Databricks job with serverless compute for workflows

Important

Because serverless compute for workflows does not support controlling egress traffic, your jobs have full access to the internet.

Serverless compute for workflows allows you to run your Azure Databricks job without configuring and deploying infrastructure. With serverless compute, you focus on implementing your data processing and analysis pipelines, and Azure Databricks efficiently manages compute resources, including optimizing and scaling compute for your workloads. Autoscaling and Photon are automatically enabled for the compute resources that run your job.

Serverless compute for workflows auto-optimization automatically optimizes compute by selecting appropriate resources such as instance types, memory, and processing engines based on your workload. Auto-optimization also automatically retries failed tasks.

Databricks automatically upgrades the Databricks Runtime version to support enhancements and upgrades to the platform while ensuring the stability of your Azure Databricks jobs. To see the current Databricks Runtime version used by serverless compute for workflows, see Serverless compute release notes.

Because cluster creation permission is not required, all workspace users can use serverless compute to run their workflows.

This article describes using the Azure Databricks Jobs UI to create and run jobs that use serverless compute. You can also automate creating and running jobs that use serverless compute with the Jobs API, Databricks Asset Bundles, and the Databricks SDK for Python.

Requirements

  • Your Azure Databricks workspace must have Unity Catalog enabled.

  • Because serverless compute for workflows uses shared access mode, your workloads must support this access mode.

  • Your Azure Databricks workspace must be in a supported region. See Feature availability.

  • Your Azure Databricks account must have serverless compute enabled. See Enable serverless compute.

Create a job using serverless compute

Note

Because serverless compute for workflows ensures that sufficient resources are provisioned to run your workloads, you might experience increased startup times when running a Azure Databricks job that requires large amounts of memory or includes many tasks.

Serverless compute is supported with the notebook, Python script, dbt, and Python wheel task types. By default, serverless compute is selected as the compute type when you create a new job and add one of these supported task types.

Create serverless task

Databricks recommends using serverless compute for all job tasks. You can also specify different compute types for tasks in a job, which might be required if a task type is not supported by serverless compute for workflows.

Configure an existing job to use serverless compute

You can switch an existing job to use serverless compute for supported task types when you edit the job. To switch to serverless compute, either:

  • In the Job details side panel click Swap under Compute, click New, enter or update any settings, and click Update.
  • Click Down Caret in the Compute drop-down menu and select Serverless.

Switch task to serverless compute

Schedule a notebook using serverless compute

In addition to using the Jobs UI to create and schedule a job using serverless compute, you can create and run a job that uses serverless compute directly from a Databricks notebook. See Create and manage scheduled notebook jobs.

Select a budget policy for your serverless usage

Important

This feature is in Public Preview.

Budget policies allow your organization to apply custom tags on serverless usage for granular billing attribution.

If your workspace uses budget policies to attribute serverless usage, you can select your job’s budget policy using the Budget policy setting in the job details UI. If you are only assigned to one budget policy, the policy is automatically selected for your new jobs.

Note

After you’ve been assigned a budget policy, your existing jobs are not automatically tagged with your policy. You must manually update existing jobs if you want to attach a policy to them.

For more on budget policies, see Attribute serverless usage with budget policies.

Set Spark configuration parameters

To automate the configuration of Spark on serverless compute, Databricks allows setting only specific Spark configuration parameters. For the list of allowable parameters, see Supported Spark configuration parameters.

You can set Spark configuration parameters at the session level only. To do this, set them in a notebook and add the notebook to a task included in the same job that uses the parameters. See Get and set Apache Spark configuration properties in a notebook.

Configure environments and dependencies

To learn how to install libaries and dependencies using serverless compute, see Install notebook dependencies.

Configure serverless compute auto-optimization to disallow retries

Serverless compute for workflows auto-optimization automatically optimizes the compute used to run your jobs and retries failed tasks. Auto-optimization is enabled by default, and Databricks recommends leaving it enabled to ensure critical workloads run successfully at least once. However, if you have workloads that must be executed at most once, for example, jobs that are not idempotent, you can turn off auto-optimization when adding or editing a task:

  1. Next to Retries, click Add (or Edit Icon if a retry policy already exists).
  2. In the Retry Policy dialog, uncheck Enable serverless auto-optimization (may include additional retries).
  3. Click Confirm.
  4. If you’re adding a task, click Create task. If you’re editing a task, click Save task.

Monitor the cost of jobs that use serverless compute for workflows

You can monitor the cost of jobs that use serverless compute for workflows by querying the billable usage system table. This table is updated to include user and workload attributes about serverless costs. See Billable usage system table reference.

View details for your Spark queries

Serverless compute for workflows has a new interface for viewing detailed runtime information for your Spark statements, such as metrics and query plans. To view query insights for Spark statements included in your jobs run on serverless compute:

  1. Click Workflows Icon Workflows in the sidebar.
  2. In the Name column, click the job name you want to view insights for.
  3. Click the specific run you want to view insights for.
  4. In the Compute section of the Task run side panel, click Query history.
  5. You are redirected to the Query History, prefiltered based on the task run ID of the task you were in.

For information on using query history, see Query history.

Limitations

For a list of serverless compute for workflows limitations, see Serverless compute limitations in the serverless compute release notes.