This article explains the current limitations of serverless compute for notebooks and jobs. It starts with an overview of the most important considerations and then provides a comprehensive reference list of limitations.
Limitations overview
Before creating new workloads or migrating workloads to serverless compute, first consider the following limitations:
Python and SQL are the only supported languages.
Only Spark connect APIs are supported. Spark RDD APIs are not supported.
Serverless compute is available to all workspace users.
Notebook tags are not supported. Use budget policies to tag serverless usage.
For streaming, only incremental batch logic can be used. There is no support for default or time-based trigger intervals. See Streaming limitations.
Limitations reference list
The following sections list the current limitations of serverless compute.
Serverless compute is based on Databricks standard access mode compute architecture (formerly called shared access mode). The most relevant limitations inherited from standard access mode are listed below, along with additional serverless-specific limitations. For a full list of standard access mode limitations, see Compute access mode limitations for Unity Catalog.
General limitations
Scala and R are not supported.
ANSI SQL is the default when writing SQL. Opt-out of ANSI mode by setting spark.sql.ansi.enabled to false.
Spark RDD APIs are not supported.
Spark Context (sc), spark.sparkContext, and sqlContext are not supported.
Individual rows must not exceed the maximum size of 128MB.
The Spark UI is not available. Instead, use the query profile to view information about your Spark queries. See Query profile.
Spark logs are not available when using serverless notebooks and jobs. Users only have access to client-side application logs.
Cross-workspace access is allowed only if the workspaces are in the same region and the destination workspace does not have an IP ACL or front-end PrivateLink configured.
Global temporary views are not supported. Databricks recommends using session temporary views or creating tables where cross-session data passing is required.
Hive SerDe tables are not supported. Additionally, the corresponding LOAD DATA command which loads data into a Hive SerDe table is not supported. Using the command will result in an exception.
Support for data sources is limited to AVRO, BINARYFILE, CSV, DELTA, JSON, KAFKA, ORC, PARQUET, ORC, TEXT, and XML.
Hive variables (for example ${env:var}, ${configName}, ${system:var}, and spark.sql.variable) or config variable references using the ${var} syntax are not supported. Using Hive variables will result in an exception.
Azure HPC is a purpose-built cloud capability for HPC & AI workload, using leading-edge processors and HPC-class InfiniBand interconnect, to deliver the best application performance, scalability, and value. Azure HPC enables users to unlock innovation, productivity, and business agility, through a highly available range of HPC & AI technologies that can be dynamically allocated as your business and technical needs change. This learning path is a series of modules that help you get started on Azure HPC - you