This article explains the current limitations of serverless compute for notebooks and jobs. It starts with an overview of the most important considerations and then provides a comprehensive reference list of limitations.
Limitations overviewâBefore creating new workloads or migrating workloads to serverless compute, first consider the following limitations:
The following sections list the current limitations of serverless compute.
Serverless compute is based on Databricks standard access mode compute architecture (formerly called shared access mode). The most relevant limitations inherited from standard access mode are listed below, along with additional serverless-specific limitations. For a full list of standard access mode limitations, see Compute access mode limitations for Unity Catalog.
General limitationsâspark.sql.ansi.enabled
to false
.spark.sparkContext
, and sqlContext
are not supported.spark.databricks.execution.timeout
property. For more details, see Configure Spark properties for serverless notebooks and jobs. This limit does not apply to serverless jobs.Trigger.AvailableNow
is supported. See Configure Structured Streaming trigger intervals..ipynb
format. If your notebook is saved in source format, serverless metadata might not be captured correctly, and some features might not function as expected.The following compute-specific features are not supported:
Dataframe and SQL cache APIs are not supported on serverless compute. Using any of these APIs or SQL commands will result in an exception.
Hive SerDe tables are not supported. Additionally, the corresponding LOAD DATA command which loads data into a Hive SerDe table is not supported. Using the command will result in an exception.
Support for data sources is limited to AVRO, BINARYFILE, CSV, DELTA, JSON, KAFKA, ORC, PARQUET, ORC, TEXT, and XML.
Hive variables (for example ${env:var}
, ${configName}
, ${system:var}
, and spark.sql.variable
) or config variable references using the ${var}
syntax are not supported. Using Hive variables will result in an exception.
Instead, use DECLARE VARIABLE, SET VARIABLE, and SQL session variable references and parameter markers ('?', or ':var') to declare, modify, and reference session state. You can also use the IDENTIFIER clause to parameterize object names in many cases.
Serverless compute supports the following data sources for DML operations (write, update, delete):
CSV
JSON
AVRO
DELTA
KAFKA
PARQUET
ORC
TEXT
UNITY_CATALOG
BINARYFILE
XML
SIMPLESCAN
ICEBERG
Serverless compute supports the following data sources for read operations:
CSV
JSON
AVRO
DELTA
KAFKA
PARQUET
ORC
TEXT
UNITY_CATALOG
BINARYFILE
XML
SIMPLESCAN
ICEBERG
MYSQL
POSTGRESQL
SQLSERVER
REDSHIFT
SNOWFLAKE
SQLDW
(Azure Synapse)DATABRICKS
BIGQUERY
ORACLE
SALESFORCE
SALESFORCE_DATA_CLOUD
TERADATA
WORKDAY_RAAS
MONGODB
RetroSearch is an open source project built by @garambo | Open a GitHub Issue
Search and Browse the WWW like it's 1997 | Search results from DuckDuckGo
HTML:
3.2
| Encoding:
UTF-8
| Version:
0.7.4