Azure Databricks documentation archive
Important
This documentation has been retired and might not be updated. The products, services, or technologies mentioned in this content are no longer supported.
In this archive, you can find earlier versions of documentation for Azure Databricks products, features, APIs, and workflows.
Compute
- Create cluster UI (legacy)
- Cluster UI preview
- Install a library with an init script (legacy)
- Cluster-named init scripts (legacy)
- Global init scripts (legacy)
Dev tools
- Manage libraries with
%conda
commands (legacy) - Explore and create tables in DBFS
- Transactional writes to cloud storage with DBIO
- Koalas
- Legacy UniForm IcebergCompatV1
- Workspace libraries (legacy)
- Databricks CLI (legacy)
- What is dbx by Databricks Labs?
- dbutils.library
- Migrate to Spark 3.x
- VScode with Git folders
Governance
- External metastores (legacy)
- Create Unity Catalog managed storage using a service principal (legacy)
- Credential passthrough (legacy)
Machine learning and AI
- Optimized LLM serving
- Model serving (legacy)
- Serverless real-time inference (preview)
- Share feature tables across workspaces (legacy)
- MLeap ML model export
- Train a PySpark model and save in MLeap format
- Set up and considerations for
ai_generate_text()
- Analyze customer reviews with
ai_generate_text()
and OpenAI
Release notes
Repos and Git source control
Security
Storage
- Azure Blob storage file source with Azure Queue Storage (legacy)
- Azure Cosmos DB
- Azure Event Hubs
- Connecting Azure Databricks and Azure Synapse with PolyBase (legacy)
- Neo4j
- Read and write XML data using the
spark-xml
library - Accessing Azure Data Lake Storage Gen1 from Azure Databricks
- Configure Delta storage credentials
- Connect to Azure Blob Storage with WASB (legacy)
Feedback
https://aka.ms/ContentUserFeedback.
Coming soon: Throughout 2024 we will be phasing out GitHub Issues as the feedback mechanism for content and replacing it with a new feedback system. For more information see:Submit and view feedback for