Share via


AI and machine learning on Databricks

Build, deploy, and manage AI and machine learning applications with Mosaic AI, an integrated platform that unifies the entire AI lifecycle from data preparation to production monitoring.

For a set of tutorials to get you started, see AI and machine learning tutorials.

Build generative AI applications

Develop and deploy enterprise-grade generative AI applications such as fine-tuned LLMs, AI agents, and retrieval-augmented generation.

Feature Description
AI Playground Prototype and test generative AI models with no-code prompt engineering and parameter tuning.
Agent Bricks Simple, no-code approach to build and optimize domain-specific, high-quality AI agent systems for common AI use cases.
Foundation Models Serve state-of-the-art LLMs including Meta Llama, Anthropic Claude, and OpenAI GPT through secure, scalable APIs.
Mosaic AI Agent Framework Build and deploy production-quality agents including RAG applications and multi-agent systems with Python.
MLflow for GenAI Measure, improve, and monitor quality throughout the GenAI application lifecycle using AI-powered metrics and comprehensive trace observability.
Vector Search Store and query embedding vectors with automatic syncing to your knowledge base for RAG applications.
Foundation Model Fine-tuning Customize foundation models with your own data to optimize performance for specific applications.

Train classic machine learning models

Create machine learning models with automated tools and collaborative development environments.

Feature Description
AutoML Automatically build high-quality models with minimal code using automated feature engineering and hyperparameter tuning.
Databricks Runtime for ML Pre-configured clusters with TensorFlow, PyTorch, Keras, and GPU support for deep learning development.
MLflow tracking Track experiments, compare model performance, and manage the complete model development lifecycle.
Feature engineering Create, manage, and serve features with automated data pipelines and feature discovery.
Databricks notebooks Collaborative development environment with support for Python, R, Scala, and SQL for ML workflows.

Train deep learning models

Use built-in frameworks to develop deep learning models.

Feature Description
Distributed training Examples of distributed deep learning using Ray, TorchDistributor, and DeepSpeed.
Best practices for deep learning on Databricks Best practices for deep learning on Databricks.
PyTorch Single-node and distributed training using PyTorch.
TensorFlow Single-node and distributed training using TensorFlow and TensorBoard.
Reference solutions Reference solutions for deep learning.

Deploy and serve models

Deploy models to production with scalable endpoints, real-time inference, and enterprise-grade monitoring.

Feature Description
Model Serving Deploy custom models and LLMs as scalable REST endpoints with automatic scaling and GPU support.
AI Gateway Govern and monitor access to generative AI models with usage tracking, payload logging, and security controls.
External models Integrate third-party models hosted outside Databricks with unified governance and monitoring.
Foundation model APIs Access and query state-of-the-art open models hosted by Databricks.

Monitor and govern ML systems

Ensure model quality, data integrity, and compliance with comprehensive monitoring and governance tools.

Feature Description
Unity Catalog Govern data, features, models, and functions with unified access control, lineage tracking, and discovery.
Lakehouse Monitoring Monitor data quality, model performance, and prediction drift with automated alerts and root cause analysis.
MLflow for Models Track, evaluate, and monitor generative AI applications throughout the development lifecycle.

Productionize ML workflows

Scale machine learning operations with automated workflows, CI/CD integration, and production-ready pipelines.

Task Description
Model Registry Manage model versions, approvals, and deployments with centralized model lifecycle management.
Lakeflow Jobs Build automated workflows and production-ready ETL pipelines for ML data processing.
Ray on Databricks Scale ML workloads with distributed computing for large-scale model training and inference.
MLOps workflows Implement end-to-end MLOps with automated training, testing, and deployment pipelines.
Git integration Version control ML code and notebooks with seamless Git integration and collaborative development.