Nóta
Teastaíonn údarú chun rochtain a fháil ar an leathanach seo. Is féidir leat triail a bhaint as shíniú isteach nó eolairí a athrú.
Teastaíonn údarú chun rochtain a fháil ar an leathanach seo. Is féidir leat triail a bhaint as eolairí a athrú.
Build, deploy, and manage AI and machine learning applications with Mosaic AI, an integrated platform that unifies the entire AI lifecycle from data preparation to production monitoring.
For a set of tutorials to get you started, see AI and machine learning tutorials.
Build generative AI applications
Develop and deploy enterprise-grade generative AI applications such as fine-tuned LLMs, AI agents, and retrieval-augmented generation.
Feature | Description |
---|---|
AI Playground | Prototype and test generative AI models with no-code prompt engineering and parameter tuning. |
Agent Bricks | Simple, no-code approach to build and optimize domain-specific, high-quality AI agent systems for common AI use cases. |
Foundation Models | Serve state-of-the-art LLMs including Meta Llama, Anthropic Claude, and OpenAI GPT through secure, scalable APIs. |
Mosaic AI Agent Framework | Build and deploy production-quality agents including RAG applications and multi-agent systems with Python. |
MLflow for GenAI | Measure, improve, and monitor quality throughout the GenAI application lifecycle using AI-powered metrics and comprehensive trace observability. |
Vector Search | Store and query embedding vectors with automatic syncing to your knowledge base for RAG applications. |
Foundation Model Fine-tuning | Customize foundation models with your own data to optimize performance for specific applications. |
Train classic machine learning models
Create machine learning models with automated tools and collaborative development environments.
Feature | Description |
---|---|
AutoML | Automatically build high-quality models with minimal code using automated feature engineering and hyperparameter tuning. |
Databricks Runtime for ML | Pre-configured clusters with TensorFlow, PyTorch, Keras, and GPU support for deep learning development. |
MLflow tracking | Track experiments, compare model performance, and manage the complete model development lifecycle. |
Feature engineering | Create, manage, and serve features with automated data pipelines and feature discovery. |
Databricks notebooks | Collaborative development environment with support for Python, R, Scala, and SQL for ML workflows. |
Train deep learning models
Use built-in frameworks to develop deep learning models.
Feature | Description |
---|---|
Distributed training | Examples of distributed deep learning using Ray, TorchDistributor, and DeepSpeed. |
Best practices for deep learning on Databricks | Best practices for deep learning on Databricks. |
PyTorch | Single-node and distributed training using PyTorch. |
TensorFlow | Single-node and distributed training using TensorFlow and TensorBoard. |
Reference solutions | Reference solutions for deep learning. |
Deploy and serve models
Deploy models to production with scalable endpoints, real-time inference, and enterprise-grade monitoring.
Feature | Description |
---|---|
Model Serving | Deploy custom models and LLMs as scalable REST endpoints with automatic scaling and GPU support. |
AI Gateway | Govern and monitor access to generative AI models with usage tracking, payload logging, and security controls. |
External models | Integrate third-party models hosted outside Databricks with unified governance and monitoring. |
Foundation model APIs | Access and query state-of-the-art open models hosted by Databricks. |
Monitor and govern ML systems
Ensure model quality, data integrity, and compliance with comprehensive monitoring and governance tools.
Feature | Description |
---|---|
Unity Catalog | Govern data, features, models, and functions with unified access control, lineage tracking, and discovery. |
Lakehouse Monitoring | Monitor data quality, model performance, and prediction drift with automated alerts and root cause analysis. |
MLflow for Models | Track, evaluate, and monitor generative AI applications throughout the development lifecycle. |
Productionize ML workflows
Scale machine learning operations with automated workflows, CI/CD integration, and production-ready pipelines.
Task | Description |
---|---|
Model Registry | Manage model versions, approvals, and deployments with centralized model lifecycle management. |
Lakeflow Jobs | Build automated workflows and production-ready ETL pipelines for ML data processing. |
Ray on Databricks | Scale ML workloads with distributed computing for large-scale model training and inference. |
MLOps workflows | Implement end-to-end MLOps with automated training, testing, and deployment pipelines. |
Git integration | Version control ML code and notebooks with seamless Git integration and collaborative development. |