Merk
Tilgang til denne siden krever autorisasjon. Du kan prøve å logge på eller endre kataloger.
Tilgang til denne siden krever autorisasjon. Du kan prøve å endre kataloger.
Build, deploy, and manage AI and machine learning applications with Mosaic AI, an integrated platform that unifies the entire AI lifecycle from data preparation to production monitoring.
For a set of tutorials to get you started, see AI and machine learning tutorials.
Build generative AI applications
Develop and deploy enterprise-grade generative AI applications such as fine-tuned LLMs, AI agents, and retrieval-augmented generation.
| Feature | Description |
|---|---|
| AI Playground | Prototype and test generative AI models with no-code prompt engineering and parameter tuning. |
| Agent Bricks | Simple approach to build and optimize domain-specific, high-quality AI agent systems for common AI use cases. |
| Foundation Models | Serve state-of-the-art LLMs including Meta Llama, Anthropic Claude, and OpenAI GPT through secure, scalable APIs. |
| Mosaic AI Agent Framework | Build and deploy production-quality agents including RAG applications and multi-agent systems with Python. |
| MLflow for GenAI | Measure, improve, and monitor quality throughout the GenAI application lifecycle using AI-powered metrics and comprehensive trace observability. |
| Vector Search | Store and query embedding vectors with automatic syncing to your knowledge base for RAG applications. |
| Serverless GPU compute | Customize single and multi-node deep learning workloads for training and fine-tuning custom models using your favorite frameworks and get state-of-the-art efficiency, performance, and quality. |
| Foundation Model Fine-tuning | Customize foundation models with your own data to optimize performance for specific applications. |
Train classic machine learning models
Create machine learning models with automated tools and collaborative development environments.
| Feature | Description |
|---|---|
| AutoML | Automatically build high-quality models with minimal code using automated feature engineering and hyperparameter tuning. |
| Databricks Runtime for ML | Pre-configured clusters with TensorFlow, PyTorch, Keras, and GPU support for deep learning development. |
| MLflow tracking | Track experiments, compare model performance, and manage the complete model development lifecycle. |
| Feature engineering | Create, manage, and serve features with automated data pipelines and feature discovery. |
| Databricks notebooks | Collaborative development environment with support for Python, R, Scala, and SQL for ML workflows. |
Train deep learning models
Use built-in frameworks to develop deep learning models.
| Feature | Description |
|---|---|
| Distributed training | Examples of distributed deep learning using Ray, TorchDistributor, and DeepSpeed. |
| Best practices for deep learning on Databricks | Best practices for deep learning on Databricks. |
| PyTorch | Single-node and distributed training using PyTorch. |
| TensorFlow | Single-node and distributed training using TensorFlow and TensorBoard. |
| Reference solutions | Reference solutions for deep learning. |
Deploy and serve models
Deploy models to production with scalable endpoints, real-time inference, and enterprise-grade monitoring.
| Feature | Description |
|---|---|
| Model Serving | Deploy custom models and LLMs as scalable REST endpoints with automatic scaling and GPU support. |
| AI Gateway | Govern and monitor access to generative AI models with usage tracking, payload logging, and security controls. |
| External models | Integrate third-party models hosted outside Databricks with unified governance and monitoring. |
| Foundation model APIs | Access and query state-of-the-art open models hosted by Databricks. |
Monitor and govern ML systems
Ensure model quality, data integrity, and compliance with comprehensive monitoring and governance tools.
| Feature | Description |
|---|---|
| Unity Catalog | Govern data, features, models, and functions with unified access control, lineage tracking, and discovery. |
| Data profiling | Monitor data quality, model performance, and prediction drift with automated alerts and root cause analysis. |
| Anomaly detection | Monitor data freshness and completeness at the catalog level. |
| MLflow for Models | Track, evaluate, and monitor generative AI applications throughout the development lifecycle. |
Productionize ML workflows
Scale machine learning operations with automated workflows, CI/CD integration, and production-ready pipelines.
| Feature | Description |
|---|---|
| Models in Unity Catalog | Use the model registry in Unity Catalog for centralized governance and to manage the model lifecycle, including deployments. |
| Lakeflow Jobs | Build automated workflows and production-ready ETL pipelines for ML data processing. |
| Ray on Databricks | Scale ML workloads with distributed computing for large-scale model training and inference. |
| MLOps workflows | Implement end-to-end MLOps with automated training, testing, and deployment pipelines. |
| Git integration | Version control ML code and notebooks with seamless Git integration and collaborative development. |