Describe MLflow deployment capabilities
MLflow is a comprehensive open-source platform that facilitates the entire machine learning lifecycle, including experimentation, reproducibility, and deployment.
When integrated with Azure Databricks, MLflow provides a robust and scalable environment for managing machine learning models from development to production.
Create endpoints to serve your models
You can use MLflow's Deployment SDK to create endpoints that deploy and serve LLMs.
MLflow is integrated with Azure Databricks' Model Serving feature, which allows you to manage your deployed LLMs. When you want to work with LLMs, you can start experimenting with one of the available foundation models in Azure Databricks:
Alternatively, you can create a new serving endpoint to deploy another LLM you want to work with. You can create an endpoint using MLflow's Deployment SDK, or use the UI where you have to specify which language model you want to create an endpoint for.
For example, you add a serving entity by connecting to an OpenAI model like GPT-4:
Once you create an endpoint, you can review all metadata and use the URL to interact with the deployed LLM:
Tip
Learn more about how to create generative AI model serving endpoints.