az ml model
Note
This reference is part of the azure-cli-ml extension for the Azure CLI (version 2.0.28 or higher). The extension will automatically install the first time you run an az ml model command. Learn more about extensions.
Manage machine learning models.
Commands
Name | Description | Type | Status |
---|---|---|---|
az ml model delete |
Delete a model from the workspace. |
Extension | GA |
az ml model deploy |
Deploy model(s) from the workspace. |
Extension | GA |
az ml model download |
Download a model from the workspace. |
Extension | GA |
az ml model list |
List models in the workspace. |
Extension | GA |
az ml model package |
Package a model in the workspace. |
Extension | GA |
az ml model profile |
Profile model(s) in the workspace. |
Extension | GA |
az ml model register |
Register a model to the workspace. |
Extension | GA |
az ml model show |
Show a model in the workspace. |
Extension | GA |
az ml model update |
Update a model in the workspace. |
Extension | GA |
az ml model delete
Delete a model from the workspace.
az ml model delete --model-id
[--path]
[--resource-group]
[--subscription-id]
[--workspace-name]
[-v]
Required Parameters
ID of model to delete.
Optional Parameters
Path to a project folder. Default: current directory.
Resource group corresponding to the provided workspace.
Specifies the subscription Id.
Name of the workspace.
Verbosity flag.
Global Parameters
Increase logging verbosity to show all debug logs.
Show this help message and exit.
Only show errors, suppressing warnings.
Output format.
JMESPath query string. See http://jmespath.org/ for more information and examples.
Name or ID of subscription. You can configure the default subscription using az account set -s NAME_OR_ID
.
Increase logging verbosity. Use --debug for full debug logs.
az ml model deploy
Deploy model(s) from the workspace.
az ml model deploy --name
[--ae]
[--ai]
[--ar]
[--as]
[--at]
[--autoscale-max-replicas]
[--autoscale-min-replicas]
[--base-image]
[--base-image-registry]
[--cc]
[--ccl]
[--cf]
[--collect-model-data]
[--compute-target]
[--compute-type]
[--cuda-version]
[--dc]
[--description]
[--dn]
[--ds]
[--ed]
[--eg]
[--entry-script]
[--environment-name]
[--environment-version]
[--failure-threshold]
[--gb]
[--gbl]
[--gc]
[--ic]
[--id]
[--key-name]
[--key-version]
[--kp]
[--ks]
[--lo]
[--max-request-wait-time]
[--model]
[--model-metadata-file]
[--namespace]
[--no-wait]
[--nr]
[--overwrite]
[--path]
[--period-seconds]
[--pi]
[--po]
[--property]
[--replica-max-concurrent-requests]
[--resource-group]
[--rt]
[--sc]
[--scoring-timeout-ms]
[--sd]
[--se]
[--sk]
[--sp]
[--st]
[--subnet-name]
[--subscription-id]
[--tag]
[--timeout-seconds]
[--token-auth-enabled]
[--tp]
[--vault-base-url]
[--version-name]
[--vnet-name]
[--workspace-name]
[-v]
Required Parameters
The name of the service deployed.
Optional Parameters
Whether or not to enable key auth for this Webservice. Defaults to False.
Whether or not to enable AppInsights for this Webservice. Defaults to False.
How often the autoscaler should attempt to scale this Webservice. Defaults to 1.
Whether or not to enable autoscaling for this Webservice. Defaults to True if num_replicas is None.
The target utilization (in percent out of 100) the autoscaler should attempt to maintain for this Webservice. Defaults to 70.
The maximum number of containers to use when autoscaling this Webservice. Defaults to 10.
The minimum number of containers to use when autoscaling this Webservice. Defaults to 1.
A custom image to be used as base image. If no base image is given then the base image will be used based off of given runtime parameter.
Image registry that contains the base image.
The number of cpu cores to allocate for this Webservice. Can be a decimal. Defaults to 0.1.
The max number of CPU cores this Webservice is allowed to use. Can be a decimal.
Path to local file containing a conda environment definition to use for the image.
Whether or not to enable model data collection for this Webservice. Defaults to False.
Name of compute target. Only applicable when deploying to AKS.
Compute type of service to deploy.
Version of CUDA to install for images that need GPU support. The GPU image must be used on Microsoft Azure Services such as Azure Container Instances, Azure Machine Learning Compute, Azure Virtual Machines, and Azure Kubernetes Service. Supported versions are 9.0, 9.1, and 10.0. If 'enable_gpu' is set, this defaults to '9.1'.
Path to a JSON or YAML file containing deployment metadata.
Description of the service deployed.
The dns name for this Webservice.
Path to local file containing additional Docker steps to run when setting up image.
Directory for Azure Machine Learning Environment for deployment. It is the same directory path as provided in 'az ml environment scaffold' command.
Whether or not to enable GPU support in the image. The GPU image must be used on Microsoft Azure Services such as Azure Container Instances, Azure Machine Learning Compute, Azure Virtual Machines, and Azure Kubernetes Service. Defaults to False.
Path to local file that contains the code to run for service (relative path from source_directory if one is provided).
Name of Azure Machine Learning Environment for deployment.
Version of an existing Azure Machine Learning Environment for deployment.
When a Pod starts and the liveness probe fails, Kubernetes will try --failure-threshold times before giving up. Defaults to 3. Minimum value is 1.
The amount of memory (in GB) to allocate for this Webservice. Can be a decimal.
The max amount of memory (in GB) this Webservice is allowed to use. Can be a decimal.
The number of gpu cores to allocate for this Webservice. Default is 1.
Path to a JSON or YAML file containing inference configuration.
Number of seconds after the container has started before liveness probes are initiated. Defaults to 310.
Key name for for encryption properties in customer-managed keys (CMK) for ACI.
Key version for for encryption properties in customer-managed keys (CMK) for ACI.
A primary auth key to use for this Webservice.
A secondary auth key to use for this Webservice.
The Azure region to deploy this Webservice to. If not specified the Workspace location will be used. More details on available regions can be found here: https://azure.microsoft.com/en-us/global-infrastructure/services/?regions=all&products=container-instances.
The maximum amount of time a request will stay in the queue (in milliseconds) before returning a 503 error. Defaults to 500.
The ID of the model to be deployed. Multiple models can be specified with additional -m arguments. Models need to be registered first.
Path to a JSON file containing model registration metadata. Multiple models can be provided using multiple -f parameters.
Kubernetes namespace in which to deploy the service: up to 63 lowercase alphanumeric ('a'-'z', '0'-'9') and hyphen ('-') characters. The first and last characters cannot be hyphens. Only applicable when deploying to AKS.
Flag to not wait for asynchronous calls.
The number of containers to allocate for this Webservice. No default, if this parameter is not set then the autoscaler is enabled by default.
Overwrite the existing service if name conflicts.
Path to a project folder. Default: current directory.
How often (in seconds) to perform the liveness probe. Default to 10 seconds. Minimum value is 1.
Path to a JSON file containing profiling results.
The local port on which to expose the service's HTTP endpoint.
Key/value property to add (e.g. key=value ). Multiple properties can be specified with multiple --property options.
The number of maximum concurrent requests per node to allow for this Webservice. Defaults to 1.
Resource group corresponding to the provided workspace.
Which runtime to use for image. Current supported runtimes are 'spark-py' and 'python'spark-py|python|python-slim.
The cname for if SSL is enabled.
A timeout to enforce for scoring calls to this Webservice. Defaults to 60000.
Path to folders that contain all files to create the image.
Whether or not to enable SSL for this Webservice. Defaults to False.
The key file needed if SSL is enabled.
The cert file needed if SSL is enabled.
Minimum consecutive successes for the liveness probe to be considered successful after having failed. Defaults to 1. Minimum value is 1.
Name of the subnet inside the vnet.
Specifies the subscription Id.
Key/value tag to add (e.g. key=value ). Multiple tags can be specified with multiple --tag options.
Number of seconds after which the liveness probe times out. Defaults to 2 second. Minimum value is 1.
Whether or not to enable token auth for this Webservice. Ignored if not deploying to AKS. Defaults to False.
The amount of traffic the version takes in an endpoint. Can be a decimal. Defaults to 0.
Vault base url for encryption properties in customer-managed keys (CMK) for ACI.
The version name in an endpoint. Defaults to endpoint name for the first version.
Name of the virtual network.
Name of the workspace.
Verbosity flag.
Global Parameters
Increase logging verbosity to show all debug logs.
Show this help message and exit.
Only show errors, suppressing warnings.
Output format.
JMESPath query string. See http://jmespath.org/ for more information and examples.
Name or ID of subscription. You can configure the default subscription using az account set -s NAME_OR_ID
.
Increase logging verbosity. Use --debug for full debug logs.
az ml model download
Download a model from the workspace.
az ml model download --model-id
--target-dir
[--overwrite]
[--path]
[--resource-group]
[--subscription-id]
[--workspace-name]
[-v]
Required Parameters
ID of model.
Target directory to download the model file to.
Optional Parameters
Overwrite if the same name file exists in target directory.
Path to a project folder. Default: current directory.
Resource group corresponding to the provided workspace.
Specifies the subscription Id.
Name of the workspace containing model to show.
Verbosity flag.
Global Parameters
Increase logging verbosity to show all debug logs.
Show this help message and exit.
Only show errors, suppressing warnings.
Output format.
JMESPath query string. See http://jmespath.org/ for more information and examples.
Name or ID of subscription. You can configure the default subscription using az account set -s NAME_OR_ID
.
Increase logging verbosity. Use --debug for full debug logs.
az ml model list
List models in the workspace.
az ml model list [--dataset-id]
[--latest]
[--model-name]
[--path]
[--property]
[--resource-group]
[--run-id]
[--subscription-id]
[--tag]
[--workspace-name]
[-v]
Optional Parameters
If provided, will only show models with the specified dataset ID.
If provided, will only return models with the latest version.
An optional model name to filter the list by.
Path to a project folder. Default: current directory.
Key/value property to add (e.g. key=value ). Multiple properties can be specified with multiple --property options.
Resource group corresponding to the provided workspace.
If provided, will only show models with the specified Run ID.
Specifies the subscription Id.
Key/value tag to add (e.g. key=value ). Multiple tags can be specified with multiple --tag options.
Name of the workspace containing models to list.
Verbosity flag.
Global Parameters
Increase logging verbosity to show all debug logs.
Show this help message and exit.
Only show errors, suppressing warnings.
Output format.
JMESPath query string. See http://jmespath.org/ for more information and examples.
Name or ID of subscription. You can configure the default subscription using az account set -s NAME_OR_ID
.
Increase logging verbosity. Use --debug for full debug logs.
az ml model package
Package a model in the workspace.
az ml model package [--cf]
[--ed]
[--entry-script]
[--environment-name]
[--environment-version]
[--ic]
[--il]
[--image-name]
[--model]
[--model-metadata-file]
[--no-wait]
[--output-path]
[--path]
[--resource-group]
[--rt]
[--sd]
[--subscription-id]
[--workspace-name]
[-v]
Optional Parameters
Path to local file containing a conda environment definition to use for the package.
Directory for Azure Machine Learning Environment for packaging. It is the same directory path as provided in 'az ml environment scaffold' command.
Path to local file that contains the code to run for service (relative path from source_directory if one is provided).
Name of Azure Machine Learning Environment for packaging.
Version of an existing Azure Machine Learning Environment for packaging.
Path to a JSON or YAML file containing inference configuration.
Label to give the built package image.
Name to give the built package image.
The ID of the model to be packaged. Multiple models can be specified with additional -m arguments. Models need to be registered first.
Path to a JSON file containing model registration metadata. Multiple models can be provided using multiple -f parameters.
Flag to not wait for asynchronous calls.
Output path for docker context. If an output path is passed, instead of building an image in the workspace ACR, a dockerfile and the necessary build context will be writen to that path.
Path to a project folder. Default: current directory.
Resource group corresponding to the provided workspace.
Which runtime to use for package. Current supported runtimes are 'spark-py' and 'python'spark-py|python|python-slim.
Path to folders that contain all files to create the image.
Specifies the subscription Id.
Name of the workspace.
Verbosity flag.
Global Parameters
Increase logging verbosity to show all debug logs.
Show this help message and exit.
Only show errors, suppressing warnings.
Output format.
JMESPath query string. See http://jmespath.org/ for more information and examples.
Name or ID of subscription. You can configure the default subscription using az account set -s NAME_OR_ID
.
Increase logging verbosity. Use --debug for full debug logs.
az ml model profile
Profile model(s) in the workspace.
az ml model profile --name
[--base-image]
[--base-image-registry]
[--cc]
[--cf]
[--description]
[--ed]
[--entry-script]
[--environment-name]
[--environment-version]
[--gb]
[--ic]
[--idi]
[--model]
[--model-metadata-file]
[--output-metadata-file]
[--resource-group]
[--sd]
[--subscription-id]
[--workspace-name]
[-v]
Required Parameters
The name of the model profile.
Optional Parameters
A custom image to be used as base image. If no base image is given then the base image will be used based off of given runtime parameter.
Image registry that contains the base image.
Double value for maximum CPU to use when profiling.
Path to local file containing a conda environment definition to use for the image.
Description of the model profile.
Directory for Azure Machine Learning Environment for deployment. It is the same directory path as provided in 'az ml environment scaffold' command.
Path to local file that contains the code to run for service (relative path from source_directory if one is provided).
Name of Azure Machine Learning Environment for deployment.
Version of an existing Azure Machine Learning Environment for deployment.
Double value for maximum Memory to use when profiling.
Path to a JSON or YAML file containing inference configuration.
ID of the Tabular Dataset to be used as input for the profile.
The ID of the model to be deployed. Multiple models can be specified with additional -m arguments. Models need to be registered first.
Path to a JSON file containing model registration metadata. Multiple models can be provided using multiple -f parameters.
Path to a JSON file where profile results metadata will be written. Used as input for model deployment.
Resource group corresponding to the provided workspace.
Path to folders that contain all files to create the image.
Specifies the subscription Id.
Name of the workspace.
Verbosity flag.
Global Parameters
Increase logging verbosity to show all debug logs.
Show this help message and exit.
Only show errors, suppressing warnings.
Output format.
JMESPath query string. See http://jmespath.org/ for more information and examples.
Name or ID of subscription. You can configure the default subscription using az account set -s NAME_OR_ID
.
Increase logging verbosity. Use --debug for full debug logs.
az ml model register
Register a model to the workspace.
az ml model register --name
[--asset-path]
[--cc]
[--description]
[--experiment-name]
[--gb]
[--gc]
[--model-framework]
[--model-framework-version]
[--model-path]
[--output-metadata-file]
[--path]
[--property]
[--resource-group]
[--run-id]
[--run-metadata-file]
[--sample-input-dataset-id]
[--sample-output-dataset-id]
[--subscription-id]
[--tag]
[--workspace-name]
[-v]
Required Parameters
Name of model to register.
Optional Parameters
The cloud path where the experiement run stores the model file.
The default number of CPU cores to allocate for this model. Can be a decimal.
Description of the model.
The name of the experiment.
The default amount of memory (in GB) to allocate for this model. Can be a decimal.
The default number of GPUs to allocate for this model.
Framework of the model to register. Currently supported frameworks: TensorFlow, ScikitLearn, Onnx, Custom, Multi.
Framework version of the model to register (e.g. 1.0.0, 2.4.1).
Full path of the model file to register.
Path to a JSON file where model registration metadata will be written. Used as input for model deployment.
Path to a project folder. Default: current directory.
Key/value property to add (e.g. key=value ). Multiple properties can be specified with multiple --property options.
Resource group corresponding to the provided workspace.
The ID for the experiment run where model is registered from.
Path to a JSON file containing experiement run metadata.
The ID for the sample input dataset.
The ID for the sample output dataset.
Specifies the subscription Id.
Key/value tag to add (e.g. key=value ). Multiple tags can be specified with multiple --tag options.
Name of the workspace to register this model with.
Verbosity flag.
Global Parameters
Increase logging verbosity to show all debug logs.
Show this help message and exit.
Only show errors, suppressing warnings.
Output format.
JMESPath query string. See http://jmespath.org/ for more information and examples.
Name or ID of subscription. You can configure the default subscription using az account set -s NAME_OR_ID
.
Increase logging verbosity. Use --debug for full debug logs.
az ml model show
Show a model in the workspace.
az ml model show [--model-id]
[--model-name]
[--path]
[--resource-group]
[--run-id]
[--subscription-id]
[--version]
[--workspace-name]
[-v]
Optional Parameters
ID of model to show.
Name of model to show.
Path to a project folder. Default: current directory.
Resource group corresponding to the provided workspace.
If provided, will only show models with the specified Run ID.
Specifies the subscription Id.
If provided, will only show models with the specified name and version.
Name of the workspace containing model to show.
Verbosity flag.
Global Parameters
Increase logging verbosity to show all debug logs.
Show this help message and exit.
Only show errors, suppressing warnings.
Output format.
JMESPath query string. See http://jmespath.org/ for more information and examples.
Name or ID of subscription. You can configure the default subscription using az account set -s NAME_OR_ID
.
Increase logging verbosity. Use --debug for full debug logs.
az ml model update
Update a model in the workspace.
az ml model update --model-id
[--add-property]
[--add-tag]
[--cc]
[--description]
[--gb]
[--gc]
[--path]
[--remove-tag]
[--resource-group]
[--sample-input-dataset-id]
[--sample-output-dataset-id]
[--subscription-id]
[--workspace-name]
[-v]
Required Parameters
ID of model.
Optional Parameters
Key/value property to add (e.g. key=value ). Multiple properties can be specified with multiple --add-property options.
Key/value tag to add (e.g. key=value ). Multiple tags can be specified with multiple --add-tag options.
The default number of CPU cores to allocate for this model. Can be a decimal.
Description to update the model with. Will replace the current description.
The default amount of memory (in GB) to allocate for this model. Can be a decimal.
The default number of GPUs to allocate for this model.
Path to a project folder. Default: current directory.
Key of tag to remove. Multiple tags can be specified with multiple --remove-tag options.
Resource group corresponding to the provided workspace.
The ID for the sample input dataset.
The ID for the sample output dataset.
Specifies the subscription Id.
Name of the workspace.
Verbosity flag.
Global Parameters
Increase logging verbosity to show all debug logs.
Show this help message and exit.
Only show errors, suppressing warnings.
Output format.
JMESPath query string. See http://jmespath.org/ for more information and examples.
Name or ID of subscription. You can configure the default subscription using az account set -s NAME_OR_ID
.
Increase logging verbosity. Use --debug for full debug logs.