Azure Machine Learning best practices for enterprise security

This article explains security best practices for planning or managing a secure Azure Machine Learning deployment. Best practices come from Microsoft and customer experience with Azure Machine Learning. Each guideline explains the practice and its rationale. The article also provides links to how-to and reference documentation.

The recommended machine learning network security architecture is a managed virtual network. An Azure Machine Learning managed virtual network secures the workspace, associated Azure resources, and all managed compute resources. It simplifies the configuration and management of network security by preconfiguring required outputs and automatically creating managed resources within the network. You can use private endpoints to allow Azure services to access the network and can optionally define outbound rules to allow the network to access the internet.

The managed virtual network has two modes that it can be configured for:

  • Allow internet outbound - This mode allows outbound communication with resources located on the internet, such as the public PyPi or Anaconda package repositories.

    A diagram of the recommended architecture with the internet outbound mode.

  • Allow only approved outbound - This mode allows only the minimum outbound communication required for the workspace to function. This mode is recommended for workspaces that must be isolated from the internet. Or where outbound access is only allowed to specific resources via service endpoints, service tags, or fully qualified domain names.

    A diagram of the recommended architecture with the only allowed outbound mode.

For more information, see Managed virtual network isolation.

If you can't use a managed virtual network due to your business requirements, you can use an Azure virtual network with the following subnets:

  • Training contains compute resources used for training, such as machine learning compute instances or compute clusters.
  • Scoring contains compute resources used for scoring, such as Azure Kubernetes Service (AKS).
  • Firewall contains the firewall that allows traffic to and from the public internet, such as Azure Firewall.

A diagram of the recommended architecture when using Azure virtual network.

The virtual network also contains a private endpoint for your machine learning workspace and the following dependent services:

  • Azure Storage account
  • Azure Key Vault
  • Azure Container Registry

Outbound communication from the virtual network must be able to reach the following Microsoft services:

  • Machine learning
  • Microsoft Entra ID
  • Azure Container Registry, and specific registries that Microsoft maintains
  • Azure Front Door
  • Azure Resource Manager
  • Azure Storage

Remote clients connect to the virtual network using Azure ExpressRoute or a virtual private network (VPN) connection.

Virtual network and private endpoint design

When designing an Azure Virtual Network, subnets, and private endpoints, consider the following requirements:

  • In general, create separate subnets for training and scoring and use the training subnet for all private endpoints.

  • For IP addressing, compute instances need one private IP each. Compute clusters need one private IP per node. AKS clusters need many private IP addresses, as described in Plan IP addressing for your AKS cluster. A separate subnet for at least AKS helps prevent IP address exhaustion.

  • The compute resources in the training and scoring subnets must access the storage account, the key vault, and the container registry. Create private endpoints for the storage account, the key vault, and the container registry.

  • Machine learning workspace default storage needs two private endpoints, one for Azure Blob Storage and another for Azure File Storage.

  • If you use Azure Machine Learning studio, the workspace and storage private endpoints should be in the same virtual network.

  • If you have multiple workspaces, use a virtual network for each workspace to create an explicit network boundary between workspaces.

Use private IP addresses

Private IP addresses minimize your Azure resources' exposure to the internet. Machine learning uses many Azure resources, and the machine learning workspace private endpoint isn't enough for end-to-end private IP. The following table shows the major resources machine learning uses and how to enable private IP for the resources. Compute instances and compute clusters are the only resources that don't have the private IP feature.

Resources Private IP solution Documentation
Workspace Private endpoint Configure a private endpoint for an Azure Machine Learning workspace
Registry Private endpoint Network isolation with Azure Machine Learning registries
Associated resources
Storage Private endpoint Secure Azure Storage accounts with service endpoints
Key Vault Private endpoint Secure Azure Key Vault
Container Registry Private endpoint Enable Azure Container Registry
Training resources
Compute instance Private IP (no public IP) Secure training environments
Compute cluster Private IP (no public IP) Secure training environments
Hosting resources
Managed online endpoint Private endpoint Network isolation with managed online endpoints
Online endpoint (Kubernetes) Private endpoint Secure Azure Kubernetes Service online endpoints
Batch endpoints Private IP (inherited from compute cluster) Network isolation in batch endpoints

Control virtual network inbound and outbound traffic

Use a firewall or Azure network security group (NSG) to control virtual network inbound and outbound traffic. For more information on inbound and outbound requirements, see Configure inbound and outbound network traffic. For more information on traffic flows between components, see Network traffic flow in a secured workspace.

Ensure access to your workspace

To ensure that your private endpoint can access your machine learning workspace, take the following steps:

  1. Make sure you have access to your virtual network using a VPN connection, ExpressRoute, or jump box virtual machine (VM) with Azure Bastion access. The public user can't access the machine learning workspace with the private endpoint, because it can be accessed only from your virtual network. For more information, see Secure your workspace with virtual networks.

  2. Make sure you can resolve the workspace fully qualified domain names (FQDNs) with your private IP address. If you use your own Domain Name System (DNS) server or a centralized DNS infrastructure, you need to configure a DNS forwarder. For more information, see How to use your workspace with a custom DNS server.

Workspace access management

When defining machine learning identity and access management controls, you can separate controls that define access to Azure resources from controls that manage access to data assets. Depending on your use case, consider whether to use self-service, data-centric, or project-centric identity and access management.

Self-service pattern

In a self-service pattern, data scientists can create and manage workspaces. This pattern is best suited for proof-of-concept situations requiring flexibility to try different configurations. The disadvantage is that data scientists need the expertise to provision Azure resources. This approach is less suitable when strict control, resource use, audit traces, and data access are required.

  1. Define Azure policies to set safeguards for resource provisioning and usage, such as allowed cluster sizes and VM types.

  2. Create a resource group for holding the workspaces and grant data scientists a Contributor role in the resource group.

  3. Data scientists can now create workspaces and associate resources in the resource group in a self-service manner.

  4. To access data storage, create user-assigned managed identities and grant the identities read-access roles on the storage.

  5. When data scientists create compute resources, they can assign the managed identities to the compute instances to gain data access.

For best practices, see Authentication for cloud-scale analytics.

Data-centric pattern

In a data-centric pattern, the workspace belongs to a single data scientist who might be working on multiple projects. The advantage of this approach is that the data scientist can reuse code or training pipelines across projects. As long as the workspace is limited to a single user, data access can be traced back to that user when auditing storage logs.

The disadvantage is that data access isn't compartmentalized or restricted on a per-project basis, and any user added to the workspace can access the same assets.

  1. Create the workspace.

  2. Create compute resources with system-assigned managed identities enabled.

  3. When a data scientist needs access to the data for a given project, grant the compute managed identity read access to the data.

  4. Grant the compute managed identity access to other required resources, such as a container registry with custom Docker images for training.

  5. Also grant the workspace's managed identity read-access role on the data to enable data preview.

  6. Grant the data scientist access to the workspace.

  7. The data scientist can now create data stores to access data required for projects and submit training runs that use the data.

Optionally, create a Microsoft Entra security group and grant it read access to data, then add managed identities to the security group. This approach reduces the number of direct role assignments on resources, to avoid reaching the subscription limit on role assignments.

Project-centric pattern

A project-centric pattern creates a machine learning workspace for a specific project, and many data scientists collaborate within the same workspace. Data access is restricted to the specific project, making the approach well suited for working with sensitive data. Also, it's straightforward to add or remove data scientists from the project.

The disadvantage of this approach is that sharing assets across projects can be difficult. It's also hard to trace data access to specific users during audits.

  1. Create the workspace

  2. Identify data storage instances required for the project, create a user-assigned managed identity, and grant the identity read access to the storage.

    Optionally, grant the workspace's managed identity access to data storage to allow data preview. You can omit this access for sensitive data not suitable for preview.

  3. Create credentialless data stores for the storage resources.

  4. Create compute resources within the workspace, and assign the managed identity to the compute resources.

  5. Grant the compute managed identity access to other required resources, such as a container registry with custom Docker images for training.

  6. Grant data scientists working on the project a role on the workspace.

    By using Azure role-based access control (RBAC), you can restrict data scientists from creating new datastores or new compute resources with different managed identities. This practice prevents access to data not specific to the project.

    Optionally, to simplify project membership management, you can create a Microsoft Entra security group for project members and grant the group access to the workspace.

Azure Data Lake Storage with credential passthrough

You can use Microsoft Entra user identity for interactive storage access from machine learning studio. Data Lake Storage with hierarchical namespace enabled allows for enhanced organization of data assets for storage and collaboration. With Data Lake Storage hierarchical namespace, you can compartmentalize data access by giving different users access control list (ACL)-based access to different folders and files. For example, you can grant only a subset of users access to confidential data.

RBAC and custom roles

Azure RBAC helps you manage who has access to machine learning resources and configure who can perform operations. For example, you might want to grant only specific users the workspace administrator role to manage compute resources.

Access scope can differ between environments. In a production environment, you might want to limit the ability of users to update inference endpoints. Instead, you might grant that permission to an authorized service principal.

Machine learning has several default roles: owner, contributor, reader, and data scientist. You can also create your own custom roles, for example to create permissions that reflect your organizational structure. For more information, see Manage access to Azure Machine Learning workspace.

Over time, the composition of your team might change. If you create a Microsoft Entra group for each team role and workspace, you can assign an Azure RBAC role to the Microsoft Entra group, and manage resource access and user groups separately.

User principals and service principals can be part of the same Microsoft Entra group. For example, when you create a user-assigned managed identity that Azure Data Factory uses to trigger a machine learning pipeline, you might include the managed identity in a ML pipelines executor Microsoft Entra group.

Central Docker image management

Azure Machine Learning provides curated Docker images that you can use for training and deployment. However, your enterprise compliance requirements might mandate using images from a private repository your company manages. Machine learning has two ways to use a central repository:

  • Use the images from a central repository as base images. The machine learning environment management installs packages and creates a Python environment where the training or inferencing code runs. With this approach, you can update package dependencies easily without modifying the base image.

  • Use the images as-is, without using machine learning environment management. This approach gives you a higher degree of control but also requires you to carefully construct the Python environment as part of the image. You need to meet all the necessary dependencies to run the code, and any new dependencies require rebuilding the image.

For more information, see Manage environments.

Data encryption

Machine learning data at rest has two data sources:

  • Your storage has all your data, including training and trained model data, except for the metadata. You're responsible for your storage encryption.

  • Azure Cosmos DB contains your metadata, including run history information like experiment name and experiment submission date and time. In most workspaces, Azure Cosmos DB is in the Microsoft subscription and encrypted by a Microsoft-managed key.

    If you want to encrypt your metadata using your own key, you can use a customer-managed key workspace. The downside is that you need to have Azure Cosmos DB in your subscription and pay its cost. For more information, see Data encryption with Azure Machine Learning.

For information on how Azure Machine Learning encrypts data in transit, see Encryption in transit.


When you deploy machine learning resources, set up logging and auditing controls for observability. Motivations for observing data might vary based on who looks at the data. Scenarios include:

  • Machine learning practitioners or operations teams want to monitor machine learning pipeline health. These observers need to understand issues in scheduled execution or problems with data quality or expected training performance. You can build Azure dashboards that monitor Azure Machine Learning data or create event-driven workflows.

  • Capacity managers, machine learning practitioners, or operations teams might want to create a dashboard to observe compute and quota utilization. To manage a deployment with multiple Azure Machine Learning workspaces, consider creating a central dashboard to understand quota utilization. Quotas are managed on a subscription level, so the environment-wide view is important to drive optimization.

  • IT and operations teams can set up diagnostic logging to audit resource access and altering events in the workspace.

  • Consider creating dashboards that monitor overall infrastructure health for machine learning and dependent resources such as storage. For example, combining Azure Storage metrics with pipeline execution data can help you optimize infrastructure for better performance or discover problem root causes.

Azure collects and stores platform metrics and activity logs automatically. You can route the data to other locations by using a diagnostic setting. Set up diagnostic logging to a centralized Log Analytics workspace for observability across several workspace instances. Use Azure Policy to automatically set up logging for new machine learning workspaces into this central Log Analytics workspace.

Azure Policy

You can enforce and audit the usage of security features on workspaces through Azure Policy. Recommendations include:

  • Enforce custom-managed key encryption.
  • Enforce Azure Private Link and private endpoints.
  • Enforce private DNS zones.
  • Disable non-Azure AD authentication, such as Secure Shell (SSH).

For more information, see Built-in policy definitions for Azure Machine Learning.

You can also use custom policy definitions to govern workspace security in a flexible manner.

Compute clusters and instances

The following considerations and recommendations apply to machine learning compute clusters and instances.

Disk encryption

The operating system (OS) disk for a compute instance or compute cluster node is stored in Azure Storage and encrypted with Microsoft-managed keys. Each node also has a local temporary disk. The temporary disk is also encrypted with Microsoft-managed keys if the workspace was created with the hbi_workspace = True parameter. For more information, see Data encryption with Azure Machine Learning.

Managed identity

Compute clusters support using managed identities to authenticate to Azure resources. Using a managed identity for the cluster allows authentication to resources without exposing credentials in your code. For more information, see Create an Azure Machine Learning compute cluster.

Setup script

You can use a setup script to automate the customization and configuration of compute instances at creation. As an administrator, you can write a customization script to use when creating all compute instances in a workspace. You can use Azure Policy to enforce the use of the setup script to create every compute instance. For more information, see Create and manage an Azure Machine Learning compute instance.

Create on behalf of

If you don't want data scientists to provision compute resources, you can create compute instances on their behalf and assign them to the data scientists. For more information, see Create and manage an Azure Machine Learning compute instance.

Private endpoint-enabled workspace

Use compute instances with a private endpoint-enabled workspace. The compute instance rejects all public access from outside the virtual network. This configuration also prevents packet filtering.

Azure Policy support

When using an Azure virtual network, you can use Azure Policy to ensure that every compute cluster or instance is created in a virtual network and specify the default virtual network and subnet. The policy isn't needed when using a managed virtual network, as the compute resources are automatically created in the managed virtual network.

You can also use a policy to disable non-Azure AD authentication, such as SSH.

Next steps

Learn more about machine learning security configurations:

Get started with a machine learning template-based deployment:

Read more articles about architectural considerations for deploying machine learning: