Edit

Share via


Partition and assign GPUs to a virtual machine

This article describes how to configure graphics processing unit (GPU) partitions and assign a partition to a virtual machine (VM). It provides instructions on how to configure GPU partition count, assign GPU partitions, and unassign GPU partitions via Windows Admin Center and PowerShell.

To provision the GPU partitioning feature, you need to complete the following steps:

Prerequisites

There are several requirements and things to consider before you begin to use the GPU partitioning feature:

Prerequisites for the host server

  • You must have Windows Server 2025 installed on the host server. If clustering is required when live migrating a virtual machine host, Windows Server 2025 Datacenter must be installed.

  • The Hyper-V role installed and configured on your server. See Install the Hyper-V role on Windows Server to found out how to get stared.

  • Install the physical GPU device of the same make, model, and size on every server of the cluster. Refer to your OEM-provided documentation when installing the GPU device on your physical servers in the cluster.

  • Install the GPU drivers on every server of the cluster by following instructions from your GPU IHVs. For NVIDIA GPU drivers, see the NVIDIA vGPU documentation.

  • Ensure that the virtualization support and SR-IOV are enabled in the BIOS of each server in the cluster. Reach out to your system vendor if you're unable to identify the correct setting in your BIOS.

  • Cluster hosts need to have Input/Output Memory Management Unit (IOMMU) DMA bit tracking capable processors. For example, processors supporting Intel VT-D or AMD-Vi.

Note

When live migrating a virtual machine with a GPU partition assigned, Hyper-V live migration will automatically fall back to using TCP/IP with compression. This has the potential effect of increasing the CPU utilization of a host. In addition, live migrations could take longer than with virtual machines without GPU partitions attached.

Prerequisites for the VMs

Prerequisites for Windows Admin Center

If you're using Windows Admin Center to provision GPU partitioning, you must install the latest version of Windows Admin Center with the GPUs extension, version 2.8.0 or later. For instructions on how to install the GPUs extensions in Windows Admin Center, see Installing an extension.

After you install the extension, it appears under the Installed extensions tab as shown in the following screenshot. Make sure the version of the GPUs extension is 2.8.0 or later.

Screenshot of the Installed extensions tab in Windows Admin Center to verify the GPUs extension is installed.

Prerequisites for PowerShell

If you're using PowerShell to provision GPU partitioning, you must run all PowerShell commands as the Administrator user.

For detailed information on how to use PowerShell commands for GPU partitioning, see the Add-VMGpuPartitionAdapter, Get-VMGpuPartitionAdapter, and Remove-VMGpuPartitionAdapter reference documentation.

Verify GPU driver installation

After you complete all the prerequisites, you must verify if the GPU driver is installed and partitionable.

Follow these steps to verify if the GPU driver is installed and partitionable using PowerShell:

  1. Connect to a server in your cluster using Remote Desktop Protocol (RDP) and run the following PowerShell command to verify if the GPU device is installed:

    Get-PnpDevice -FriendlyName "<device-friendly-name>"
    

    For example, to list the NVIDIA A2 GPU devices, enter the following command:

    Get-PnpDevice -FriendlyName "NVIDIA A2"
    

    Here's a sample output:

    Status    Class    FriendlyName
    ------    -----    ------------
    OK        Display  NVIDIA A2
    

    For NVIDIA GPU devices, you can also run the NVIDIA System Management Interface, nvidia-smi to list the GPUs on the host server.

    If the driver is installed, you see an output similar to the following sample:

    nvidia-smi
    

    Here's a sample output:

    Wed Nov 30 15:22:36 2022
    +-----------------------------------------------------------------------------+
    | NVIDIA-SMI 527.27       Driver Version: 527.27       CUDA Version: N/A      |
    |-------------------------------+----------------------+----------------------+
    | GPU  Name            TCC/WDDM | Bus-Id        Disp.A | Volatile Uncorr. ECC |
    | Fan  Temp  Perf  Pwr:Usage/Cap|         Memory-Usage | GPU-Util  Compute M. |
    |                               |                      |               MIG M. |
    |===============================+======================+======================|
    |   0  NVIDIA A2          WDDM  | 00000000:65:00.0 Off |                    0 |
    |  0%   24C    P8     5W /  60W |  15192MiB / 15356MiB |      0%      Default |
    |                               |                      |                  N/A |
    +-------------------------------+----------------------+----------------------+
    |   1  NVIDIA A2          WDDM  | 00000000:66:00.0 Off |                    0 |
    |  0%   24C    P8     5W /  60W |  15192MiB / 15356MiB |      0%      Default |
    |                               |                      |                  N/A |
    +-------------------------------+----------------------+----------------------+
    
    +-----------------------------------------------------------------------------+
    | Processes:                                                                  |
    |  GPU   GI   CI        PID   Type   Process name                  GPU Memory |
    |        ID   ID                                                   Usage      |
    |=============================================================================|
    |  No running processes found                                                 |
    +-----------------------------------------------------------------------------+
    
  2. Run the following command to verify that the host server has the required GPU adapters installed by listing the GPUs that support GPU partitioning.

    Get-VMHostPartitionableGpu | FL Name,ValidPartitionCounts
    

    Note down the Name and ValidPartitionCounts values from the output of the Get-VMHostPartitionableGpu command. You use them later to configure partition count.

    Here's a sample output of the Get-VMHostPartitionableGpu command, which shows that two partitionable GPU drivers are installed on this server. The configured partition count for the first GPU is 16 and for the second GPU is 4.

    PS C:\Users> Get-VMHostPartitionableGPU
    
    
    Name                    : \\?\PCI#VEN_10DE&DEV_25B6&SUBSYS_157E10DE&REV_A1#4&18416dc3&0&0000#{064092b3-625e-43bf-9eb5-dc845897dd59}
    ValidPartitionCounts    : {16, 8, 4, 2...}
    
    Name                    : \\?\PCI#VEN_10DE&DEV_25B6&SUBSYS_157E10DE&REV_A1#4&5906f5e&0&0010#{064092b3-625e-43bf-9eb5-dc845897dd59}
    ValidPartitionCounts    : {16, 8, 4, 2...}
    

Configure GPU partition count

Each partitionable GPU comes with a set of valid partition counts that's predefined by its OEM. You can't define the number of partitions a GPU can have. However, you can configure the partition count to any of the valid count from within the supported set.

Follow these steps to configure GPU partition count in PowerShell:

Note

When using PowerShell, you must manually ensure the GPU configuration is homogenous across all the servers in your cluster.

  1. Refer to the Name and ValidPartitionCounts values you noted earlier when you ran the Get-VMHostPartitionableGpu command. See the PowerShell tab in the Verify GPU driver installation previous section.

  2. Connect to the server whose GPU partition count you want to configure.

  3. Run the following command to configure the partition count. Replace GPU-name with the Name value and partition-count with one of the supported counts from the ValidPartitionCounts value:

    Set-VMHostPartitionableGpu -Name "<GPU-name>" -PartitionCount <partition-count>
    

    For example, the following command configures the partition count to 4:

    PS C:\Users> Set-VMHostPartitionableGpu -Name "\\?\PCI#VEN_10DE&DEV_25B6&SUBSYS_157E10DE&REV_A1#4&18416dc3&0&0000#{064092b3-625e-43bf-9eb5-dc845897dd59}" -PartitionCount 4
    

    You can run the Get-VMHostPartitionableGpu | FL Name,ValidPartitionCounts,PartitionCount again to verify if the partition count is set to 4.

    Here's a sample output:

    PS C:\Users> Get-VMHostPartitionableGpu | FL Name,ValidPartitionCounts,PartitionCount
    
    Name                    : \\?\PCI#VEN_10DE&DEV_25B6&SUBSYS_157E10DE&REV_A1#4&18416dc3&0&0000#{064092b3-625e-43bf-9eb5-dc845897dd59}
    ValidPartitionCounts    : {16, 8, 4, 2...}
    PartitionCount          : 4
    
    Name                    : \\?\PCI#VEN_10DE&DEV_25B6&SUBSYS_157E10DE&REV_A1#4&5906f5e&0&0010#{064092b3-625e-43bf-9eb5-dc845897dd59}
    ValidPartitionCounts    : {16, 8, 4, 2...}
    PartitionCount          : 4
    
  4. To keep the configuration homogeneous, repeat the partition count configuration steps on each server in your cluster.

Assign GPU partition to a VM

Save your workloads before assigning partition to the VM.

Note

Currently, you can assign only a single GPU partition to a VM. Both the VM and the GPU partition need to be on the same host machine. We recommend that you plan ahead and determine the GPU partition size based on your workload performance requirements.

Follow these steps to assign GPU partition to a VM using PowerShell:

  1. Connect to the server hosting the VM to which you want to assign a GPU partition.

  2. Run the following command to configure the cluster VM resource’s default offline action as force-shutdown rather than save.

    Get-ClusterResource -name vmname | Set-ClusterParameter -Name "OfflineAction" -Value 3
    

    Note

    If a failure occurs, you need to shutdown the VM, drain the server, and manually fail over the VM to another server.

  3. To assign the partition, run the following command.

    Add-VMGpuPartitionAdapter -VMName $VMName
    
  4. To verify if the partition is assigned, run the following command.

    Get-VMGpuPartitionAdapter -VMName $VMName | FL InstancePath,PartitionId,PartitionVfLuid
    

    For example, run the following commands to first assign a partition to the mytestgpu-vm1 VM and then verify the assignment:

    $VMname = "mytestgpu-vm1"
    Add-VMGpuPartitionAdapter -VMName $VMName
    Get-VMGpuPartitionAdapter -VMName $VMName | FL InstancePath,PartitionId,PartitionVfLuid
    

    Here's a sample output:

    InstancePath            : \\?\PCI#VEN_10DE&DEV_25B6&SUBSYS_157E10DE&REV_A1#4&18416dc3&0&0000#{064092b3-625e-43bf-9eb5-dc845897dd59}
    PartitionId             : 0
    PartitionVfLuid         : 10-973909323
    
  5. Start the VM using PowerShell or Windows Admin Center to resolve the partitions.

    The next step is to start the VM using PowerShell or Windows Admin Center. Once the VM is up and running, it shows a GPU in Device Manager.

Unassign a partition from a VM

You can unassign a GPU partition from the VM if you no longer need it to run your workloads. Unassigning the partition frees up the GPU partition resource, which you can reassign to another VM later.

You must save your workloads before unassigning partitions.

Follow these steps to unassign a GPU partition from a VM using PowerShell:

  1. Connect to the server hosting the VM from which you want to unassign the GPU partition.

  2. Run the following command to configure the cluster VM resource’s default offline action as force-shutdown rather than save.

    Get-ClusterResource -name vmname | Set-ClusterParameter -Name "OfflineAction" -Value 3
    

    Note

    If a failure occurs, you need to shutdown the VM, drain the server, and manually fail over the VM to another server.

  3. Run the following command to remove the partition:

    Remove-VMGpuPartitionAdapter -VMName $VMname
    
  4. Run the following command to confirm the partition is removed. If the partition is removed, this command doesn't return anything.

    Get-VMGpuPartitionAdapter -VMName $VMName