Offline seeding for MARS using customer-owned disks with Azure Import/Export
This article describes how to send the initial full backup data from MARS to Azure using customer-owned disks instead of sending it via the network. Learn about sending the initial full backup data from DPM/MABS to Azure using customer-owned disks.
Azure Backup has several built-in efficiencies that save network and storage costs during the initial full backups of data to Azure. Initial full backups typically transfer large amounts of data and require more network bandwidth when compared to subsequent backups that transfer only the deltas/incrementals. Through the process of offline seeding, Azure Backup can use disks to upload the offline backup data to Azure.
The Azure Backup offline-seeding process is tightly integrated with the Azure Import/Export service. You can use this service to transfer initial backup data to Azure by using disks. If you have terabytes (TBs) of initial backup data that need to be transferred over a high-latency and low-bandwidth network, you can use the offline-seeding workflow to ship the initial backup copy, on one or more hard drives to an Azure datacenter.
To perform the offline backup:
- Instead of sending the backup data over the network, write the backup data to a staging location.
- Use the AzureOfflineBackupDiskPrep utility to write the data in the staging location to one or more SATA disks.
- As part of the preparatory work, the AzureOfflineBackupDiskPrep utility creates an Azure import job. Send the SATA drives to the nearest Azure datacenter, and reference the import job to connect the activities.
- At the Azure datacenter, the data on the disks is copied to an Azure storage account.
- Azure Backup copies the backup data from the storage account to the Recovery Services vault, and incremental backups are scheduled.
The following diagram provides an overview of the offline-seeding flow:
Ensure that you use the latest MARS agent (version 2.0.9250.0 or higher) before following the below sections. Learn more.
The following Azure Backup features or workloads support the use of offline backup for:
- Backup of files and folders with the Microsoft Azure Recovery Services (MARS) Agent, also referred to as the Azure Backup Agent.
- Backup of all workloads and files with System Center Data Protection Manager (DPM).
- Backup of all workloads and files with Microsoft Azure Backup Server.
Offline backup isn't supported for system state backups done by using the Azure Backup agent.
The following prerequisites and workflow apply only to offline backup of files and folders using the latest Azure Recovery Services Agent. To perform offline backups for workloads using System Center DPM or Azure Backup Server, see Offline backup workflow for DPM and Azure Backup Server.
Before you start the offline backup workflow, complete the following prerequisites:
Make sure that only the latest version of the Azure Backup Agent is installed on the Windows Server or Windows client, as applicable, and the computer is registered with the Recovery Services vault.
Azure PowerShell 3.7.0 is required on the computer running the Azure Backup Agent. Download and install the 3.7.0 version of Azure PowerShell.
Create an Azure storage account in the same subscription as the Recovery Services vault.
Make sure you have the necessary permissions to create the Microsoft Entra application. The offline backup workflow creates a Microsoft Entra application in the subscription associated with the Azure storage account. The goal of the application is to provide Azure Backup with secure and scoped access to the Azure Import/Export service, which is required for the offline backup workflow.
Register the Microsoft.DataBox resource provider with the subscription that contains the Azure storage account. To register the resource provider:
On the main menu, select Subscriptions.
If you're subscribed to multiple subscriptions, select the subscription you plan to use for the offline backup. If you use only one subscription, then your subscription appears.
On the subscription menu, select Resource providers to view the list of providers.
In the list of providers, scroll down to Microsoft.DataBox. If the Status is NotRegistered, select Register.
A staging location, which might be a network share or any additional drive on the computer, internal or external, with enough disk space to hold your initial copy, is created. For example, if you want to back up a 500-GB file server, ensure that the staging area is at least 500 GB. (A smaller amount is used due to compression.)
When you send disks to Azure, use only 2.5-inch SSD or 2.5-inch or 3.5-inch SATA II/III internal hard drives. You can use hard drives up to 10 TB. Check the Azure Import/Export service documentation for the latest set of drives that the service supports.
The SATA drives must be connected to a computer (referred to as a copy computer) from where the copy of backup data from the staging location to the SATA drives is done. Ensure that BitLocker is enabled on the copy computer.
This section describes the offline backup workflow so that your data can be delivered to an Azure datacenter and uploaded to Azure Storage. If you have questions about the import service or any aspect of the process, see the Azure Import/Export service overview documentation.
Initiate offline backup
When you schedule a backup on the Recovery Services Agent, you see this page.
Select the option Transfer using my own disks.
Use the Azure Data Box option to transfer initial backup data offline. This option saves the effort required to procure your own Azure-compatible disks. It delivers Microsoft-proprietary, secure, and tamperproof Azure Data Box devices to which backup data can be directly written to by the Recovery Services Agent.
Select Next, and fill in the boxes carefully.
The boxes that you fill in are:
- Staging Location: The temporary storage location to which the initial backup copy is written. The staging location might be on a network share or a local computer. If the copy computer and source computer are different, specify the full network path of the staging location.
- Azure Resource Manager Storage Account: The name of the Resource Manager type storage account (general purpose v1 or general purpose v2) in any Azure subscription.
- Azure Storage Container: The name of the destination blob storage container in the Azure storage account where the backup data is imported before being copied to the Recovery Services vault.
- Azure Subscription ID: The ID for the Azure subscription where the Azure storage account is created.
- Azure Import Job Name: The unique name by which the Azure Import/Export service and Azure Backup track the transfer of data sent on disks to Azure.
After you fill in the boxes, select Next. Save the Staging Location and the Azure Import Job Name information. It's required to prepare the disks.
When prompted, sign in to your Azure subscription. You must sign in so that Azure Backup can create the Microsoft Entra application. Enter the required permissions to access the Azure Import/Export service.
Finish the workflow. On the Azure Backup Agent console, select Back Up Now.
On the Confirmation page of the wizard, select Back Up. The initial backup is written to the staging area as part of the setup.
After the operation finishes, the staging location is ready to be used for disk preparation.
Prepare SATA drives and ship to Azure
The AzureOfflineBackupDiskPrep utility prepares the SATA drives that are sent to the nearest Azure datacenter. This utility is available in the Azure Backup Agent installation directory in the following path:
*\Microsoft Azure Recovery Services Agent\Utils\\*
Go to the directory, and copy the AzureOfflineBackupDiskPrep directory to another computer where the SATA drives are connected. On the computer with the connected SATA drives, ensure that:
The copy computer can access the staging location for the offline-seeding workflow by using the same network path that was provided in the workflow in the "Initiate offline backup" section.
BitLocker is enabled on the copy computer.
Azure PowerShell 3.7.0 is installed.
The copy computer can access the Azure portal. If necessary, the copy computer can be the same as the source computer.
If the source computer is a virtual machine, then the copy computer must be a different physical server or client machine from the source computer.
Open an elevated command prompt on the copy computer with the AzureOfflineBackupDiskPrep utility directory as the current directory. Run the following command:
.\AzureOfflineBackupDiskPrep.exe s:<Staging Location Path>
Parameter Description s:<Staging Location Path> This mandatory input is used to provide the path to the staging location that you entered in the workflow in the "Initiate offline backup" section.
When you run the command, the utility requests the selection of the Azure import job that corresponds to the drives that need to be prepared. If only a single import job is associated with the provided staging location, you see a page like this one.
Enter the drive letter without the trailing colon for the mounted disk that you want to prepare for transfer to Azure.
Provide confirmation for the formatting of the drive when prompted.
You're prompted to sign in to your Azure subscription. Enter your credentials.
The tool then begins to prepare the disk and copy the backup data. You might need to attach additional disks when prompted by the tool if the provided disk doesn't have sufficient space for the backup data.
After successful copy of the data from the staging location to the disks, the tool shows the following details:
- The list of disks prepared for seeding.
- The name of the storage account, resource group, and country/region of the Import/Export Job.
The tool lists the fields required to create the Import/Export Job.* Enter the following details:
Required Parameter Detail Contact Name Name of the contact for the Import/Export Job Contact Number Phone number of the contact for the Import/Export Job Valid Email Id Email ID to notify for the Import/Export Job Shipping Address The return shipping address Country Return shipping country/region Postal Code Return shipping postal code
*All fields are required.
You can edit these parameters in future in the Azure portal for the Import/Export Job.
After you enter these parameters and run the tool successfully, you receive a confirmation of the successful creation of the import job.
The tool also displays the Azure data centre address to which the disks need to be shipped along with a list of supported carriers.
Ship the disks to the address that the tool provided. Keep the tracking number for future reference and update it in the Azure portal as soon as possible.
No two Azure import jobs can have the same tracking number. Ensure that drives created by the utility under a single Azure import job are shipped together in a single package and that there is a single unique tracking number for the package. Don't combine drives prepared as part of separate Azure import jobs in a single package.
Update the tracking and shipping details on the Azure import job
This section helps you update the Azure import job shipping details, which include details about:
- The name of the carrier that delivers the disks to Azure.
- Return-shipping details for your disks.
- Modify the notification email for the import job.
Update the tracking details
To update the tracking details, follow these steps:
Sign in to the Azure subscription.
On the main menu, select All services.
On the All services pane, enter Azure Data Box in the search box, and then select it from the search result.
On the Azure Data Box menu, the list of all Azure Data Box jobs under the selected subscription appears (including Import/Export).
Enter Import/Export on the search box to filter the Import/Export jobs, or enter the job name directly, and then select the newly created import job to view its details.
If you've multiple subscriptions, select the subscription used to import the backup data.
Select the job, and then on the Overview pane, add the Carrier and Tracking Number to update the Tracking information.
Add return-shipping details
To add the return-shipping details, follow these steps:
- Select Job Details under General, and then Edit Address.
- Update the carrier, carrier account number, contact details, and the return shipping address details
- Select Save.
Edit notification email
To update the email addresses that are notified on the Import job progress, select Edit notification details.
Ensure that the carrier information and tracking number are updated within two weeks of Azure import job creation. Failure to verify this information within two weeks can result in the job being deleted and drives not being processed.
Time to process the drives
The amount of time it takes to process an Azure import job varies. Process time is based on factors like shipping time, job type, type and size of the data being copied, and the size of the disks provided. The Azure Import/Export service doesn't have an SLA. After disks are received, the service strives to complete the backup data copy to your Azure storage account in 7 to 10 days.
Monitor Azure import job status
To monitor the status of your import job from the Azure portal, go to the Azure Data Box pane and select the job.
For more information on the status of the import jobs, see Monitor Azure Import/Export Jobs.
Finish the workflow
After the import job successfully completes, initial backup data is available in your storage account. At the time of the next scheduled backup, Azure Backup copies the contents of the data from the storage account to the Recovery Services vault.
At the time of the next scheduled backup, Azure Backup performs an incremental backup.
Clean up resources
After the initial backup is finished, you can safely delete the data imported to the Azure Storage container and the backup data in the staging location.
- For any questions about the Azure Import/Export service workflow, see Use the Microsoft Azure Import/Export service to transfer data to Blob storage.