แก้ไข

แชร์ผ่าน


Migrate Db2 Data into SQL Server (Db2ToSQL)

After you successfully synchronize the converted objects with SQL Server, you can migrate data from Db2 to SQL Server using SQL Server Migration Assistant (SSMA) for Db2.

Set migration options

Before migrating data to SQL Server, review the project migration options in the Project Settings dialog box.

  • By using this dialog box, you can set options such as migration batch size, table locking, constraint checking, null value handling, and identity value handling. For more information about the Project Migration Settings, see Project Settings (Migration).

Client-side data migration

  • To initiate data-migration on the client side, select the Client Side Data Migration Engine option in the Project Settings dialog box.

  • In Project Settings, the Client Side Data Migration Engine option is set.

    Note

    The Client-Side Data Migration Engine resides inside the SSMA application and is, therefore, not dependent on the availability of the extension pack.

Migrate data to SQL Server

Migrating data is a bulk-load operation that moves rows of data from Db2 tables into SQL Server tables in transactions. The number of rows loaded into SQL Server in each transaction is configured in the project settings.

To view migration messages, make sure that the Output pane is visible. Otherwise, navigate to View > Output.

  1. Verify that:

    • The Db2 providers are installed on the computer that is running SSMA.

    • You synchronized the converted objects with the SQL Server database.

  2. In Db2 Metadata Explorer, select the objects that contain the data that you want to migrate:

    • To migrate data for all schemas, select the check box next to Schemas.

    • To migrate data or omit individual tables, first expand the schema, expand Tables, and then select or clear the check box next to the table.

  3. Right-click Schemas in Db2 Metadata Explorer, and then select Migrate Data. You can also migrate data for individual objects or categories of objects: Right-click the object or its parent folder; select the Migrate Data option.

  4. In the Connect to Db2 dialog box, enter the connection credentials, and then select Connect. For more information on connecting to Db2, see Connect to Db2 database

    For connecting to the target database SQL Server, enter the connection credentials in the Connect to SQL Server dialog box, and select Connect. For more information on connecting to SQL Server, see Connect to SQL Server

    Messages appear in the Output pane. When the migration is complete, the Data Migration Report appears. If any data didn't migrate, select the row that contains the errors, and then select Details. When you're finished with the report, select Close. For more information on Data Migration Report, see Data Migration Report (Db2ToSQL).

Migrate data at scale

When you need to migrate a large amount of data at scale within a short duration, offline migration is a suitable option. You can use the Azure Data Factory pipeline feature to migrate on-premises or external Oracle data sources to Azure SQL Database or Azure SQL Managed Instance (platform as a service, or PaaS), or SQL Server on Azure VM (infrastructure as a service, or IaaS). The process creates a new data factory, or you can use an existing data factory when you migrate to Azure SQL.

  1. For a large offline migration, select Migrate data at scale.

  2. Authenticate with your Microsoft Entra credentials. Once you enter the username, select Connect.

  3. Select the tenant and the subscription that already has the Azure SQL target configured and select Next.

  4. Provide the data source credentials, followed by the target Azure SQL credentials.

  5. For this migration, you can either use existing Azure resources (a resource group and the data factory name), or choose New to create a resource group and Azure data factory for migration.

  6. To migrate an external data source, you need an integration runtime. Either select New and follow the steps to create the integration runtime or use an existing integration service. To create an integration service, you need to configure a gateway installation, and supply a key to configure the integration runtime.

  7. Finally, provide a unique data migration name. This name can be an alphanumeric value only. Avoid any special characters.

  8. If the target tables contain data, they are truncated and reloaded. You will see a warning dialog box. Select OK to proceed, or Cancel to avoid any truncate and load activity.

  9. It takes a few minutes to create the Azure data factory components. A status bar indicates the pipeline creation progress. If the pipeline is created successfully, the following message is written in the output log:

    A data factory pipeline <PipelineName> is created for the data migration activity.
    
  10. To monitor the data migration, select the monitoring URL, or visit the data factory monitoring page in the Azure portal.