Jaa


Migrate Azure Data Lake Analytics to Azure Synapse Analytics

Azure Data Lake Analytics will be retired on 29 February 2024. Learn more with this announcement.

If you're already using Azure Data Lake Analytics, you can create a migration plan to Azure Synapse Analytics for your organization.

Microsoft launched Azure Synapse Analytics that aims at bringing both data lakes and data warehouse together for a unique big data analytics experience. It will help you gather and analyze your data to solve data inefficiency, and help your teams work together. Moreover, Synapse’s integration with Azure Machine Learning and Power BI will allow the improved ability for organizations to get insights from its data and execute machine learning to all its smart apps.

The document shows you how to do the migration from Azure Data Lake Analytics to Azure Synapse Analytics.

  • Step 1: Assess readiness
  • Step 2: Prepare to migrate
  • Step 3: Migrate data and application workloads
  • Step 4: Cutover from Azure Data Lake Analytics to Azure Synapse Analytics

Step 1: Assess readiness

  1. Look at Apache Spark on Azure Synapse Analytics, and understand key differences of Azure Data Lake Analytics and Spark on Azure Synapse Analytics.

    Item Azure Data Lake Analytics Spark on Synapse
    Pricing Per Analytic Unit-hour Per vCore-hour
    Engine Azure Data Lake Analytics Apache Spark
    Default Programing Language U-SQL T-SQL, Python, Scala, Spark SQL and .NET
    Data Sources Azure Data Lake Storage Azure Blob Storage, Azure Data Lake Storage
  2. Review the Questionnaire for Migration Assessment and list those possible risks for considering.

Step 2: Prepare to migrate

  1. Identify jobs and data that you'll migrate.

    • Take this opportunity to clean up those jobs that you no longer use. Unless you plan to migrate all your jobs at one time, take this time to identify logical groups of jobs that you can migrate in phases.
    • Evaluate the size of the data and understand Apache Spark data format. Review your U-SQL scripts and evaluate the scripts rewriting efforts and understand the Apache Spark code concept.
  2. Determine the impact that a migration will have on your business. For example, whether you can afford any downtime while migration takes place.

  3. Create a migration plan.

Step 3: Migrate data and application workload

  1. Migrate your data from Azure Data Lake Storage Gen1 to Azure Data Lake Storage Gen2.

    Azure Data Lake Storage Gen1 retirement will be in February 2024, see the official announcement. We’d suggest migrating the data to Gen2 in the first place. See Understand Apache Spark data formats for Azure Data Lake Analytics U-SQL developers and move both the file and the data stored in U-SQL tables to make them accessible to Azure Synapse Analytics. More details of the migration guide can be found here.

  2. Transform your U-SQL scripts to Spark. Refer to Understand Apache Spark code concepts for Azure Data Lake Analytics U-SQL developers to transform your U-SQL scripts to Spark.

  3. Transform or re-create your job orchestration pipelines to new Spark program.

Step 4: Cut over from Azure Data Lake Analytics to Azure Synapse Analytics

After you're confident that your applications and workloads are stable, you can begin using Azure Synapse Analytics to satisfy your business scenarios. Turn off any remaining pipelines that are running on Azure Data Lake Analytics and retire your Azure Data Lake Analytics accounts.

Questionnaire for Migration Assessment

Category Questions Reference
Evaluate the size of the Migration How many Azure Data Lake Analytics accounts do you have? How many pipelines are in use? How many U-SQL scripts are in use? The more data and scripts to be migrated, the more UDO/UDF are used in scripts, the more difficult it is to migrate. The time and resources required for migration need to be well planned according to the scale of the project.
Data source What’s the size of the data source? What kinds of data format for processing? Understand Apache Spark data formats for Azure Data Lake Analytics U-SQL developers
Data output Will you keep the output data for later use? If the output data is saved in U-SQL tables, how to handle it? If the output data will be used often and saved in U-SQL tables, you need change the scripts and change the output data to Spark supported data format.
Data migration Have you made the storage migration plan? Migrate Azure Data Lake Storage from Gen1 to Gen2
U-SQL scripts transform Do you use UDO/UDF (.NET, python, etc.)?If above answer is yes, which language do you use in your UDO/UDF and any problems for the transform during the transform?Is the federated query being used in U-SQL? Understand Apache Spark code concepts for Azure Data Lake Analytics U-SQL developers

Next steps