Uredi

Deli z drugimi prek


Load data into Azure Data Lake Storage Gen2 with Azure Data Factory

APPLIES TO: Azure Data Factory Azure Synapse Analytics

Tip

Try out Data Factory in Microsoft Fabric, an all-in-one analytics solution for enterprises. Microsoft Fabric covers everything from data movement to data science, real-time analytics, business intelligence, and reporting. Learn how to start a new trial for free!

Azure Data Lake Storage Gen2 is a set of capabilities dedicated to big data analytics, built into Azure Blob storage. It allows you to interface with your data using both file system and object storage paradigms.

Azure Data Factory (ADF) is a fully managed cloud-based data integration service. You can use the service to populate the lake with data from a rich set of on-premises and cloud-based data stores and save time when building your analytics solutions. For a detailed list of supported connectors, see the table of Supported data stores.

Azure Data Factory offers a scale-out, managed data movement solution. Due to the scale-out architecture of ADF, it can ingest data at a high throughput. For details, see Copy activity performance.

This article shows you how to use the Data Factory Copy Data tool to load data from Amazon Web Services S3 service into Azure Data Lake Storage Gen2. You can follow similar steps to copy data from other types of data stores.

Tip

For copying data from Azure Data Lake Storage Gen1 into Gen2, refer to this specific walkthrough.

Prerequisites

  • Azure subscription: If you don't have an Azure subscription, create a free account before you begin.
  • Azure Storage account with Data Lake Storage Gen2 enabled: If you don't have a Storage account, create an account.
  • AWS account with an S3 bucket that contains data: This article shows how to copy data from Amazon S3. You can use other data stores by following similar steps.

Create a data factory

  1. If you have not created your data factory yet, follow the steps in Quickstart: Create a data factory by using the Azure portal and Azure Data Factory Studio to create one. After creating it, browse to the data factory in the Azure portal.

    Home page for the Azure Data Factory, with the Open Azure Data Factory Studio tile.

  2. Select Open on the Open Azure Data Factory Studio tile to launch the Data Integration application in a separate tab.

Load data into Azure Data Lake Storage Gen2

  1. In the home page of Azure Data Factory, select the Ingest tile to launch the Copy Data tool.

  2. In the Properties page, choose Built-in copy task under Task type, and choose Run once now under Task cadence or task schedule, then select Next.

    Properties page

  3. In the Source data store page, complete the following steps:

    1. Select + New connection. Select Amazon S3 from the connector gallery, and select Continue.

      Source data store s3 page

    2. In the New connection (Amazon S3) page, do the following steps:

      1. Specify the Access Key ID value.
      2. Specify the Secret Access Key value.
      3. Select Test connection to validate the settings, then select Create.

      Specify Amazon S3 account

    3. In the Source data store page, ensure that the newly created Amazon S3 connection is selected in the Connection block.

    4. In the File or folder section, browse to the folder and file that you want to copy over. Select the folder/file, and then select OK.

    5. Specify the copy behavior by checking the Recursively and Binary copy options. Select Next.

    Screenshot that shows the source data store page.

  4. In the Destination data store page, complete the following steps.

    1. Select + New connection, and then select Azure Data Lake Storage Gen2, and select Continue.

      Destination data store page

    2. In the New connection (Azure Data Lake Storage Gen2) page, select your Data Lake Storage Gen2 capable account from the "Storage account name" drop-down list, and select Create to create the connection.

      Specify Azure Data Lake Storage Gen2 account

    3. In the Destination data store page, select the newly created connection in the Connection block. Then under Folder path, enter copyfroms3 as the output folder name, and select Next. ADF will create the corresponding ADLS Gen2 file system and subfolders during copy if it doesn't exist.

      Screenshot that shows the destination data store page.

  5. In the Settings page, specify CopyFromAmazonS3ToADLS for the Task name field, and select Next to use the default settings.

    Settings page

  6. In the Summary page, review the settings, and select Next.

    Summary page

  7. On the Deployment page, select Monitor to monitor the pipeline (task).

  8. When the pipeline run completes successfully, you see a pipeline run that is triggered by a manual trigger. You can use links under the Pipeline name column to view activity details and to rerun the pipeline.

    Monitor pipeline runs

  9. To see activity runs associated with the pipeline run, select the CopyFromAmazonS3ToADLS link under the Pipeline name column. For details about the copy operation, select the Details link (eyeglasses icon) under the Activity name column. You can monitor details like the volume of data copied from the source to the sink, data throughput, execution steps with corresponding duration, and used configuration.

    Monitor activity runs

    Monitor activity run details

  10. To refresh the view, select Refresh. Select All pipeline runs at the top to go back to the "Pipeline runs" view.

  11. Verify that the data is copied into your Data Lake Storage Gen2 account.