Quickstart: Create your first pipeline to copy data
In this quickstart, you build a data pipeline to move a Sample dataset to the Lakehouse. This experience shows you a quick demo about how to use pipeline copy activity and how to load data into Lakehouse.
Prerequisites
To get started, you must complete the following prerequisites:
- A Microsoft Fabric tenant account with an active subscription. Create a free account.
- Make sure you have a Microsoft Fabric enabled Workspace: Create a workspace.
Create a data pipeline
Navigate to Power BI.
Select the Power BI icon in the bottom left of the screen, then select Data factory to open homepage of Data Factory.
Navigate to your Microsoft Fabric workspace. If you created a new workspace in the prior Prerequisites section, use this one.
Select Data pipeline and then input a pipeline name to create a new pipeline.
Copy data using pipeline
In this session, you start to build your first pipeline by following below steps about copying from a sample dataset provided by pipeline into Lakehouse.
Step 1: Start with the Copy assistant
After selecting Copy data on the canvas, the Copy assistant tool will be opened to get started.
Step 2: Configure your source
Choose the Public Holidays sample data, and then select Next.
On the Connect to data source page of the assistant, the preview for the Public Holidays sample data is displayed, and then click Next.
Step 3: Configure your destination
Select Lakehouse and then Next.
Select Create new Lakehouse, and enter a Lakehouse name, then select Next.
Configure and map your source data to the destination Lakehouse table. Select tables and provide a Table name under Root folder, then choose the Overwrite option for Table action, and select Next.
Step 4: Review and create your copy activity
Review your copy activity settings in the previous steps and select OK to finish. Or you can revisit the previous steps in the tool to edit your settings, if needed.
The Copy activity is added to your new data pipeline canvas. All settings including advanced settings for the activity are available in the tabs below the pipeline canvas when the created Copy data activity is selected.
Run and schedule your data pipeline
Switch to the Home tab and select Run. A confirmation dialog is displayed. Then select Save and run to start the activity.
You can monitor the running process and check the results on the Output tab below the pipeline canvas. Select the run details button (with the glasses icon highlighted) to view the run details.
The run details show how much data was read and written and various other details about the run.
You can also schedule the pipeline to run with a specific frequency as required. Below is an example scheduling the pipeline to run every 15 minutes.
Related content
The pipeline in this sample shows you how to copy sample data to Lakehouse. You learned how to:
- Create a data pipeline.
- Copy data with the Copy Assistant.
- Run and schedule your data pipeline.
Next, advance to learn more about monitoring your pipeline runs.
Atsauksmes
https://aka.ms/ContentUserFeedback.
Drīzumā: 2024. gada laikā mēs pakāpeniski pārtrauksim izmantot “GitHub problēmas” kā atsauksmju par saturu mehānismu un aizstāsim to ar jaunu atsauksmju sistēmu. Papildinformāciju skatiet:Iesniegt un skatīt atsauksmes par