Edit

Share via


Move data from Oracle to Fabric Lakehouse via Pipeline and On-premises Data Gateway

In this tutorial, you build a pipeline to move data from an on-premises Oracle database to a Lakehouse destination.

Prerequisites

To start, you must complete the following prerequisites:

  • Install an on-premises data gateway in your local environment. You can get more details about how to install and configure an on-premises data gateway here: Install an on-premises data gateway

Note

Fabric pipeline activities can only be executed by a version of the on-premises data gateway that is 3000.222.5 or higher.

Create a pipeline

  1. Navigate to Power BI.

  2. Select the Power BI icon in the bottom left of the screen, then select Fabric to open homepage of Data Factory.

  3. Select + New item.

  4. Search for and select Pipeline and then input a pipeline name to create a new pipeline. to create a new pipeline.

    Screenshot showing the new pipeline button in the newly created workspace.

    Screenshot showing the name of creating a new pipeline.

Copy data using pipeline

In this session, you start to build your pipeline by following below steps about copying data from on-premises Oracle database to Lakehouse.

Add a copy activity

  1. Open an existing pipeline or create a new pipeline.

  2. Add a copy activity either by selecting Add pipeline activity > Copy activity or by selecting Copy data > Add to canvas under the Activities tab.

    Screenshot showing two ways to add a copy activity.

Configure your source under the source tab

  1. Create a new connection to your data source.

    Screenshot showing where to create a connection.

    a. Choose the data source type from the pop-up window. Select Oracle Database, and then select Continue.

    Screenshot showing how to select the data source.

    b. It navigates to the connection creation page. Fill in the required connection information on the panel, and then select Create. For the details of connection creation for each type of data source, you can refer to each connector article.

    Screenshot showing New connection page.

    c. Specify a table to be copied. Select Preview data to preview your source table. You can also use Query and Stored procedure to read data from your source.

Configure your destination under the destination tab

  1. Choose your destination type. It could be either your internal first class data store from your workspace, such as Lakehouse, or your external data stores. You use Lakehouse as an example.

  2. Choose to use Lakehouse in Workspace data store type. Select + New, and it navigates you to the Lakehouse creation page. Specify your Lakehouse name and then select Create.

    Screenshot showing Lakehouse creation.

  3. Specify a table or set up the file path to define the file or folder as the destination. Here select Tables and specify a table to write data.

Run and schedule your pipeline

After completing the configuration of your pipeline, run the pipeline to trigger the copy activity. You can also schedule your pipeline run if needed.

  1. Switch to the Home tab and select Run. A confirmation dialog is displayed. Then select Save and run to start the activity.

    Screenshot of saving and running activity.

  2. You can monitor the running process and check the results on the Output tab below the pipeline canvas. Select the run details button (with the glasses icon highlighted) to view the run details.

    Screenshot of the output of the pipeline.

  3. You can also schedule the pipeline to run with a specific frequency as required. Below is an example scheduling the pipeline to run every 15 minutes. You can also specify the Start time and End time for your schedule. If you don't specify a start time, the start time is the time your schedule applies. If you don't specify an end time, your pipeline run keeps recurring every 15 minutes.

    Screenshot of scheduling the pipeline.