Edit

Share via


Move data from Oracle to Fabric Lakehouse via Pipeline and On-premises Data Gateway

In this tutorial, you build a data pipeline to move data from an on-premises Oracle database to a Lakehouse destination.

Prerequisites

To start, you must complete the following prerequisites:

  • Install an on-premises data gateway in your local environment. You can get more details about how to install and configure an on-premises data gateway here: Install an on-premises data gateway

Note

Fabric pipeline activities can only be executed by a version of the on-premises data gateway that is 3000.222.5 or higher.

Create a data pipeline

  1. Navigate to Power BI.

  2. Select the Power BI icon in the bottom left of the screen, then select Data factory to open homepage of Data Factory.

  3. Select Data pipeline and then input a pipeline name to create a new pipeline.

    Screenshot showing the new data pipeline button in the newly created workspace.

    Screenshot showing the name of creating a new pipeline.

Copy data using data pipeline

In this session, you start to build your data pipeline by following below steps about copying data from on-premises Oracle database to Lakehouse.

Add a copy activity

  1. Open an existing data pipeline or create a new data pipeline.

  2. Add a copy activity either by selecting Add pipeline activity > Copy activity or by selecting Copy data > Add to canvas under the Activities tab.

    Screenshot showing two ways to add a copy activity.

Configure your source under the source tab

  1. Create a new connection to your data source.

    Screenshot showing where to create a connection.

    a. Choose the data source type from the pop-up window. Select Oracle Database, and then select Continue.

    Screenshot showing how to select the data source.

    b. It navigates to the connection creation page. Fill in the required connection information on the panel, and then select Create. For the details of connection creation for each type of data source, you can refer to each connector article.

    Screenshot showing New connection page.

    c. Specify a table to be copied. Select Preview data to preview your source table. You can also use Query and Stored procedure to read data from your source.

Configure your destination under the destination tab

  1. Choose your destination type. It could be either your internal first class data store from your workspace, such as Lakehouse, or your external data stores. You use Lakehouse as an example.

  2. Choose to use Lakehouse in Workspace data store type. Select + New, and it navigates you to the Lakehouse creation page. Specify your Lakehouse name and then select Create.

    Screenshot showing Lakehouse creation.

  3. Specify a table or set up the file path to define the file or folder as the destination. Here select Tables and specify a table to write data.

    Screenshot showing where to find Table settings.

Run and schedule your data pipeline

After completing the configuration of your data pipeline, run the data pipeline to trigger the copy activity. You can also schedule your data pipeline run if needed.

  1. Switch to the Home tab and select Run. A confirmation dialog is displayed. Then select Save and run to start the activity.

    Screenshot of saving and running activity.

  2. You can monitor the running process and check the results on the Output tab below the data pipeline canvas. Select the run details button (with the glasses icon highlighted) to view the run details.

    Screenshot of the output of the data pipeline.

  3. You can also schedule the data pipeline to run with a specific frequency as required. Below is an example scheduling the data pipeline to run every 15 minutes. You can also specify the Start time and End time for your schedule. If you don't specify a start time, the start time is the time your schedule applies. If you don't specify an end time, your data pipeline run keeps recurring every 15 minutes.

    Screenshot of scheduling the data pipeline.