Редактиране

Споделяне чрез


Quickstart: Create an Apache Airflow Job

Note

Apache Airflow job is powered by Apache Airflow.
Apache Airflow is an open-source platform used to programmatically create, schedule, and monitor complex jobs. It allows you to define a set of tasks, called operators, that can be combined into directed acyclic graphs (DAGs) to represent data pipelines.

Apache Airflow Job provides a simple and efficient way to create and manage Apache Airflow environments, enabling you to run your orchestration jobs at scale with ease. In this quickstart, let's create a simple Apache Airflow job to familiarize yourself with the environment and functionalities of Apache Airflow Job.

Prerequisites

  • Enable Apache Airflow Job in your Tenant.

Note

Since Apache Airflow job is in preview state, you need to enable it through your tenant admin. If you already see Apache Airflow Job, your tenant admin may have already enabled it.

  1. Go to Admin Portal -> Tenant Settings -> Under Microsoft Fabric -> Expand 'Users can create and use Apache Airflow Job (preview)' section.
  2. Select Apply.

Screenshot to enable Apache Airflow in tenant.

Create an Apache Airflow Job

  1. You can use an existing workspace or Create a new workspace.

  2. Expand + New dropdown -> Click on More Options -> Under Data Factory section -> Select Apache Airflow Job (preview)

    Screenshot shows click on more options.

    Screenshot to select Apache Airflow Job.

  3. Give a suitable name to your project and click on the "Create" button.

Create a DAG File

  1. Click on "New DAG file" card -> give the name to the file and Click on "Create" button.

    Screenshot to name the DAG file.

  2. A boilerplate DAG code is presented to you. You can edit the file as per your requirements.

    Screenshot presents boilerplate DAG file in Microsoft Fabric.

  3. Click on "Save icon".

    Screenshot presents how to save DAG file in Microsoft Fabric.

Run a DAG

  1. Begin by clicking on the "Run DAG" button.

    Screenshot to run the DAG from data workflows UI.

  2. Once initiated, a notification will promptly appear indicating the DAG is running.

  3. To monitor the progress of the DAG run, simply click on "View Details" within the notification center. This action will redirect you to the Apache Airflow UI, where you can conveniently track the status and details of the DAG run.

    Screenshot to navigate to Apache Airflow UI from notification center.

Monitor your Apache Airflow DAG in Apache Airflow UI

The saved dag files are loaded in the Apache Airflow UI. You can monitor them by clicking on the "Monitor in Apache Airflow" button.

Screenshot to monitor the Airflow DAG.

Screenshot presents the loaded Airflow DAG.