Muokkaa

Jaa


Quickstart: Ingest data using Azure Synapse Pipelines (Preview)

In this quickstart, you learn how to load data from a data source into Azure Synapse Data Explorer pool.

Prerequisites

  • An Azure subscription. Create a free Azure account.

  • Create a Data Explorer pool using Synapse Studio or the Azure portal

  • Create a Data Explorer database.

    1. In Synapse Studio, on the left-side pane, select Data.

    2. Select + (Add new resource) > Data Explorer pool, and use the following information:

      Setting Suggested value Description
      Pool name contosodataexplorer The name of the Data Explorer pool to use
      Name TestDatabase The database name must be unique within the cluster.
      Default retention period 365 The time span (in days) for which it's guaranteed that the data is kept available to query. The time span is measured from the time that data is ingested.
      Default cache period 31 The time span (in days) for which to keep frequently queried data available in SSD storage or RAM, rather than in longer-term storage.
    3. Select Create to create the database. Creation typically takes less than a minute.

  • Create a table

    1. In Synapse Studio, on the left-side pane, select Develop.
    2. Under KQL scripts, Select + (Add new resource) > KQL script. On the right-side pane, you can name your script.
    3. In the Connect to menu, select contosodataexplorer.
    4. In the Use database menu, select TestDatabase.
    5. Paste in the following command, and select Run to create the table.
    .create table StormEvents (StartTime: datetime, EndTime: datetime, EpisodeId: int, EventId: int, State: string, EventType: string, InjuriesDirect: int, InjuriesIndirect: int, DeathsDirect: int, DeathsIndirect: int, DamageProperty: int, DamageCrops: int, Source: string, BeginLocation: string, EndLocation: string, BeginLat: real, BeginLon: real, EndLat: real, EndLon: real, EpisodeNarrative: string, EventNarrative: string, StormSummary: dynamic)
    

    Tip

    Verify that the table was successfully created. On the left-side pane, select Data, select the contosodataexplorer more menu, and then select Refresh. Under contosodataexplorer, expand Tables and make sure that the StormEvents table appears in the list.

  • Get the Query and Data Ingestion endpoints. You'll need the query endpoint to configure your linked service.

    1. In Synapse Studio, on the left-side pane, select Manage > Data Explorer pools.

    2. Select the Data Explorer pool you want to use to view its details.

      Screenshot of the Data Explorer pools screen, showing the list of existing pools.

    3. Make a note of the Query and Data Ingestion endpoints. Use the Query endpoint as the cluster when configuring connections to your Data Explorer pool. When configuring SDKs for data ingestion, use the data ingestion endpoint.

      Screenshot of the Data Explorer pools properties pane, showing the Query and Data Ingestion URI addresses.

Create a linked service

In Azure Synapse Analytics, a linked service is where you define your connection information to other services. In this section, you'll create a linked service for Azure Data Explorer.

  1. In Synapse Studio, on the left-side pane, select Manage > Linked services.

  2. Select + New.

    Screenshot of the Linked services screen, showing the list of existing services and highlighting the add new button.

  3. Select the Azure Data Explorer service from the gallery, and then select Continue.

    Screenshot of the new Linked services pane, showing the list of available services and highlighting the add new Azure Data Explorer service.

  4. In the New Linked Services page, use the following information:

    Setting Suggested value Description
    Name contosodataexplorerlinkedservice The name for the new Azure Data Explorer linked service.
    Authentication method Managed Identity The authentication method for the new service.
    Account selection method Enter manually The method for specifying the Query endpoint.
    Endpoint https://contosodataexplorer.contosoanalytics.dev.kusto.windows.net The Query endpoint you made a note of earlier.
    Database TestDatabase The database where you want to ingest data.

    Screenshot of the new Linked services details pane, showing the fields that need to be completed for the new service.

  5. Select Test connection to validate the settings, and then select Create.

Create a pipeline to ingest data

A pipeline contains the logical flow for an execution of a set of activities. In this section, you'll create a pipeline containing a copy activity that ingests data from your preferred source into a Data Explorer pool.

  1. In Synapse Studio, on the left-side pane, select Integrate.

  2. Select + > Pipeline. On the right-side pane, you can name your pipeline.

    Screenshot showing the selection for creating a new pipeline.

  3. Under Activities > Move & transform, drag Copy data onto the pipeline canvas.

  4. Select the copy activity and go to the Source tab. Select or create a new source dataset as the source to copy data from.

  5. Go to the Sink tab. Select New to create a new sink dataset.

    Screenshot of the pipeline copy activity, showing the selection for creating a new sink.

  6. Select the Azure Data Explorer dataset from the gallery, and then select Continue.

  7. In the Set properties pane, use the following information, and then select OK.

    Setting Suggested value Description
    Name AzureDataExplorerTable The name for the new pipeline.
    Linked service contosodataexplorerlinkedservice The linked service you created earlier.
    Table StormEvents The table you created earlier.

    Screenshot of the pipeline copy activity set properties pane, showing the fields that need to be completed for the new sink.

  8. To validate the pipeline, select Validate on the toolbar. You see the result of the Pipeline validation output on the right side of the page.

Debug and publish the pipeline

Once you've finished configuring your pipeline, you can execute a debug run before you publish your artifacts to verify everything is correct.

  1. Select Debug on the toolbar. You see the status of the pipeline run in the Output tab at the bottom of the window.

  2. Once the pipeline run succeeds, in the top toolbar, select Publish all. This action publishes entities (datasets and pipelines) you created to the Synapse Analytics service.

  3. Wait until you see the Successfully published message. To see notification messages, select the bell button in the top-right.

Trigger and monitor the pipeline

In this section, you manually trigger the pipeline published in the previous step.

  1. Select Add Trigger on the toolbar, and then select Trigger Now. On the Pipeline Run page, select OK.

  2. Go to the Monitor tab located in the left sidebar. You see a pipeline run that is triggered by a manual trigger.

  3. When the pipeline run completes successfully, select the link under the Pipeline name column to view activity run details or to rerun the pipeline. In this example, there's only one activity, so you see only one entry in the list.

  4. For details about the copy operation, select the Details link (eyeglasses icon) under the Activity name column. You can monitor details like the volume of data copied from the source to the sink, data throughput, execution steps with corresponding duration, and used configurations.

  5. To switch back to the pipeline runs view, select the All pipeline runs link at the top. Select Refresh to refresh the list.

  6. Verify your data is correctly written in the Data Explorer pool.

Next steps