Редагувати

Поділитися через


Build real-time dashboard with Power BI dataset produced from Stream Analytics no code editor

This article describes how you can use the no code editor to easily create a Stream Analytics job to produce processed data into Power BI dataset. It continuously reads from your Event Hubs, processes and outputs the data into Power BI dataset to build the real-time Power BI dashboard.

Prerequisites

  • Your Azure Event Hubs resources must be publicly accessible and not be behind a firewall or secured in an Azure Virtual Network
  • You should have an existing Power BI workspace, and you have the permission to create dataset there.
  • The data in your Event Hubs must be serialized in either JSON, CSV, or Avro format.

Develop a Stream Analytics job to create Power BI dataset with selected data

  1. In the Azure portal, locate and select the Azure Event Hubs instance.

  2. Select Features > Process Data and then select Start on the Build near real-time data dashboard with Power BI card.

    Screenshot showing the Process data page of an event hub.

  3. Enter a name for the Stream Analytics job, then select Create.

    Screenshot showing where to enter a job name.

  4. Specify the Serialization type of your data in the Event Hubs window and the Authentication method that the job uses to connect to the Event Hubs. Then select Connect.

    Screenshot showing the Event Hubs connection configuration.

  5. When the connection is established successfully and you have data streams flowing into your Event Hubs instance, you immediately see two things:

    • Fields that are present in the input data. You can choose Add field or select the three dot symbol next to a field to remove, rename, or change its type.

      Screenshot showing the Event Hubs field list where you can remove, rename, or change the field type.

    • A live sample of incoming data in the Data preview table under the diagram view. It automatically refreshes periodically. You can select Pause streaming preview to see a static view of the sample input data.

      Screenshot showing sample data under Data Preview.

  6. Select the Manage tile. In the Manage fields configuration panel, choose the fields you want to output. If you want to add all the fields, select Add all fields.

    Screenshot that shows the Manage field operator configuration.

  7. Select Power BI tile. In the Power BI configuration panel, fill in needed parameters and connect.

    • Dataset: it's the Power BI destination where the Azure Stream Analytics job output data is written into.
    • Table: it's the table name in the Dataset where the output data goes to.

    Screenshot that shows the Power BI output configuration.

  8. Optionally, select Get static preview/Refresh static preview to see the data preview that will be ingested in event hub.
    Screenshot showing the Get static preview/Refresh static preview option.

  9. Select Save and then select Start the Stream Analytics job.
    Screenshot showing the Save and Start options.

  10. To start the job, specify:

    • The number of Streaming Units (SUs) the job runs with. SUs represents the amount of compute and memory allocated to the job. We recommended that you start with three and then adjust as needed.
    • Output data error handling – It allows you to specify the behavior you want when a job’s output to your destination fails due to data errors. By default, your job retries until the write operation succeeds. You can also choose to drop such output events.
      Screenshot showing the Start Stream Analytics job options where you can change the output time, set the number of streaming units, and select the Output data error handling options.
  11. After you select Start, the job starts running within two minutes, and the metrics will be open in tab section.

    Screenshot that shows the job metrics after it's started.

    You can also see the job under the Process Data section on the Stream Analytics jobs tab. Select Open metrics to monitor it or stop and restart it, as needed.

    Screenshot of the Stream Analytics jobs tab where you view the running jobs status.

Build the real-time dashboard in Power BI

Now, you have the Azure Stream Analytics job running and the data is continuously written into the table in the Power BI dataset you've configured. You can now create the real-time dashboard in Power BI workspace.

  1. Go to the Power BI workspace, which you've configured in above Power BI output tile, and select + New in the top left corner, then choose Dashboard to give the new dashboard a name. Screenshot of the pbi dashboard creation.
  2. Once the new dashboard is created, you'll be led to the new dashboard. Select Edit, and choose + Add a tile in the top menu bar. A right pane is open. Select Custom Streaming Data to go to next page. Screenshot of the pbi dashboard adding tile.
  3. Select the streaming dataset (for example nocode-pbi-demo-xujx) which you've configured in Power BI node, and go to next page. Screenshot of the pbi dashboard adding tile with selected dataset.
  4. Fill in the tile details, and follow the next step to complete the tile configuration. Screenshot of the pbi dashboard adding tile with configured details.
  5. Then, you can adjust its size and get the continuously updated dashboard as shown in the following example. Screenshot of the pbi dashboard report.

Considerations when using the Event Hubs Geo-replication feature

Azure Event Hubs recently launched the Geo-Replication feature in public preview. This feature is different from the Geo Disaster Recovery feature of Azure Event Hubs.

When the failover type is Forced and replication consistency is Asynchronous, Stream Analytics job doesn't guarantee exactly once output to an Azure Event Hubs output.

Azure Stream Analytics, as producer with an event hub an output, might observe watermark delay on the job during failover duration and during throttling by Event Hubs in case replication lag between primary and secondary reaches the maximum configured lag.

Azure Stream Analytics, as consumer with Event Hubs as Input, might observe watermark delay on the job during failover duration and might skip data or find duplicate data after failover is complete.

Due to these caveats, we recommend that you restart the Stream Analytics job with appropriate start time right after Event Hubs failover is complete. Also, since Event Hubs Geo-replication feature is in public preview, we don't recommend using this pattern for production Stream Analytics jobs at this point. The current Stream Analytics behavior will improve before the Event Hubs Geo-replication feature is generally available and can be used in Stream Analytics production jobs.

Next steps

Learn more about Azure Stream Analytics and how to monitor the job you've created.