Hello @Hanna
One way to achieve this is by using Azure Stream Analytics to stream the Data Factory activity logs to Databricks Delta table. Here are the high-level steps to achieve this:
- Create a Delta table in Databricks to store the Data Factory activity logs.
- Create an Azure Stream Analytics job to stream the Data Factory activity logs to the Delta table.
- Configure the Stream Analytics job to run every 5 minutes.
- Verify that the logs are being inserted into the Delta table. Here is an example of how you can create a Delta table in Databricks:
CREATE TABLE activity_logs ( eventTime timestamp, operationName string, status string, activityId string, pipelineName string, runId string, message string ) USING DELTA LOCATION '/mnt/delta/activity_logs';
Once you have created the Delta table, you can use Azure Stream Analytics to stream the Data Factory activity logs to the table. Here is an example of how you can configure the Stream Analytics job:
- In the Azure portal, navigate to your Stream Analytics job.
- Click on Inputs and then click on Add.
- Select Data Factory as the input source and configure the input settings.
- Click on Outputs and then click on Add.
- Select Databricks Delta as the output sink and configure the output settings.
- Click on Query and then enter the following query:
SELECT CAST(eventTime AS timestamp) AS eventTime, operationName, status, activityId, pipelineName, runId, message INTO activity_logs FROM DataFactoryInput
- Click on Save and then click on Start to start the Stream Analytics job. Once the job is running, it will stream the Data Factory activity logs to the Delta table every 5 minutes. You can then use Databricks to create a dashboard and monitor the activity logs in real-time.
I hope that this response has addressed your query and helped you overcome your challenges. If so, please mark this response as Answered. This will not only acknowledge our efforts, but also assist other community members who may be looking for similar solutions.