Data Flow Duration

Derek Horrall 201 Reputation points
2020-07-08T15:19:07.683+00:00

I can't figure out where the bottleneck is in my Mapping Data Flow. The duration says it took around 15 minutes to run, but if I look at the details, it only seems that actual sink processing time is 4 minutes. I don't understand where the remaining 11 minutes is spent. I suspect I am being charged for these 11 mystery minutes.

Perhaps I am interpreting the 'data flow' details visual incorrectly?

11624-data-flow-timing.png

11614-data-flow-timing-total.png

Azure Data Factory
Azure Data Factory
An Azure service for ingesting, preparing, and transforming data at scale.
{count} votes

1 answer

Sort by: Most helpful
  1. MarkKromer-MSFT 5,231 Reputation points Microsoft Employee Moderator
    2020-07-09T17:27:09.073+00:00

    Click on the Sink in your data flow in the monitoring UI. The right-hand panel will show the total processing time at the bottom.

    Now, look at the "Stage time" in that same fly-in panel. The difference between those 2 durations is the amount of time it took Databricks to write your data to your destination linked service and perform any post-execution clean-up.

    And, yes, before you ask ... we can make this better and clearer and we are working on it! :)

    QQ: Are you performing any post-execution scripts or sink to single file in your sink configuration?


Your answer

Answers can be marked as 'Accepted' by the question author and 'Recommended' by moderators, which helps users know the answer solved the author's problem.