An Azure service for ingesting, preparing, and transforming data at scale.
Hi Sudhakar P
Thanks for the detailed overview of your Azure Data Factory (ADF) and Databricks pipeline. It's great that you're looking to optimize both performance and cost. Based on your current setup and the screenshot you shared, here are some suggestions:
Performance optimization with Parallel Execution
You are already on the right track using a ForEach loop with Sequential unchecked, which enables parallel execution.
To avoid overwhelming the Databricks cluster or exceeding ADF concurrency limits, I recommend setting a controlled Batch count - for example, start with 5 or 10 depending on the cluster size and table workload.
You can dynamically process tables in parallel by grouping them logically (e.g., dimensions vs. facts, or Bronze vs. Silver vs. Gold layers) and creating separate ForEach loops for each group.
Cost management tips
Each Databricks Notebook activity in ADF incurs compute and activity cost. Running too many in parallel can spike costs.
Use ADF's cost monitoring (via Azure Cost Management or Log Analytics) to track how activity parallelism impacts daily spend.
To optimize Databricks usage:
- Reuse existing clusters with auto-scaling and auto-termination enabled.
- Avoid spinning up separate clusters per notebook if not required.
- Use Job Clusters only when isolation is needed.
Architecture & Design best practices
Consider breaking the pipeline into modular stages: e.g., one pipeline for Bronze load, one for Silver, and one for Gold. Trigger each stage conditionally or sequentially.
Use pipeline parameters and metadata-driven design so that logic remains dynamic and maintainable.
For high-volume tables, isolate them into their own ForEach loop or child pipeline to prevent bottlenecks.
I hope this information helps. Please do let us know if you have any further queries.
Kindly consider upvoting the comment if the information provided is helpful. This can assist other community members in resolving similar issues.
Thank you.