Hi @Dankerlui, Patrick
Thank you for your question and using the Microsoft Q&A platform.
Your understanding is correct. In Azure Synapse Analytics, Data Flows and Pipelines do not directly run on the Spark pools that you create. Instead, they operate on integration runtimes, which are managed environments where the execution of these activities takes place.
The Spark version for integration runtimes is indeed managed automatically by Microsoft. This means that as long as you are using Data Flows or Pipelines, the underlying Spark version is updated by Microsoft without requiring manual intervention. You do not need to manually upgrade Spark versions for integration runtimes as you would for custom Spark pools.
To summarize:
- Spark Pools: These are manually managed, and you have to create or update them to use specific Spark versions (like 3.4).
- Integration Runtimes: The Spark version is automatically managed, and Microsoft will handle updates, including the transition from Spark 3.3 to 3.4.
For more information, please refer to the below documents:
https://learn.microsoft.com/en-us/azure/synapse-analytics/spark/apache-spark-version-support
https://learn.microsoft.com/en-us/azure/synapse-analytics/spark/apache-spark-33-runtime
Hope this helps. Do let us know if you any further queries.
If this answers your query, do click Accept Answer
and Yes
for was this answer helpful. And, if you have any further query do let us know.