SQL Server Big Data Clusters Spark 3 upgrade guide
Applies to: SQL Server 2019 (15.x)
Important
The Microsoft SQL Server 2019 Big Data Clusters add-on will be retired. Support for SQL Server 2019 Big Data Clusters will end on February 28, 2025. All existing users of SQL Server 2019 with Software Assurance will be fully supported on the platform and the software will continue to be maintained through SQL Server cumulative updates until that time. For more information, see the announcement blog post and Big data options on the Microsoft SQL Server platform.
This article contains important information and guidance for migrating Apache Spark 2.4 workloads to Spark version 3.1. This is required in order to upgrade from SQL Server Big Data Clusters CU12 to CU13, and above.
Introduction of Apache Spark 3 on SQL Server Big Data Clusters
Up to cumulative update 12 (CU12), big data clusters relied on the Apache Spark 2.4 line, which reached its end of life in May 2021. Consistent with our commitment to continuous improvements of the Big Data and Machine Learning capabilities brought by the Apache Spark engine, CU13 brings in the current release of Apache Spark, version 3.1.2.
A new performance baseline
This new version of Apache Spark brings performance benefits over big data processing workloads. Using the reference TCP-DS 10TB workload in our tests we were able to reduce runtime from 4.19 hours to 2.96 hours, a 29.36% improvement achieved just by switching engines using the same hardware and configuration profile on SQL Server Big Data Clusters, no additional application optimizations. The improvement mean of individual query runtime is 36%.
Upgrade guidance
Spark 3 is a major release and contains breaking changes. Following the same established best practice in the SQL Server universe, it's recommended:
- Review this article entirely.
- Review the official Apache Spark 3 Migration Guide.
- Perform a side-by-side deployment of a new big data cluster version CU13 with your current environment.
- (Optional) Leverage the new azdata HDFS distributed copy capability to have a subset of your data needed for validation.
- Validate your current workload with Spark 3 before upgrading.
- Reassess enforced Spark optimizations in your code and table definition strategies. Spark 3 brings new shuffling, partitioning, and Adaptive Query Execution enhancements. This is a great opportunity to reevaluate previous decisions and try leverage the newer engine out-of-the-box features.
What happens during cluster upgrade?
The cluster upgrade process will deploy Spark pods with the new version and the refreshed runtime for Apache Spark. After the upgrade, there will be no Spark 2.4 components any longer.
Persistent configuration changes made through the configuration framework will be preserved.
User libraries and artifacts loaded directly into HDFS will be preserved. Yet, make sure that those libraries and artifacts are compatible with Spark 3.
Warning
Customizations made directly to the pods will be lost, make sure you validate and re-apply those if still applicable to Spark 3.
Breaking changes
Spark 3 is not fully backward compatible with 2.4, the breaking changes are mainly caused by three parts:
- Scala 2.12 used by Spark 3 is incompatible with Scala 2.11 used by Spark 2.4
- Spark 3 API changes and deprecations
- SQL Server Big Data Clusters runtime for Apache Spark library updates
Scala 2.12 used by Spark 3 is incompatible with Scala 2.11
If running Spark jobs based on Scala 2.11 jars, it is required to rebuild it using Scala 2.12. Scala 2.11 and 2.12 are mostly source compatible, but not binary compatible. For more information, see Scala 2.12.0.
Below changes are needed:
- Change Scala version for all Scala dependencies.
- Change Spark version for all Spark dependencies.
- Change all Spark dependencies have provided scope except external dependencies such as
spark-sql-kafka-0-10
.
Here is a sample pom.xml as below:
<properties>
<spark.version>3.1.2</spark.version>
<scala.version.major>2.12</scala.version.major>
<scala.version.minor>10</scala.version.minor>
<scala.version>${scala.version.major}.${scala.version.minor}</scala.version>
</properties>
<dependencies>
<dependency>
<groupId>org.scala-lang</groupId>
<artifactId>scala-library</artifactId>
<version>${scala.version}</version>
<scope>provided</scope>
</dependency>
<dependency>
<groupId>org.apache.spark</groupId>
<artifactId>spark-core_${scala.version.major}</artifactId>
<version>${spark.version}</version>
<scope>provided</scope>
</dependency>
<dependency>
<groupId>org.apache.spark</groupId>
<artifactId>spark-sql_${scala.version.major}</artifactId>
<version>${spark.version}</version>
<scope>provided</scope>
</dependency>
<dependency>
<groupId>org.apache.spark</groupId>
<artifactId>spark-sql-kafka-0-10_${scala.version.major}</artifactId>
<version>${spark.version}</version>
</dependency>
</dependencies>
Spark 3 API changes and deprecations
Review the official Apache Spark 3 Migration Guide, which covers all API changes in detail.
Some captured highlights are:
Breaking change | Action |
---|---|
spark-submit parameters yarn-client and yarn-clustermodes are removed in Spark 3 |
Use spark-submit --master yarn --deploy-mode client or --deploy-mode cluster instead.Details refer https://spark.apache.org/docs/latest/running-on-yarn.html |
HiveContext class is removed |
Use SparkSession.builder.enableHiveSupport() instead |
The order of argument is reversed in the TRIM method | Use TRIM(str, trimStr) instead of TRIM(trimStr, str) |
Due to the upgrade to Scala 2.12, DataStreamWriter.foreachBatch is not a compatible source with Scala program |
Update your Scala source code to distinguish between Scala function and Java lambda. |
SQL Server Big Data Clusters runtime for Apache Spark library updates
As covered by the SQL Server Big Data Clusters runtime for Apache Spark specification, all default Python, R, and Scala libraries were updated on the CU13 release. Also, many libraries were added to provide a better out-of-the-box experience.
- Make sure that your workload works with the newer library set.
- Review if a custom loaded library is now part of the default package baseline, and adjust your jobs specifications to remove the custom library to allow the job to use the shipped library.
Frequently Asked Questions
How to solve strange java.lang.NoSuchMethodError or java.lang.ClassNotFoundException
This error is most likely caused by Spark or Scala version conflict. Double-check the below and rebuild your project.
- Make sure all Scala versions are updated.
- Make sure all Spark dependencies are updated with correct scala version and spark version.
- Make sure all Spark dependencies have provided scope except spark-sql-kafka-0-10.
SparkUpgradeException due to calendar mode change
There are changes in Spark 3.0 calendar model. You may see exception like this when writing calendar column in Spark SQL:
Caused by: org.apache.spark.SparkUpgradeException:
You may get a different result due to the upgrading of Spark 3.0:
writing dates before 1582-10-15 or timestamps before 1900-01-01T00:00:00Z into Parquet INT96 files can be dangerous,
as the files may be read by Spark 2.x or legacy versions of Hive later,
which uses a legacy hybrid calendar that is different from Spark 3.0+'s Proleptic Gregorian calendar.
See more details in SPARK-31404.
You can set spark.sql.legacy.parquet.int96RebaseModeInWrite to 'LEGACY' to
rebase the datetime values w.r.t. the calendar difference during writing, to get maximum interoperability.
Or set spark.sql.legacy.parquet.int96RebaseModeInWrite to 'CORRECTED' to
write the datetime values as it is, if you are 100% sure that the written files
will only be read by Spark 3.0+ or other systems that use Proleptic Gregorian calendar.
Solution: Set configuration spark.sql.legacy.parquet.int96RebaseModeInWrite to LEGACY or CORRECTED, as explained above. Below is a possible solution in PySpark code:
spark.conf.set("spark.sql.legacy.parquet.int96RebaseModeInWrite","CORRECTED")
Next steps
For more information, see Introducing SQL Server Big Data Clusters.