Since you're not using Hive Metastore and are running into issues with the Hive Warehouse Connector (HWC) on HDInsight 5.1, there are a few ways you can disable or remove it:
- You can prevent the Hive Warehouse Connector from loading by setting the following in your Spark configuration:
--conf spark.driver.extraClassPath="" --conf spark.executor.extraClassPath=""
- If you have cluster access, navigate to the directory where the JAR is located (typically:
/usr/hdp/current/spark2-client/jars/
) and rename or remove thehive-warehouse-connector-assembly-*.jar
file. Then, restart the Spark services to apply the changes. - To tell Spark to ignore Hive integration, set this option when running your job:
This ensures Spark doesn’t attempt to use Hive for metadata storage.--conf spark.sql.catalogImplementation=in-memory
- If you're planning to create new HDInsight clusters, you can use a script action during cluster deployment to remove or disable the HWC JAR file automatically.
Hope this helps! Let me know if you need further clarification or run into any issues.
Kindly consider upvoting the comment if the information provided is helpful. This can assist other community members in resolving similar issues.