Databricks readstream writestream to Azure Synapse

sakuraime 2,326 Reputation points
2020-12-10T06:08:39.147+00:00

I am having an issue on writing stream to Azure synapse with the following error .

let's have a look and see if there is idea ?

46766-3142-synapsestreamissue.jpg

Azure Synapse Analytics
Azure Synapse Analytics
An Azure analytics service that brings together data integration, enterprise data warehousing, and big data analytics. Previously known as Azure SQL Data Warehouse.
5,004 questions
Azure Databricks
Azure Databricks
An Apache Spark-based analytics platform optimized for Azure.
2,222 questions
0 comments No comments
{count} votes

1 answer

Sort by: Most helpful
  1. PRADEEPCHEEKATLA 90,231 Reputation points
    2020-12-10T10:22:44.01+00:00

    Hello @sakuraime ,

    You can resolve this issue by using the below configuration:

    spark.conf.set("spark.hadoop.mapreduce.fileoutputcommitter.cleanup.skipped","true")  
    

    46902-image.png

    For “tempDir”, we recommend you use a dedicated Blob storage container for the Azure Synapse Analytics SQL pools.

    Here is the python example of structured streaming:

    # Set up the Blob storage account access key in the notebook session conf.  
    spark.conf.set(  
      "fs.azure.account.key.<your-storage-account-name>.blob.core.windows.net",  
      "<your-storage-account-access-key>")  
      
    # Prepare streaming source; this could be Kafka or a simple rate stream.  
    df = spark.readStream \  
      .format("rate") \  
      .option("rowsPerSecond", "100000") \  
      .option("numPartitions", "16") \  
      .load()  
      
    # Apply some transformations to the data then use  
    # Structured Streaming API to continuously write the data to a table in Azure Synapse.  
      
    df.writeStream \  
      .format("com.databricks.spark.sqldw") \  
      .option("url", "jdbc:sqlserver://<the-rest-of-the-connection-string>") \  
      .option("tempDir", "wasbs://<your-container-name>@<your-storage-account-name>.blob.core.windows.net/<your-directory-name>") \  
      .option("forwardSparkAzureStorageCredentials", "true") \  
      .option("dbTable", "<your-table-name>") \  
      .option("checkpointLocation", "/tmp_checkpoint_location") \  
      .start()  
    

    The output of the Notebook:

    46885-image.png

    Reference: Azure Databricks - Azure Synapse Analytics - Usage (Streaming).

    Hope this helps. Do let us know if you any further queries.

    ------------

    • Please accept an answer if correct. Original posters help the community find answers faster by identifying the correct answer. Here is how.
    • Want a reminder to come back and check responses? Here is how to subscribe to a notification.

Your answer

Answers can be marked as Accepted Answers by the question author, which helps users to know the answer solved the author's problem.