When trying to use a Data Factory pipeline to copy from Azure Databricks Delta Lake to Azure Blob Storage (DelimitedText), I'm getting the following error when using the copy data activity:
ErrorCode=AzureDatabricksCommandError,Hit an error when running the command in Azure Databricks. Error details: shaded.databricks.org.apache.hadoop.fs.azure.AzureException: shaded.databricks.org.apache.hadoop.fs.azure.AzureException: Unable to access container <my-container> in account <my-storage-account>.blob.core.windows.net using anonymous credentials, and no credentials found for them in the configuration.
Caused by: shaded.databricks.org.apache.hadoop.fs.azure.AzureException: Unable to access container <my-container> in account <my-storage-account>.blob.core.windows.net using anonymous credentials, and no credentials found for them in the configuration.
Caused by: hadoop_azure_shaded.com.microsoft.azure.storage.StorageException:
Caused by: java.net.UnknownHostException: <my-storage-account>.blob.core.windows.net.
I can switch the source dataset to Azure SQL Database and it runs without issue. I can also read and write to the same blob storage container without issue in the same pipeline. The credentials are saved with the underlying Azure Blob Storage linked service and the connection validates just fine. So the issue can't be with the blob storage container.
Inputs:
{
"source": {
"type": "AzureDatabricksDeltaLakeSource",
"exportSettings": {
"type": "AzureDatabricksDeltaLakeExportCommand"
}
},
"sink": {
"type": "DelimitedTextSink",
"storeSettings": {
"type": "AzureBlobStorageWriteSettings"
},
"formatSettings": {
"type": "DelimitedTextWriteSettings",
"quoteAllText": true,
"fileExtension": ".csv"
}
},
"enableStaging": false
}
Seems like either the above error message isn't accurate, or there's a bug somewhere.
Has anyone else run into this?