Error while reading file abfss:REDACTED_LOCAL_PART@miglandingzonelake.dfs.core.windows.net - source folder deleted

GB 6 Reputation points
2021-02-12T12:25:37.48+00:00

Hello, I am getting this intermittent error when running an ADF pipeline. After receiving this error the storage account blob container folder is emptied of all files! I presume this what redacted means? I have deleted the DataSet and recreated with the same name and the error returned after working for a while? Recreating with a new name seems to have avoided the error thus far? Is there a way to monitor the cache mentioned? To have an error that deletes source files is a little scary to say the least.

Hopefully someone can help.

Operation on target <data flow name> failed: {"StatusCode":"DFExecutorUserError","Message":"Job failed due to reason: Error while reading file abfss:REDACTED_LOCAL_PART@miglandingzonelake.dfs.core.windows.net/<path/filename.csv>. It is possible the underlying files have been updated. You can explicitly invalidate the cache in Spark by running 'REFRESH TABLE tableName' command in SQL or by recreating the Dataset/DataFrame involved."

The DataSet used is as follows. It is used a source and sink in the data flow.
{
"name": "<ldz_name>",
"properties": {
"linkedServiceName": {
"referenceName": "<LandingZoneDataLake>",
"type": "LinkedServiceReference"
},
"folder": {
"name": "ldz_name"
},
"annotations": [],
"type": "DelimitedText",
"typeProperties": {
"location": {
"type": "AzureBlobFSLocation",
"fileSystem": "ldz_name"
},
"columnDelimiter": ",",
"escapeChar": "\",
"firstRowAsHeader": true,
"quoteChar": "\""
},
"schema": []
}
}

<Aside - I have replaced actual folder and filenames above>

Azure Data Factory
Azure Data Factory
An Azure service for ingesting, preparing, and transforming data at scale.
10,127 questions
{count} votes