Azure Data Factory still generating deleted log even with "After completition" set to "No Action"

Azevedo Emanoel 1 Reputation point
2024-05-19T13:19:22.4433333+00:00

Hi,

I’m facing the issue below.

I have a Data Flow that inserts a filter into an Azure SQL DB. I’ve set it to “Delete source files” after completion, but it’s still generating a log in the Blob storage, which I don’t want.

So, I changed it to “No action again,” but it’s still generating the log in the Blob Storage.

The Data flow:

User's image

The log, the one that I don't want to be generated anymore:

User's image

I have already created a new data flow, changed the name, and so on, but nothing is resolving the issue.

Azure Data Factory
Azure Data Factory
An Azure service for ingesting, preparing, and transforming data at scale.
9,805 questions
{count} votes

1 answer

Sort by: Most helpful
  1. Vlad Costa 90 Reputation points
    2024-05-19T23:14:04.8066667+00:00

    This is a known behaviour in Azure Data Factory (ADF) where a branch exists whether or not any rows are fed to it, so the Sink will still run. In the case of storage sinks, a 0-byte blob will be created.

    Here are a couple of suggestions that might help:

    • Use Cache Sink: If you don’t wish to write output to an external source, you can use a cache sink. It writes data into the Spark cache instead of a data store. In mapping data flows, you can reference this data within the same flow many times using a cache lookup.
    • Delete Empty Files Programmatically: If you still want to delete empty zero-byte files, you can use ADF or a programmatic method to delete them at the end of execution.

    References:

    https://stackoverflow.com/questions/68404511/zero-bytes-files-are-getting-created-by-adf-data-flow

    Also, if my response helped resolve your issue, could you please mark this answer as accepted? Thank you!

    0 comments No comments