Hello Babu Musanga,
Thanks for posting your question in the Microsoft Q&A forum.
- Azure Data Factory's data flow treats empty JSON files (with just
[]
) as valid files with zero rows. This means that even though the files are technically empty, they are still processed and counted as part of the data flow execution. So remove or exclude the empty JSON files from the source folder before running the data flow pipeline. - Data flows in Azure Data Factory use partitioning to parallelize the data processing. If the partitioning strategy is not optimal for your dataset, it may lead to incomplete or inconsistent results.
- Azure Data Factory's data flow engine may apply certain optimizations or transformations that could potentially cause data loss or inconsistencies in specific scenarios.
Please don't forget to close up the thread here by upvoting and accept it as an answer if it is helpful