I have a Data Factory pipeline that currently copies files daily from a Google Storage account down to an Azure Storage Blob ADL Gen2 enabled.
Source several different files, File1, File2, File3 etc, all have a data range in the file name File1_20200101_20200102.csv.gzip and they are .csv and zipped.
I was able to connect and use a Binary source and Binary target and just grab all files that were created/modified yesterday. Also part of the target, I unzip the files so they are just .csv.
I want to make sure that I'm setting up the structures correctly in the blob storage for it to function as a DL.
BlobContainer1/RAW/GoogleSource/File1_20200101_20200102.csv.gzip
From what I'm reading, I should probably have BlobContainer1/RAW/GoogleSource/File1/{year}/{month}/{day}/File1_20200101_20200102.csv.gzip, would that be correct?
If so, is it possible to dynamically determine the folder path based on each file name that is being pulled in, OR, do I have to create a separate copy pipeline for each File that is being copied over?
[Note: As we migrate from MSDN, this question has been posted by an Azure Cloud Engineer as a frequently asked question]
MSDN Source: Copy Multiple files into ADL Gen2