Azure Data Factory

Imran Mondal 246 Reputation points
2021-03-13T20:04:50.507+00:00

EveryDay My Data Bricks job dumping data in YYYY/MM/DD/HH format folder in Blob Storage
like this -

hourlydata/2021/03/10/00/part-00000-tid-8b61-4d85-a6cb-249245ccad76-39980-1-c000.csv
hourlydata/2021/03/10/01/part-00000-tid-9c81-dc555aaa0f8-41519-1-8c3-46ca-97-c000.csv
hourlydata/2021/03/10/02/part-00000-tid-9c81-dc555aaa0f8-41519-1-8c3-46ca-97-c000.csv

77445-image.png

I have a data set that resides under a folder path where the date is dynamic, inside each hour folder along with CSV files, we have other files also. But I want to select only the CSV file.

As soon as my Databricks Jobs dumps data every hour, I would like to read those new data using data factory and load it into a database, please suggest

Azure Data Factory
Azure Data Factory
An Azure service for ingesting, preparing, and transforming data at scale.
11,624 questions
{count} votes

1 answer

Sort by: Most helpful
  1. Vaibhav Chaudhari 38,916 Reputation points Volunteer Moderator
    2021-03-15T06:45:35.013+00:00

    Try using Event based triggers in ADF. You should configure it such way that as soon as new file is created ending with .csv, pipeline will use this file to copy/process

    See point 6 in below doc

    https://learn.microsoft.com/en-us/azure/data-factory/how-to-create-event-trigger

    ----------

    Please don't forget to Accept Answer and Up-vote if the response helped -- Vaibhav


Your answer

Answers can be marked as Accepted Answers by the question author, which helps users to know the answer solved the author's problem.