Gathering Analytics from txt files from Blob container with complex file paths

HarriS 1 Reputation point
2021-01-26T01:07:52.64+00:00

I want to be able to perform analytics on several txt files that are stored in a blob container within a storage account.

The file path looks something like

shoppingcenter\location\2020-05\shop\myfile.txt

there's several location folders, several date folders, and several shop folders.

I want to be able to analyse all the txt files that start with myfile_.txt together.

What would be the best approach to produce analytics (graphs and statistics) from all the data combined?

I assume this is an azure data factory question.

Azure Data Lake Storage
Azure Data Lake Storage
An Azure service that provides an enterprise-wide hyper-scale repository for big data analytic workloads and is integrated with Azure Blob Storage.
1,466 questions
Azure Blob Storage
Azure Blob Storage
An Azure service that stores unstructured data in the cloud as blobs.
2,840 questions
Azure Data Factory
Azure Data Factory
An Azure service for ingesting, preparing, and transforming data at scale.
10,679 questions
0 comments No comments
{count} votes

1 answer

Sort by: Most helpful
  1. Ashley Mallia 1 Reputation point
    2021-01-26T06:37:15.537+00:00

    Hi Harri,

    The file connector accepts wildcards and has options like recursive and flatten, which will allow you to match files of a specific name at any point in the hierarchy.

    Refer to the docs here:
    https://learn.microsoft.com/en-us/azure/data-factory/connector-file-system

    Best Regards,
    Ashley


Your answer

Answers can be marked as Accepted Answers by the question author, which helps users to know the answer solved the author's problem.