How to bring data from a website and ingest into data lake storage and compare the data with other datasets

Sourav 130 Reputation points
2023-10-29T12:19:15.2233333+00:00

Hello,

We have a requirement to bring data from a publicly available source website, the data can be accessed programmatically over API, we need to bring these datasets and compare it with other data within Azure. We need to ensure data protection and encryption.

I could think of couple of options-

  1. We manually keep the file in a shared folder on-prem location and ADF pipeline to pick these files, write some python scripts within notebook to compare the data.
  2. We manually upload these files in share point site and and ADF pipeline to pick these files, write some python scripts within notebook to compare the data.

3.Use ADF pipeline Web or API connector for linked service and bring the data into data lake ,write some python scripts within notebook to compare the data.

These are assumptions hence I am looking for an expert advise and more detail level solution approach, keeping in mind we need to protect these data while in-transit and while we store the data in azure. Some sort of encryption techniques and capability to ensure data is protected through out the process.

Thanks in advance.

Regards,

Sourav

Azure Data Lake Storage
Azure Data Lake Storage
An Azure service that provides an enterprise-wide hyper-scale repository for big data analytic workloads and is integrated with Azure Blob Storage.
1,562 questions
Azure Databricks
Azure Databricks
An Apache Spark-based analytics platform optimized for Azure.
2,534 questions
Azure Data Factory
Azure Data Factory
An Azure service for ingesting, preparing, and transforming data at scale.
11,646 questions
Microsoft 365 and Office | SharePoint | For business | Windows
{count} votes

Your answer

Answers can be marked as Accepted Answers by the question author, which helps users to know the answer solved the author's problem.