How to bring data from a website and ingest into data lake storage and compare the data with other datasets
Hello,
We have a requirement to bring data from a publicly available source website, the data can be accessed programmatically over API, we need to bring these datasets and compare it with other data within Azure. We need to ensure data protection and encryption.
I could think of couple of options-
- We manually keep the file in a shared folder on-prem location and ADF pipeline to pick these files, write some python scripts within notebook to compare the data.
- We manually upload these files in share point site and and ADF pipeline to pick these files, write some python scripts within notebook to compare the data.
3.Use ADF pipeline Web or API connector for linked service and bring the data into data lake ,write some python scripts within notebook to compare the data.
These are assumptions hence I am looking for an expert advise and more detail level solution approach, keeping in mind we need to protect these data while in-transit and while we store the data in azure. Some sort of encryption techniques and capability to ensure data is protected through out the process.
Thanks in advance.
Regards,
Sourav
Azure Data Lake Storage
Azure Databricks
Azure Data Factory
Microsoft 365 and Office | SharePoint | For business | Windows
