best practice
Hi, What is the best practice to execute the appropriate pipeline(s) with parameters based on the file(s) being landed in container A? Thank you
Custom Event Trigger for multiple synapse pipelines
Hi everyone, We have a process in place which sends and event to an event grid once a staging table is loaded. The payload of the event includes the table name. The next step is to develop a trigger which executes the correspondent pipeline to move…
ADF Pipeline concurrency for different pipelines
Hello Team, I am having two different pipelines. Both the pipelines are configured to process data and insert in to a same table. So when these pipelines run parallelly, sometimes the first pipeline already inserts the data and the second pipeline has…
Latest file extraction by scanning multiple folders and copy to azure sql database
Hi, I have a folder structure in ADLS Gen2 in container source that goes as below: /table1/2021/01/20/file1.parquet /table1/2021/02/20/file2.parquet /table1/2022/01/10/file3.parquet /table1/2022/01/12/file4.parquet …
Read XML content as a parametar
HI All , I have one XML file .... I need to read the file and wants to pass the total data into one variable as a single column .....is it possible to do that in ADF ? Thanks in Advance
Unable to launch the ADFv2
I have tried launching adfv2 both in safari and chrome(normal/incognito) browsers but nothing worked for me. Also I verified cookies,it’s fine and deleted browsing data. After these many attempts still issue persists. Kindly suggest . I have tried…
Write to cosmos DB using ADF
Hi Experts, I need to create a pipeline with source as "Azure SQL DB" and the target / sink being cosmos db. There might be few nested objects. Can someone provide some links with sample pipelines for the same? Thanks
Capture Pipeline error messages
Hello, My requirement is to capture the error message of pipeline execution in Data factory and store it into a table. Currently I'm able to do it activity wise only ('@{activity('Copy data').output.errors[0].Message}'), which captures the message for…
Is Data factory really needs cluster configuration for running? Can we run datafactory without any cluster configuration ?
Hi team, I am doing PAAS based project using data factory. We need push values from excel to mysql. For this do we really need cluster configuration or it will run with in azure without any cost apart from Data factory execution cost. Help and…
Release pipelines
Hello, In the development environment, I create sample pipelines to test, etc. When I decide to deploy to production using the CI/CD release pipelines, what happens to the sample pipelines in dev? will they get migrated to production as well? if so,…
ELT frameworks (Azure)
Generally speaking, which Azure data platform is most typically utilized for data processing in an ELT framework? Azure Data Factory or Azure Databricks...
ADF to schedule databricks streaming job
Hello, We are using Azure Data Factory (ADF) to schedule databricks streaming job. Requirement: The streaming job may need to run all the time. If it is cancelled or failed, then we need to automatically restart it. Problem: Currently we…
Select cluster option is changing after deplymento to stage and production in ADF
In Dev we have created linked service for databricks with choosing "existing interactive cluster" option . But after deployment to stage and production the configuration is changing to "New job cluster". What could be the issue? …
ADF: Unable to load json file >100 MB into a database column
Hi, I have a requirement to load the json files (ADF Logs) into our database via Azure Data Factory (Copy Activity). The datatype of the column is "[nvarchar]". There will be around 20 files/day and the file at 21:00 hrs is having size more…
Querying a data source - Azure Data Factory
Hi, I'm implementing some ADF pipelines. I need to query the data source (a relational database) in order to understand the right data filters before the data ingestion. Is it possible (f.e. at linked service and/or data set)? Thanks
Mapping parquet format datasets to Azure Datawarehouse
Hi, I have a pipeline that copies all the files (*.parquet) present in different folders into different tables using foreach loop, Now I have particular set of columns defined for each destination tables, for example Table A contains 100 columns, while…
How to connect to multiple servers to run the same query dynamically in Azure data factory or in Azure
I have a requirement to connect to multiple servers to run the same complex sql query and should combine the results of all those outputs as a single result set and should pass that to the next component for further processing. Could you please let me…
Downloading multiple file types from dynamic url in ADF
I'm fetching a JSON object, { id:"1123123", "download_url":"https://file-examples-com.github.io/uploads/2017/02/file_example_XLS_10.xls", "status":1} { id:"1123123",…
Dynamically retrieving name of storage account in Azure Data Factory
I have multiple pipelines in different environments (dev, test, stage etc...). Is there a way that I can dynamically retrieve the name of the storage account (the name tells you the environment it's running in) in azure data factory so that I can make…
How to accomplish pagination with "rel = next"
For a REST api endpoint, the client data feed, gives us the pagination value in the header. Please can you advise on how to accomplish this. The relative URL for the dataset is https://clientname.com/api/v2/tickets?updated_since=2019-01-19 When the above…