Efficient Log Handling and Data Retention in Azure Data Factory and Databricks
I need to create a solution to send logs from Azure Data Factory to the Databricks Unity Catalog. I'm considering the following structure: Whenever an activity run results in either failure or success, the corresponding log will be sent to Azure Logic…
How to connect (create a linked service) an ADF instance to an SFTP Server that is hosted on an Azure VM?
I am trying to get my ADF instance to connect to my SFTP Server that is hosted on an Azure VM but when testing the connection it always Times Out. The ADF and VM hosting the SFTP are on different VNets and this is because I personally don't have…
How to have back of everything for Azure Active Directory?
I want to have backup of everything from Azure Active Directory. How to do it whats the procedure please help me guide it. If 3rd software also can be used Please help. I use Azure active directory and Intune.
![](https://techprofile.blob.core.windows.net/images/nangnqqNAkuWTT7031X8vg.png?8DCA0F)
Data Factory Logs to Databricks
I need to create a way to send logs from Data Factory to the Databricks Catalog. What is the most cost-effective and efficient method to achieve this?
Captures logs from Azure Data Factory and inserts them into a Delta Table in Databricks
Good morning, I need assistance in creating a project that captures logs from Azure Data Factory and inserts them into a Delta Table in Databricks. The key requirements for this project are as follows: No Duplicate Logs: Ensuring that the logs are not…
How to Correctly Pass and Use Boolean Values from Datafactory to Databricks Notebook
How can I correctly pass a Boolean value from Datafactory to a Databricks notebook and use it in conditional logic? I configured a pipeline in Datafactory that calls a Databricks notebook. I attempted to pass a Boolean parameter from Datafactory as a…
Data Factory Logs --> Catolog Databricks
Good morning, I need assistance in creating a project that captures logs from Azure Data Factory and inserts them into a Delta Table in Databricks. The key requirements for this project are as follows: No Duplicate Logs: Ensuring that the logs are not…
ADF Data Flow - CDC - Schema Drift not working
Hi, I have setup ADF pipeline for data sync of tables between Azure SQL Database to SQL Managed Instance databases. I am using Data Flow and CDC to track changes and sink activity to replicate it to target. Since there are multiple tables involved, I am…
How do you access the Airflow CLI/API/DB in Workflow Orchestration Manager?
Working with Airflow, one often requires access to the Airflow CLI/API or even access to the underlying meta database. For example: to import or export connections start backfills manage Variables (using scripts) manage DB items that are…
how to fetch data from Azure Active Directory(AD) by using either ADF or databricks
To fetch data from Azure Active Directory (AD) using either Azure Data Factory (ADF) or Azure Databricks, Pleae let me know in detail. thanks
How to flatten nested json array in ADF where the first node does not have key and starts with values which is not constant
I am trying to parse nested JSON in my pipeline to store the values such as table_name, lastSuccessfulWriteTimestamp, totalProcessedRecordsCount, dataFilesPath, schemaHistory_value1, schemaHistory_value2 from below JSON data using ADF dataflow activity.…
I wan use Show billing report by pipeline in ADF, and enabled this via CI/CD
I want enabled the feature "Show billing report by pipeline " in azure Data Factory, but I don't know how can I enabled this via Terraform deployment or ARM template deployment. How Can I do that ?
"Extra"."Api_Issue" row limit from Data Factory
Im working with the API from Data Factory and whas doing this query: select * from "Extra"."Api_Issue" Previously it was retrieving like 60K rows but from 24th of June it is only getting EXACTLY 12.400 each day from this table. Has it…
ADF Global Parameters validation failed using NPM Run validate command
Below CICD YAML deployment code is getting failed "customCommand: 'run build validate $(Build.Repository.LocalPath) /subscriptions........" is getting failed due to usage of global parameters in various pipelines."with below…
How to Handle Case Sensitivity in Dynamic JSON Input Keys ("A" vs "a") for Derived Columns in ADF Dataflow Pipeline?
I am working on a dataflow pipeline in ADF with an input dataset schema that includes a column 'A'. The dataflow processes various JSON files with dynamic schemas, where some JSON files use 'A' and others use 'a' as keys. Since the schema is case…
SAS token generation by Databricks to access CSV files from ADLS container folder
Hi Team, There are some csv files zips inside the ADLS container folder. These zip files need to be downloaded for data correction. Downloading the file requires SAS token embedded with zip file path. Databricks has been used to generate the token and…
![](https://techprofile.blob.core.windows.net/images/hAFzqf2_AwAAAAAAAAAAAA.png?8DB62B)
Upsert Data in SQL Server Using Synapse Notebook and Pyspark
How do I use pyspark in a synapse notebook to upsert data in SQL Server? I am able to read the table with the following code: df = spark.read.jdbc(url=jdbc_url, table="Dim.table_name", properties=properties) But I am not sure how to upsert…
How to use/pass comparison query operators (like Greater Than $gt/Less Than or Equal To $lte, etc) in Filter condition in Source Dataset of Azure COSMOS for Mongo DB in the COPY activity?
Hello Team, I am sourcing data from Azure COSMOS DB for Mongo DB and load the same to Azure SQL Server DB. The data has to be read/loaded incrementally from source in every run based on a Date column. Could you please help me on how I can pass/use the…
Convert UTC timezone to US
Hi All, I need to convert the timezone from UTC to the US timezone. Can anyone please tell me what function I need to use: convertTimezone or convertfromutc? This is because my sharepoint is in the US time…
![](https://techprofile.blob.core.windows.net/images/c4mRkDtLdU6n6fSv2i9uBg.png?8D843C)
Best Practices for Automating Pipeline Execution Data Collection in Azure Data Factory
Hello everyone, I am looking for the best practices to create an automated workflow for collecting execution data from Azure Data Factory (ADF) pipelines, storing this data in Azure Data Lake Storage Gen2, and consolidating it into a single table for…