Copy Dataverse data into Azure SQL using Synapse Link
Im using Azure Synapse pipeline to copy data from Dataverse to Azure SQL DB. For large tables, it is only copying 10 million records and it is not inserting after. I have about 15 mil records. I dont see any settings limitation anywhere on the pipeline.…
Azure Data Factory CopyData from DB on-premise to Azure using SHIR
Hi, I had set up Self Hosted Integration Runtime on my on-premise VM with azure cloud using Azure Data Factory. The connection is already been tested and checked green on both integration runtime and linked services. My goal is to be able to copy data…
Copy activity failed because you have multiple concurrent copy activities runs writing to the same file
Hi All, I am migrating 1000 of SQL tables from on premise SQL server to Azure Blob storage. I am using ADF for each and copy activity to do so. However, while processing the tables in parallel/concurrent, I am getting below error - Failure happened…
PATCH method not allowed for Storage Services REST API "Path - Update"
I try to use the Set Access Control option described in https://learn.microsoft.com/en-us/rest/api/storageservices/data-lake-storage-gen2?view=rest-storageservices-datalakestoragegen2-2019-12-12. Although I was able to run other methods (as Path - List)…
Why Data asset is not supported when try to create a AutoML job?
Hi, I successuffly created a data asset (folder_URI type) with uploaded imges in an Azure Blob storage (registered storage source), but when I try to create a job in Azure AutoML the Data asset shows (not supported). Any idea what is the issue? Thanks
Is there a way to use query acceleration on .avro files
I have a data lake storage resource in which I am storing .avro files. I now have a need to transfer the data in those files to a database. I've managed to make that work. But I would love to optimize it by only downloading data I need. I read this…
About right naming convention on paths and blobs
Hi, I'm used to design ADLS partition in the following format: source={source}/table={table}/date={date}/content.csv When a table in any source doesn't have data in certain date, that partition won't exists. In the other hand, not every table will…
Azure Datalake query accelaration error XML specified is not syntactically valid.\nRequestId:e3204a59-901e-005f-7612-d2ee5f000000\nTime:2021-11-05T06:58:32.8652708Z\r\nStatus: 400 (XML specified is not syntactically valid.
I am using query acceleration to access and filter data lake- https://learn.microsoft.com/en-us/azure/storage/blobs/data-lake-storage-query-acceleration-how-to?tabs=dotnet%2Cazure-cli I followed the same steps as per the document to the letter. but…
Power BI JSON file 1000 row limit
Hi all, I am facing issues with Power BI only pulling in 1000 rows. I have exported JSON data into azure data lake storage gen 2, which then our business analyst was able to pull in the files and create separate tables in a pbix file. The issue he has…
Why is the activity log stored in the adls gen2 container not a valid json
Why is the activity log stored in the adls gen2 container not a valid json? It's line by line, and the outer layer is missing '[]' and comma separation
partitioning of Data lake
I have just the below link on data partitioning of Azure data lake : https://learn.microsoft.com/en-us/azure/architecture/best-practices/data-partitioning But it is too brief and does not cover this important topic extensively. I request you to please…
DataLakeServiceClient.DeleteFileSystem() method on non-ADLSv2 upgraded storage accounts
The .NET codebase I work on interacts with some storage accounts that are upgraded to ADLSv2 and some that are not. I noticed that the DataLakeServiceClient.DeleteFileSystem() method works in both cases. Can I get a confirmation that the method's…
Azure datalake and data consistency
Hi Experts, Azure Data Lake Storage (ADLS) does not natively provide full ACID (Atomicity, Consistency, Isolation, Durability) transaction support unlike traditional relational databases designed to support ACID transactions. This raises the following…
Best way to grant SQL DB (serverless sql pool) access to end user.
What is the best way to grant access to users for SQL DB(serverless sql pool) build on top of synapse azure lake db. How should SQL DB (views or external tables) access the lake database (storage container). Should we use RBAC, Managed Identity, ACLS,…
Data lake performance and improvement
The data lake performance depends on critical factors such as partitioning on date and if available in data region etc. I have already partitioned the data lake in a hierarchal structure LandingZone Subject area YYYY MM …
can abac be applied to a data lake?
Can abac be applied to a data lake storage account or does it have to be rbac with potential ACL?
Azure data lake folder structure
Hi, the link provided in the below thread is not working, any way to find the information or a new URL/link: FAQs About Organizing a Data…
Copy Activity very slow for Azure data factory
Hi All, we are having issue where copy activity from Azure Table storage to data factory is taking over 50mins on daily basis. it used to take 15-20mins until a week back. What's the source of the slow copy activity? Azure Table storage What's the sink…
Filepath() ,metadata functions in serverless view on delta format
Hi , Earlier i was creating serverless views with FORMAT='PARQUET' using openrowset wildcard paths ("/synapsetesing/test/*.parquet"), so i was able to use metadata functions like filepath etc , now when i create views on FORMAT='DELTA' ,…
ADF pipeline to read the data from UC table to adls gen2 account
Hello Team, We have a requirement to create Azure Datafactory pipeline to read the data from UC table, access on the table is provided ( to Azure Datafactory Managed Identity) and copy the data into adls gen2. Is there a way or article to implement this?…