Hi ,
I am copying large data from ADLS gen 1 to ASDW(Azure synapse Analytics) using Data Factory (polybase set to yes and mentioned blob storage settings).
The Source data is *.parquet format and is partitioned. Copy activity fails. In first attempt I gave the wildcard path of all the parquet files. It threw this error: Operation failed as split count exceeding upper bound of 1000000
In the second attempt, I reduced the copy size to just one partition, I mentioned just one partition folder in ADLS - but it gave another error: " file is not a parquet file (too small)" . Please help me to resolve this. It looks like configuration issue but where should this be done? Thanks
"dataRead": 25884501402,
"filesRead": 1674684,
"sourcePeakConnections": 137,
"copyDuration": 5796,
"throughput": 4361.255,
"sqlDwPolyBase": true,
"errors": [
{
"Code": 11404,
"Message": "ErrorCode=FailedDbOperation,'Type=Microsoft.DataTransfer.Common.Shared.HybridDeliveryException,Message=Error happened when loading data into SQL Data Warehouse.
,Source=Microsoft.DataTransfer.ClientLibrary,''Type=System.Data.SqlClient.SqlException,Message=110825;
Operation failed as split count exceeding upper bound of 1000000.,
Source=.Net SqlClient Data Provider,SqlErrorNumber=110825,Class=16,ErrorCode=-2146232060,State=1,
Errors=[{Class=16,Number=110825,State=1,Message=110825;Operation failed as split count exceeding upper bound of 1000000.,},],'",
"EventType": 0,
"Category": 5,
"Data": {},
"MsgId": null,
"ExceptionType": null,
"Source": null,
"StackTrace": null,
"InnerEventInfos": []