Hi @Vikranth-AI
Thank you for using Microsoft Q&A platform and thanks for your question.
I understand that you are facing an issue with Synapse ingestion with PySpark. The error message you are seeing is due to the row size limit of 1MB in Synapse (polybase).
Use below code to get size of each row.
import sys
rows = df.collect()
for rw in rows:
print(str((sys.getsizeof(''.join(rw[0:]))))+" bytes")
This gives you size in bytes as below:
Hope this helps. Do let us know if you any further queries.
If this answers your query, do click Accept Answer
and Yes
for was this answer helpful. And, if you have any further query do let us know.