@Rakesh Kumar - Thanks for the question and using MS Q&A platform.
This is an expected behaviour when run any spark job to create these files.
When DBIO transactional commit is enabled, metadata files starting with started and committed will accompany data files created by Apache Spark jobs. Generally you shouldn’t alter these files directly. Rather, use the VACUUM command to clean the files.
A combination of below three properties will help to disable writing all the transactional files which start with "_".
We can disable the transaction logs of spark parquet write using
spark.sql.sources.commitProtocolClass = org.apache.spark.sql.execution.datasources.SQLHadoopMapReduceCommitProtocol
This will help to disable the committed and started files but still _SUCCESS, _common_metadata and _metadata files will generate.
We can disable the _common_metadata and _metadata files using
parquet.enable.summary-metadata=false
We can also disable the _SUCCESS file using
mapreduce.fileoutputcommitter.marksuccessfuljobs=false
For more details, refer "Transactional Writes to Cloud Storage with DBIO" and "Stop Azure Databricks auto creating files" and "How do I prevent _success and _committed files in my write output?".
Hope this helps. Do let us know if you any further queries.
If this answers your query, do click Accept Answer
and Yes
for was this answer helpful. And, if you have any further query do let us know.