How to monitor REST API "run submit" job on Azure Databricks?

philmarius-new 126 Reputation points
2021-03-29T10:48:36.833+00:00

Reposted here for visibility.

We're moving away from notebooks and putting our pyspark workflows into .py files that are being uploaded to DBFS via CI/CD pipelines and then being run by the run submit API endpoint. However, we're struggling to monitor these jobs for failures.

We've setup the spark-monitoring scripts on our Azure Databricks instance and it's successfully feeding logs back to Azure Monitor. However, when running deliberately failing jobs, we cannot tell via the logs whether the job has failed or not, making monitoring the ETL workflows we have problematic.

Has anyone done something similar to this before and how did they do it?

Azure Databricks
Azure Databricks
An Apache Spark-based analytics platform optimized for Azure.
1,916 questions
{count} votes