Best practices for submitting spark batch jobs in Azure HDinsights.

Bipin Singh 21 Reputation points
2021-07-19T08:22:48.617+00:00

Hi, I'm looking to submit my pyspark scripts in HDInsight.
Currently, HDInsight provided Livy for job submission, using curl.
However, If I want to productionize it, then what authentication mechanism to use.
Also, How can I check the progress of my spark application and create an alert on various instances.

Thanks

Azure HDInsight
Azure HDInsight
An Azure managed cluster service for open-source analytics.
210 questions
0 comments No comments
{count} votes

Accepted answer
  1. MartinJaffer-MSFT 26,086 Reputation points
    2021-07-21T06:54:46.957+00:00

    Hello @Bipin Singh and welcome to Microsoft Q&A.

    Under 'Best Practicies' documentation, there are a number of options for submitting, and monitoring jobs.

    These options include making an app in VScode, an application calling REST API for livy, as part of a Data Factory pipeline, and others.
    Curl is a client application. You can write a script to automate using curl calls, or, make a full application to do the same thing using some other url library. In the end, everything is at some point an http request.

    It all really depends upon what tools you want to use, and what sort of application you are building.

    Or you can go direct to the livy docs.

    Does this help?

    1 person found this answer helpful.

0 additional answers

Sort by: Most helpful

Your answer

Answers can be marked as Accepted Answers by the question author, which helps users to know the answer solved the author's problem.