Share via

Facing issue while calling mssparkutils.notebook.runMuktiple([notebook_list along with parameters]) in azure synapse analytics

karthik raja 0 Reputation points
2024-10-10T06:48:34.6166667+00:00

Here I try to call few notebooks in another common notebook. Where magic commands is not working to have control over the notebooks when it comes execute the required notebook. So I tried write through mssparkutils.notebook.runMultiple(), in this scenario I want those notebooks to run parallel without waiting for other notebooks to complete. But it's failing with the error msg which I added in the picture

Azure Synapse Analytics
Azure Synapse Analytics

An Azure analytics service that brings together data integration, enterprise data warehousing, and big data analytics. Previously known as Azure SQL Data Warehouse.


1 answer

Sort by: Most helpful
  1. PRADEEPCHEEKATLA 91,861 Reputation points
    2024-10-10T09:20:13.4433333+00:00

    @karthik raja - Thanks for the question and using MS Q&A platform.

    It seems like you are facing an error while running a notebook in Azure Synapse Analytics. The error message "py4JJavaError saying failed with status code: 400, response: system level submitter mismatch. Kernel" indicates that there is a mismatch between the submitter of the job and the system level submitter. This error can occur when the user who submitted the job is different from the user who is running the job.

    To resolve this issue, you can try the following steps:

    • Transient issues can sometimes occur in Azure Synapse Analytics, and retrying the operation can often resolve the issue.
    • You can try creating a new notebook or a new Spark SQL pool and see if the issue persists.
    • If the issue continues to occur, you may want to check if there are any service outages or maintenance activities that could be causing the issue. Please do share the Synapse Spark runtime verision and the region of the synapse workspace.

    I tried to run the mssparkutils.notebook.help("runMultiple") from our end and able to execute without any issues.

    The method mssparkutils.notebook.runMultiple() allows you to run multiple notebooks in parallel or with a predefined topological structure. The API is using a multi-thread implementation mechanism within a spark session, which means the compute resources are shared by the reference notebook runs.

    User's image

    Here is the status view of notebook run: Notebook1

    User's image

    Here is the status view of notebook run: Notebook2

    User's image

    In the above example both the notebooks named(Notebook1 and Notebook2) ran using the same Apache spark application named Livy ID 12

    In case, if you experiencing the same issue - I would suggest you to share the Synapse Spark runtime verision and the region of the synapse workspace along with the screenshot of the error message to investigate further.

    Hope this helps. Do let us know if you have any further queries.


    If this answers your query, do click Accept Answer and Yes for was this answer helpful. And, if you have any further query do let us know.

    Was this answer helpful?


Your answer

Answers can be marked as 'Accepted' by the question author and 'Recommended' by moderators, which helps users know the answer solved the author's problem.