Thank you for reaching out to Microsoft Q&A forum!
The model
parameter in the RunFunction
class specifies a model asset for the parallel job but does not automatically provide it like in BatchDeployment
. It doesn't trigger errors for non-existent models, indicating it's more for metadata. You need to manage model loading explicitly in your entry script, as there’s no built-in way to access the model during job execution.
To run parallel jobs in Azure Machine Learning, you can use the Azure CLI or Python SDK. This process involves splitting a task into mini-batches, distributing them across multiple compute nodes, and configuring inputs, data division, and compute resources.
- Setup: Ensure you have an Azure ML account, workspace, and necessary SDKs installed.
- Define Parallel Job: Create a parallel job step in your pipeline, specifying input data, instance count, mini-batch size, and error handling settings.
- Automation: Utilize optional settings for automatic error handling and resource monitoring.
- Pipeline Integration: Incorporate the parallel job as a step within your pipeline, binding inputs and outputs to coordinate with other steps.
This approach can significantly reduce execution time and improve efficiency in tasks like model training and batch inferencing.
Please look into: Use parallel jobs in pipelines.
I hope you understand. Thank you.