Optimize model with Intel OneAPI on Azure ML

Karnik Kanojia 0 Reputation points
2024-07-14T14:51:14.0633333+00:00

I have already successfully deployed my deep learning model as an online endpoint using the following docs. Now I'm seeking to optimize the inference timings. The VM size is following Standard_DS3_v2 which belongs to DSv2 series. My questions are the following:

  1. Is there any Azure ML curated environment that has the Intel OneAPI build and I can stack my docker context upon it to optimize things?
  2. Are there other ways to optimize the things too?
Azure Machine Learning
0 comments No comments
{count} votes

1 answer

Sort by: Most helpful
  1. Azar 31,600 Reputation points MVP Volunteer Moderator
    2024-07-14T18:58:11.1133333+00:00

    Hi there Karnik Kanojia

    Thanks for using QandA platform

    For optimizing, you can create a custom Docker image with Intel OneAPI and use it as your base. Start with a Dockerfile using the intel/oneapi-basekit image, build and push it to Azure Container Registry, and configure your Azure ML endpoint to use this custom image.

    also, try model quantization, pruning, using more powerful VM instances like the NC series, optimizing your inference pipeline,

    If this helps kindly accept the response thanks much,


Your answer

Answers can be marked as 'Accepted' by the question author and 'Recommended' by moderators, which helps users know the answer solved the author's problem.