Azure AI foundry - can't do serverless deployments

PLAMEN YORDANOV 10 Reputation points
2025-02-20T11:06:22.6133333+00:00

Having difficulties deploying serverless Microsoft AI models (PHI-3-small-8k-instruct)

Following this tutorial https://learn.microsoft.com/en-us/azure/ai-studio/how-to/deploy-models-serverless?tabs=azure-ai-studio#deploy-the-model-to-a-serverless-api-endpoint, but all deployments are made as "Azure AI services" tagged instead of "serverless".

Is there anything I am missing ? On the AzureAI Foundry portal I just go with the deployment wizard in Models + endpoints -> Deploy model, filter by "serverless" deployment type, choose my model and it always gets deployed as "Azure AI services" tagged. Region is Sweden-Central which supports the phi-3-small-8k instruct serverless deployment.

Azure AI services
Azure AI services
A group of Azure services, SDKs, and APIs designed to make apps more intelligent, engaging, and discoverable.
3,621 questions
{count} votes

2 answers

Sort by: Most helpful
  1. PLAMEN YORDANOV 10 Reputation points
    2025-02-22T12:03:59.26+00:00

    If anybody comes across this- seems like MS has silently added a preview feature to deploy models to the Azure AI Inference service. I only found this exists when I tried to deploy a non-MS model.

    When you disable this preview feature you will see the standard deployment wizard deployment options- Managed Compute or Serverless.

    Screenshot 2025-02-21 at 17.23.41

    1 person found this answer helpful.

  2. Deleted

    This answer has been deleted due to a violation of our Code of Conduct. The answer was manually reported or identified through automated detection before action was taken. Please refer to our Code of Conduct for more information.

    1 deleted comment

    Comments have been turned off. Learn more

Your answer

Answers can be marked as Accepted Answers by the question author, which helps users to know the answer solved the author's problem.