Share via

Enabling Chat Completions API for GPT-5.1-mini deployment (currently only Responses API works)

Rokas Stankevicius 0 Reputation points
2026-05-13T06:51:39.88+00:00

Hi Microsoft Support,

We have a GPT-5.1-mini deployment running on Azure OpenAI and need help enabling the Chat Completions API for it.

Current situation:

  • Our deployment works fine with the Responses API (/openai/responses)
  • When we send requests to the Chat Completions endpoint (/openai/deployments/{deployment-name}/chat/completions), we get a 404 Not Found error
  • It looks like Azure only exposes the Responses API by default for this model

We are getting intermittent 404 "Resource not found" errors when calling our GPT-5.1-mini deployment via the Chat Completions API. The behaviour is inconsistent and we need help identifying the root cause.

Problem summary

The same deployment works from some environments and fails from others with a 404 error. In some cases it works for a period of time, then stops working without any change on our side.

What we have verified

  • The API key is correct (same key works in environments that succeed).
  • The resource name and endpoint URL are correct (copied directly from Azure AI Foundry > Deployment > Target URI).
  • The deployment name in our request matches the deployment shown as active in the Azure portal.
  • The deployment status in the portal shows as Succeeded / active.
  • No configuration changes were made between the working and failing states.

Important context

We built our Chat Completions API integration based on working examples from previous deployments (GPT-4.1-mini and similar). The same URL format, headers, and request structure that worked reliably for GPT-4.1-mini now fail intermittently on GPT-5.1-mini, even though the only thing we changed was the deployment name and model.

What we need:

  1. Can GPT-5.1-mini be used with the Chat Completions API on Azure? If yes, how do we enable it on our existing deployment?
  2. If it requires a new deployment type or API version, please share the exact configuration (API version, endpoint path, deployment settings).
  3. If Chat Completions is not supported for this model, please confirm so we can plan around it.
Azure OpenAI in Foundry Models

1 answer

Sort by: Most helpful
  1. Q&A Assist
    2026-05-13T06:52:02.39+00:00
    AI generated content. This question contains an answer created with AI from Q&A Assist. Learn more

    GPT-5.1-mini can be used with the Chat Completions API in Azure OpenAI in Foundry, but it must be called through the correct endpoint shape and API version for chat completions.

    Key points based on the available information:

    1. Supported Chat Completions endpoint shape

    For chat completions in the Foundry Models REST API preview, the documented endpoint is:

    POST https://{endpoint}/openai/deployments/{deployment-id}/chat/completions?api-version=2025-04-01-preview
    

    Required elements:

    • endpoint: https://{your-resource-name}.openai.azure.com
    • deployment-id: the name of the GPT-5.1-mini deployment
    • api-version: 2025-04-01-preview (for the preview chat completions API)

    Headers (one of the following auth methods):

    • Authorization: Bearer {Azure_OpenAI_Auth_Token} with a token obtained for https://cognitiveservices.azure.com
    • or api-key: {your-azure-openai-key}

    If the current calls are using a different api-version, a different base URL, or a path that does not match the above, a 404 Not Found is expected. Ensure:

    • The base URL is exactly https://{your-resource-name}.openai.azure.com (no /openai/v1/ suffix for this preview REST shape).
    • The path is /openai/deployments/{deployment-id}/chat/completions.
    • The query string includes ?api-version=2025-04-01-preview.
    1. Relationship to Responses API

    The Responses API and Chat Completions API are two different surfaces over the same models:

    • Responses API: typically called via /openai/responses (often through the Foundry SDKs like AIProjectClient and get_openai_client().responses.create(...)).
    • Chat Completions API: called via /openai/deployments/{deployment-id}/chat/completions with the preview API version.

    The same deployment (for example, a GPT-5.1-mini deployment) can be addressed via:

    • responses.create(model=<deployment-name>, ...) using the Foundry Responses client, or
    • POST /openai/deployments/{deployment-name}/chat/completions?api-version=2025-04-01-preview using the Chat Completions REST API.

    No special “deployment type” toggle is documented for enabling chat completions separately from responses; the critical requirement is using the correct endpoint and API version.

    1. Checklist to resolve the 404

    Validate the following against the current implementation:

    1. Endpoint
      • Must be: https://{your-resource-name}.openai.azure.com
      • Do not append /openai/v1/ when using the Foundry Models REST preview shape shown above.
    2. Path
      • Must be: /openai/deployments/{deployment-name}/chat/completions
      • {deployment-name} must exactly match the deployment name configured in Azure OpenAI / Foundry.
    3. API version
      • Use: api-version=2025-04-01-preview for chat completions in the Foundry Models REST API preview.
    4. Authentication
      • Use either:
        • Authorization: Bearer {token} where the token is obtained via:
                 az account get-access-token --resource https://cognitiveservices.azure.com
          
          and the scope https://cognitiveservices.azure.com/.default, or
        • api-key: {your-azure-openai-api-key}.

    If any of these elements are incorrect (for example, using an older api-version, a different base URL, or a mismatched deployment name), the service can return 404 Not Found even though the deployment works via the Responses API.

    1. When to assume chat completions are not available

    From the provided documentation, chat completions are exposed via the preview endpoint above. There is no indication in the context that GPT-5.1-mini is excluded from chat completions. If, after aligning the endpoint, path, and api-version as described, the deployment still returns 404 specifically for chat completions while Responses continues to work, this would indicate a service-side or configuration issue that requires support investigation rather than a model limitation.

    In that case, capture:

    • Full request URL (with api-version),
    • Deployment name,
    • Resource name/region,
    • Exact HTTP status and body,

    and provide them to Azure support for further analysis.


    References:

    Was this answer helpful?

Your answer

Answers can be marked as 'Accepted' by the question author and 'Recommended' by moderators, which helps users know the answer solved the author's problem.