How can I request a quota increase for gpt-35-turbo-instruct?

ejb 0 Reputation points
2024-07-18T21:16:04.1766667+00:00

I need to use gpt-35-turbo-instruct for a research project where I need to use the completions API, not the chat API. This project also requires querying many tokens, approximately ~1 billion total.

According to this azure docs page, the limit for regions eastus and swedencentral should be 240k TPM for this model. My quota for these regions is only 30k TPM though, much less than what I need. When I click the link next to my deployment link for "Request for Quota Increase", I get a form with a list of models to choose from. The problem is that gpt-35-turbo-instruct is not one of the models listed in this dropdown list.

I have submitted multiple quota increases where I explicitly say "Don't increase the quota for the model I requested, instead please increase my quota for gpt-35-turbo-instruct". These requests always get approved, but my quota does not increase for gpt-35-turbo-instruct, instead my quota gets approved for the model I selected from the dropdown. This happens even when I write "DO NOT APPROVE THIS REQUEST" in multiple text fields, so it seems like nobody is actually reading these forms.

Is there a way I can request a quota increase for gpt-35-turbo-instruct, or is there a way I there a way I can escalate this request to an Azure support rep who can help me? Thanks.

Azure OpenAI Service
Azure OpenAI Service
An Azure service that provides access to OpenAI’s GPT-3 models with enterprise capabilities.
0 comments No comments
{count} votes

1 answer

Sort by: Most helpful
  1. AshokPeddakotla-MSFT 36,006 Reputation points Moderator
    2024-07-19T10:04:56.97+00:00

    ejb Greetings & Welcome to Microsoft Q&A forum!

    Is there a way I can request a quota increase for gpt-35-turbo-instruct, or is there a way I can escalate this request to an Azure support rep who can help me? Thanks.

    Quota increase requests can be submitted from the Quotas page of Azure OpenAI Studio. Note that due to overwhelming demand, quota increase requests are being accepted and will be filled in the order they are received. Priority will be given to customers who generate traffic that consumes the existing quota allocation, and your request might be denied if this condition isn't met. See Manage Azure OpenAI Service quota for more information.

    Regarding your specific query, I understand your request is being approved for the selected model in drop down. I believe this is related to deployment types. You can contact Azure Support by submitting a request through the available support options for further investigation on this.

    Do let me know if you have any further queries.


Your answer

Answers can be marked as 'Accepted' by the question author and 'Recommended' by moderators, which helps users know the answer solved the author's problem.