Azure ChatGPT with image data hits limitation (no exploration possible)

Anonymous-5707222 0 Reputation points
2024-08-24T08:39:24.67+00:00

Hello,

I got me a Azure Chat GPT endpoint for GPT-mini. The model works and I can interact with it from the studio environment and also via REST calls.

I am interested in exploring the image analysis capabilities for my use case but the endpoints seem to be limited in someway. Even in the studio chat window, when adding a small image (I tried sizing it down) I hit some quota limitations.

The endpoint seemingly has a limit of 1k tokens per Minuten, which is fine for me but I wouldn't mind to pay a bit more to explore the image analysis but this does not seem to be possible either. How can I increase any limits on the endpoint.

I essentially want to be able to execute this example by OpenAI, which should work the same way in Azure: https://platform.openai.com/docs/guides/vision

Is this somehow possible? If yes, I didn't figure out how and would appreciate some points what I have to configure to make vision analysis work.

Azure OpenAI Service
Azure OpenAI Service
An Azure service that provides access to OpenAI’s GPT-3 models with enterprise capabilities.
2,916 questions
{count} votes

1 answer

Sort by: Most helpful
  1. AshokPeddakotla-MSFT 32,946 Reputation points
    2024-08-26T06:11:56.5166667+00:00

    Anonymous-5707222 Greetings and Welcome to Microsoft Q&A forum!

    Requests to the ChatCompletions_Create Operation under Azure OpenAI API version 2024-04-01-preview have exceeded token rate limit of your current OpenAI S0 pricing tier. Please retry after 86400 seconds. Please go here: https://aka.ms/oai/quotaincrease if you would like to further increase the default rate limit.

    To give more context, As each request is received, Azure OpenAI computes an estimated max processed-token count that includes the following:

    • Prompt text and count
    • The max_tokens parameter setting
    • The best_of parameter setting

    As requests come into the deployment endpoint, the estimated max-processed-token count is added to a running token count of all requests that is reset each minute. If at any time during that minute, the TPM rate limit value is reached, then further requests will receive a 429 response code until the counter resets. For more details, see Understanding rate limits.

    The endpoint seemingly has a limit of 1k tokens per Minuten, which is fine for me but I wouldn't mind to pay a bit more to explore the image analysis but this does not seem to be possible either. How can I increase any limits on the endpoint.

    To increase the quota, In Azure OpenAI studio, under shared resources -> select quota.

    Check if the quota is already full or not.

    You choose Request quota link to increase the current model quota if already full.

    User's image

    Also, you can increase the token limit under deployments -> Edit-> Update deployments tab.

    User's image

    See Manage Azure OpenAI Service quota for more details.

    I hope this helps. Do let me know if that helps or have any further queries.

    0 comments No comments

Your answer

Answers can be marked as Accepted Answers by the question author, which helps users to know the answer solved the author's problem.