429 error while running model in AI foundry

Shubham Jalewa 0 Reputation points
2025-06-13T07:19:16.7733333+00:00

Hello team,

Yesterday while running DeepSeek via Azure AI foundry we received a 429 error . We went through the documents where it said that this error occurs when there are too many requests. Can you help me with the same how to fix it.

Regards

Shubham

Azure AI services
Azure AI services
A group of Azure services, SDKs, and APIs designed to make apps more intelligent, engaging, and discoverable.
3,602 questions
0 comments No comments
{count} votes

2 answers

Sort by: Most helpful
  1. Prashanth Veeragoni 4,930 Reputation points Microsoft External Staff Moderator
    2025-06-13T09:55:33.61+00:00

    Hi Shubham Jalewa,

    Azure OpenAI's quota feature enables assignment of rate limits to your deployments, up-to a global limit called your “quota”. Quota is assigned to your subscription on a per-region, per-model basis in units of Tokens-per-Minute (TPM). Your subscription is onboarded with a default quota for most models.

    Refer to this document for default TPM values. You can allocate TPM among deployments until reaching quota. If you exceed a model's TPM limit in a region, you can reassign quota among deployments or request a quota increase. Alternatively, if viable, consider creating a deployment in a new Azure region in the same geography as the existing one.

    TPM rate limits are based on the maximum tokens estimated to be processed when the request is received. It is different than the token count used for billing, which is computed after all processing is completed. Azure OpenAI calculates a max processed-token count per request using:

    ·    Prompt text and count

    ·    The max_tokens setting

    ·    The best_of setting

    This estimated count is added to a running token count of all requests, which resets every minute. A 429 response code is returned once the TPM rate limit is reached within the minute.

    To minimize issues related to rate limits, it's a good idea to use the following techniques:

    1.  Implement retry logic in your application.

    2.  Avoid sharp changes in the workload. Increase the workload gradually.

    3.  Test different load increase patterns.

    4.  Increase the quota assigned to your deployment. Move quota from another deployment, if necessary.

    Remember to optimize these settings based on your specific needs.

    Resources:

    ·   Optimizing Azure OpenAI: A Guide to Limits, Quotas, and Best Practices

    ·   Azure OpenAI Service quotas and limits

    ·   Azure OpenAI Insights: Monitoring AI with Confidence

    Hope this helps. If you have any follow-up questions, please let me know. I would be happy to help.

    **

    Please do not forget to "Accept the answer” and “up-vote” wherever the information provided helps you, this can be beneficial to other community members.

    Thank you! 


  2. Danny Dang 85 Reputation points Independent Advisor
    2025-06-16T07:14:18.9566667+00:00

    Hi Shubham,

    Thank you for contacting Q&A Forum. The 429 error you're encountering typically occurs when the number of requests exceeds the allowed quota. To resolve this issue, you need to increase the quota for your DeepSeek model via a support request.

    Steps to Fix:

    1. Understand the Error: The 429 error indicates that there are too many requests being made. You can find more details in the Azure documentation here.
    2. Request Quota Increase: To increase the quota, follow the instructions provided in the below Azure official documentation.
    3. Submit a Support Request: Navigate to the Azure portal and submit a support request to increase the quota for your DeepSeek model. Ensure you provide all necessary details about your current usage and the required increase.
    4. Implement Retry Logic: While waiting for the quota increase, implement retry logic in your application to handle 429 errors gracefully. This can help manage the load and reduce the frequency of errors.

    References:

    If I have answered your question, please accept this answer as a token of appreciation and don't forget to give a thumbs up for "Was it helpful"!

    Best Regards,


Your answer

Answers can be marked as Accepted Answers by the question author, which helps users to know the answer solved the author's problem.