An Azure service that provides access to OpenAI’s GPT-3 models with enterprise capabilities.
Hi GS,
Thanks for sharing your observation. Yes, you are right that the model is available in that region, but what you are experiencing is a different issue. It is not about availability; it is about performance.
Even when a model is supported in a region, performance can vary depending on a few factors.
One common reason is regional load. Some regions like East US2 are heavily used, so requests may take longer or responses may feel slower or inconsistent. This does not mean the model is not working, but the backend capacity may be under higher demand at that time.
Another thing to check is your deployment configuration. If the deployment is using a lower throughput setting or limited capacity, it can affect response speed and quality. Increasing the capacity or checking the tokens per minute allocation can sometimes improve performance.
It is also worth verifying the prompt and usage pattern. Larger prompts, long conversation history, or complex instructions can increase latency and sometimes give the impression of poor performance.
You can also try testing the same model in a different region if possible. If the same request gives better results in another region, then it confirms the issue is related to regional load or infrastructure differences rather than the model itself.
In simple terms, supported region means you can use the model there, but performance is not always exactly the same across all regions.
If you can share a bit more about what kind of delay or issue you are seeing, such as slow response or low-quality output, it will help to narrow down the exact cause.
I Hope this helps. Do let me know if you have any further queries.
If this answers your query, please do click Accept Answer and Yes for was this answer helpful.
Thankyou!