An Azure service that provides access to OpenAI’s GPT-3 models with enterprise capabilities.
Hi Aditya Purohit,
Following up with an update from engineering.
Based on our investigation, the elevated latency observed on GPT‑4.1‑mini and related models in the Central/South India region was due to a service-side issue, which is being tracked under INC 747008667.
Issue:
Requests were experiencing higher-than-expected end‑to‑end latency (>1000 ms) due to a backend service regression affecting request handling and routing in the region.
Resolution:
Engineering has applied a service-side mitigation to address the latency issue. The fix restores normal request handling behavior and improves response times. No configuration or code changes are required on the customer side. The improvements will be reflected automatically as the fix completes rollout.
If you are still seeing higher latency after this update, please share:
Deployment name
Region
Approximate request timestamps
This will help us validate that traffic is hitting the updated path.
If you have any remaining questions or additional details to share, feel free to let us know. We’ll be glad to provide further clarification or guidance.
If this answers your query, please do click Accept Answer and Yes for was this answer helpful.
Thank you!