Increased latency after switching to GPT-4 version 1106-Preview

Anna-Marie Barborikova 40 Reputation points
2023-12-13T12:26:16.7566667+00:00

We recently switched from GPT-4 version 0613 to version 1106-Preview and have experienced a significant increase in latencies. We have deployed our models in Sweden Central and East US 2, and the latency increase is similar for both regions. We are unsure if this is due to the new version being in preview or if it is related to the model's workload in the given location. Can anyone help provide some insight? Thank you.

Latency in East US 2 for GPT-4 version 0613:

Screenshot of latency in East US 2 for GPT-4 version 0613

Latency in East US 2 for GPT-4 version 1106-Preview:

Screenshot of latency in East US 2 for GPT-4 version 1106-Preview

Azure OpenAI Service
Azure OpenAI Service
An Azure service that provides access to OpenAI’s GPT-3 models with enterprise capabilities.
4,081 questions
{count} votes

1 answer

Sort by: Most helpful
  1. navba-MSFT 27,540 Reputation points Microsoft Employee Moderator
    2023-12-14T09:40:54.7333333+00:00

    @Anna Dombajova Thanks for sharing the details. Please note that this latency issue is an expected behavior.

    Reason:
    This latency is expected considering that gpt-4 version 1106-Preview has more capacity. As of now, we do not offer Service Level Agreements (SLAs) for response times from the Azure OpenAI service.

    Action Plan:
    This article talks about Azure OpenAI service about improving the latency performance.

    Here are some of the best practices to lower latency:

    • Model latency: If model latency is important to you we recommend trying out our latest models in the GPT-3.5 Turbo model series.
    • Lower max tokens: OpenAI has found that even in cases where the total number of tokens generated is similar the request with the higher value set for the max token parameter will have more latency.

    Lower total tokens generated: The fewer tokens generated the faster the overall response will be. Remember this is like having a for loop with n tokens = n iterations. Lower the number of tokens generated and overall response time will improve accordingly.

    Streaming: Enabling streaming can be useful in managing user expectations in certain situations by allowing the user to see the model response as it is being generated rather than having to wait until the last token is ready.

    Please let me know if you have any follow-up questions. I would be happy to answer it.

    0 comments No comments

Your answer

Answers can be marked as Accepted Answers by the question author, which helps users to know the answer solved the author's problem.