Prompt caching in Azure OpenAI?

Christian 100 Reputation points
2024-10-02T11:39:05.5933333+00:00

OpenAi recently announced Prompt Caching in the API. Apparently it caches input tokens when the promt is larger than 1024 tokens.
https://openai.com/index/api-prompt-caching/

Is this feature enable in Azure OpenAI? If it's not, is there an ETA?

Azure OpenAI Service
Azure OpenAI Service
An Azure service that provides access to OpenAI’s GPT-3 models with enterprise capabilities.
3,198 questions
{count} votes

Accepted answer
  1. Abel Wenning 81 Reputation points
    2024-10-22T18:20:14.3933333+00:00

    Yes, as of yesterday, per Microsoft:
    Currently only the following models support prompt caching with Azure OpenAI:

    • o1-preview-2024-09-12
    • o1-mini-2024-09-12

    Reference: https://learn.microsoft.com/en-us/azure/ai-services/openai/how-to/prompt-caching

    Cc: @YutongTie-MSFT , @ Christian-7033 @TienHoang-5634, @Dexter Awoyemi , @koritadenadakddp3-4938 (sorry, I don't know how to effectively tag the other 3 [or 8] people)

    1 person found this answer helpful.
    0 comments No comments

1 additional answer

Sort by: Most helpful
  1. Saravanan Palanivel 5 Reputation points
    2024-10-24T08:43:30.15+00:00

    As of 23rd October, prompt caching is extended for following models as well

    • gpt-4o-2024-05-13
    • gpt-4o-2024-08-06
    • gpt-4o-mini-2024-07-18

Your answer

Answers can be marked as Accepted Answers by the question author, which helps users to know the answer solved the author's problem.