Batch Support For Azure OpenAI Embedding

Sicong Yang 30 Reputation points Microsoft Employee
2023-04-04T05:46:12.8566667+00:00

According to Azure OpenAI Service REST API reference - Azure OpenAI | Microsoft Learn, currently AOAI's embedding API only accept a max array of 1. I'm just curious about why it doesn’t allow batched embeddings. When would you support batch embeddings?

Azure OpenAI Service
Azure OpenAI Service
An Azure service that provides access to OpenAI’s GPT-3 models with enterprise capabilities.
4,106 questions
{count} vote

Accepted answer
  1. romungi-MSFT 48,916 Reputation points Microsoft Employee Moderator
    2023-04-04T09:45:29.1433333+00:00

    @Sicong Yang I think due to the current demand for the service and compute limitations there are certain limits on the rate and token sizes that can be accepted by the API but there is an option to increase your quota limits through a request submitted through this form. Depending on the model used you can increase requests per minute (RPM) throughput & total tokens (prompt + completion). The requests are similar to the original request to gain access to Azure OpenAI but it can help you improve the current speed if request is approved. Thanks!! If this answers your query, do click Accept Answer and Yes for was this answer helpful. And, if you have any further query do let us know.

    1 person found this answer helpful.

0 additional answers

Sort by: Most helpful

Your answer

Answers can be marked as Accepted Answers by the question author, which helps users to know the answer solved the author's problem.