@Sicong Yang I think due to the current demand for the service and compute limitations there are certain limits on the rate and token sizes that can be accepted by the API but there is an option to increase your quota limits through a request submitted through this form. Depending on the model used you can increase requests per minute (RPM) throughput & total tokens (prompt + completion). The requests are similar to the original request to gain access to Azure OpenAI but it can help you improve the current speed if request is approved. Thanks!!
If this answers your query, do click Accept Answer
and Yes
for was this answer helpful. And, if you have any further query do let us know.
Batch Support For Azure OpenAI Embedding
Sicong Yang
30
Reputation points Microsoft Employee
According to Azure OpenAI Service REST API reference - Azure OpenAI | Microsoft Learn, currently AOAI's embedding API only accept a max array of 1. I'm just curious about why it doesn’t allow batched embeddings. When would you support batch embeddings?
Azure OpenAI Service
Azure OpenAI Service
An Azure service that provides access to OpenAI’s GPT-3 models with enterprise capabilities.
4,106 questions
Accepted answer
-
romungi-MSFT 48,916 Reputation points Microsoft Employee Moderator
2023-04-04T09:45:29.1433333+00:00