Azure OpenAI GPT Turbo 3.5 Instruct Model

Test Admin 176 Reputation points
2023-10-24T12:54:25.2+00:00

We are switching the Azure Open AI text-davinci-003 model to gpt-35-turbo-instruct (0914) model (Standard tier.)

Region : EastUS

We created one resource in Azure Open AI services and deployed the gpt-35-turbo-instruct model.

As it is mentioned in the Microsoft documentation that gpt-35-turbo-instruct (0914) only accepts the maximum tokens - 4097 tokens
https://learn.microsoft.com/en-us/azure/ai-services/openai/concepts/models#gpt-35-models

But when we use this model, maxtoken is accepted above 4096 tokens(Accepting more than 8000 tokens)

Sample code used to check this model,

String azureKey = "t#######";
String deploymentOrModelId = "deploymentID"; //deploymentid created in Azure portal(gpt-35-turbo-instruct) under Azure open AI services

String endpoint = "https://wyz.openai.azure.com/";

OpenAIClient client = new OpenAIClientBuilder().endpoint(endpoint)
		.credential(new AzureKeyCredential(azureKey)).buildClient();

List<String> prompt = new ArrayList<>();
prompt.add("what is tree");

CompletionsOptions options = new CompletionsOptions(prompt);
options.setMaxTokens(800);
options.setPresencePenalty(0.0);
options.setFrequencyPenalty(0.0);
options.setTemperature(1.0);
options.setTopP(0.5);

Completions completions = client.getCompletions(deploymentOrModelId, options);

Dependencies used,

<dependency>
	<groupId>com.azure</groupId>
	<artifactId>azure-sdk-bom</artifactId>
	<version>1.2.15</version>
	<type>pom</type>
	<scope>import</scope>
</dependency>

<dependency> 
  <groupId>com.azure</groupId> 
  <artifactId>azure-ai-openai</artifactId> 
  <version>1.0.0-beta.2</version> 
</dependency>

Kindly suggest us to solve this problem, is there any setting need to be checked in Azure portal?

Azure OpenAI Service
Azure OpenAI Service
An Azure service that provides access to OpenAI’s GPT-3 models with enterprise capabilities.
4,080 questions
{count} votes

Accepted answer
  1. Saurabh Sharma 23,846 Reputation points Microsoft Employee Moderator
    2023-11-03T19:41:17.09+00:00

    @Test Admin This looks like an expected behavior, however, it is not recommended to exceed this token limit. You can find the below note in the documentation over here as well. User's image

    Please let me know if you have any other questions.

    Thanks

    Saurabh


    Please 'Accept as answer' and Upvote if it helped so that it can help others in the community looking for help on similar topics.

    1 person found this answer helpful.

0 additional answers

Sort by: Most helpful

Your answer

Answers can be marked as Accepted Answers by the question author, which helps users to know the answer solved the author's problem.