Chat in Azure AI Studio fails when max tokens set at 5000

matsuo_basho 30 Reputation points
2024-07-25T21:55:22.5366667+00:00

I'm using the Chat playground in Azure AI Studio, running Meta Llama 3.1 70B instruct with a Serverless compute.

When I set the max tokens to 500, it works fine. However, when I set it to 5000, I get an error:
Request failed with status code 500. Clear the output to start a new dialog.

According to Meta's announcement, the model's context length is 128K, so why is there an issue?

User's image

Please let me know if this is some sort of MS quota issue and if so, whether it's something I have to request and wait for. Perhaps it's the fact that I'm running this on serverless compute.

Azure AI services
Azure AI services
A group of Azure services, SDKs, and APIs designed to make apps more intelligent, engaging, and discoverable.
2,896 questions
{count} votes

Accepted answer
  1. YutongTie-MSFT 52,776 Reputation points
    2024-07-29T16:51:54.5733333+00:00

    Hello Mastuo,

    Thanks for reaching out to us again, since this issue need to be checked with more details, we have redirect it to support team. Please kindly accept the answer here so that we can close this issue in the public forum.

    Once you get your issue solved, please post the solution here if it is convenient to share.

    I hope it helps.

    Regards,

    Yutong

    0 comments No comments

1 additional answer

Sort by: Most helpful
  1. matsuo_basho 30 Reputation points
    2024-08-01T10:44:42.4766667+00:00

    As of now, this is a bug with no solution since the Microsoft team has been investigating this for a few days.

    0 comments No comments

Your answer

Answers can be marked as Accepted Answers by the question author, which helps users to know the answer solved the author's problem.