What fine tune method is used under the fine-tune API for gpt-3.5 in Azure OpenAI?

Weiyu Yan 20 Reputation points Microsoft Employee

I don't know if this is mentioned in the doc but I wonder what's under the hood of fine tuning jobs. Is it Lora? Qlora?

response = client.fine_tuning.jobs.create(
    training_file = training_file_id,
    validation_file = validation_file_id,
    model = "gpt-35-turbo-0613", # Enter base model name. Note that in Azure OpenAI the model name contains dashes and cannot contain dot/period characters.
    seed = 105 # seed parameter controls reproducibility of the fine-tuning job. If no seed is specified one will be generated automatically.
Azure OpenAI Service
Azure OpenAI Service
An Azure service that provides access to OpenAI’s GPT-3 models with enterprise capabilities.
2,490 questions
0 comments No comments
{count} votes

Accepted answer
  1. santoshkc 5,975 Reputation points Microsoft Vendor

    Hi @Weiyu Yan,

    Thank you for reaching out to Microsoft Q&A forum!

    According to the Azure OpenAI documentation, the fine-tuning method used for GPT-3.5 is not explicitly mentioned. However, Azure OpenAI uses a technique called Low Rank Approximation (LoRA) to fine-tune models in a way that reduces their complexity without significantly affecting their performance. This method works by approximating the original high-rank matrix with a lower rank one, thus only fine-tuning a smaller subset of "important" parameters during the supervised training phase, making the model more manageable and efficient. For users, this makes training faster and more affordable than other techniques.

    The large quantities of high-quality training and validation data will be used in gpt-35-turbo-0613 for better performance.

    For mor info: Azure OpenAI GPT-3.5 Turbo fine-tuning and Customize a model with fine-tuning.

    I hope this information helps. Do let us know if you any further queries.

    If this answers your query, do click Accept Answer and Yes for was this answer helpful.

    1 person found this answer helpful.
    0 comments No comments

0 additional answers

Sort by: Most helpful