Azure OpenAI GPT-4 Turbo with Vision tool (preview)
Azure OpenAI GPT-4 Turbo with Vision tool enables you to leverage your AzureOpenAI GPT-4 Turbo with Vision model deployment to analyze images and provide textual responses to questions about them.
Important
Azure OpenAI GPT-4 Turbo with Vision tool is currently in public preview. This preview is provided without a service-level agreement, and is not recommended for production workloads. Certain features might not be supported or might have constrained capabilities. For more information, see Supplemental Terms of Use for Microsoft Azure Previews.
Prerequisites
Create AzureOpenAI resources
Create Azure OpenAI resources with instruction.
Create a GPT-4 Turbo with Vision deployment
Go to Azure OpenAI Studio and sign in with the credentials associated with your Azure OpenAI resource. During or after the sign-in workflow, select the appropriate directory, Azure subscription, and Azure OpenAI resource.
Under Management, select Deployments and Create a GPT-4 Turbo with Vision deployment by selecting model name:
gpt-4
and model versionvision-preview
.
Connection
Set up connections to provisioned resources in prompt flow.
Type | Name | API KEY | API Type | API Version |
---|---|---|---|---|
AzureOpenAI | Required | Required | Required | Required |
Inputs
Name | Type | Description | Required |
---|---|---|---|
connection | AzureOpenAI | the AzureOpenAI connection to be used in the tool | Yes |
deployment_name | string | the language model to use | Yes |
prompt | string | Text prompt that the language model uses to generate its response. The Jinja template for composing prompts in this tool follows a similar structure to the chat API in the LLM tool. To represent an image input within your prompt, you can use the syntax ![image]({{INPUT NAME}}) . Image input can be passed in the user , system and assistant messages. |
Yes |
max_tokens | integer | the maximum number of tokens to generate in the response. Default is 512. | No |
temperature | float | the randomness of the generated text. Default is 1. | No |
stop | list | the stopping sequence for the generated text. Default is null. | No |
top_p | float | the probability of using the top choice from the generated tokens. Default is 1. | No |
presence_penalty | float | value that controls the model's behavior with regard to repeating phrases. Default is 0. | No |
frequency_penalty | float | value that controls the model's behavior with regard to generating rare phrases. Default is 0. | No |
Outputs
Return Type | Description |
---|---|
string | The text of one response of conversation |
Next step
Learn more about how to process images in prompt flow.