Summary
In this module, you learned about vision-enabled generative AI models and how to implement chat solutions that include image-based input.
Vision-enabled models let you create AI solutions that can understand images and respond to related questions or instructions. Beyond just identifying objects in pictures, some models can also use reasoning based on what they see. For instance, they can interpret a chart or assess if an object is damaged.
Tip
For more information about analyzing images with the OpenAI Responses API, see, see Images and vision in the OpenAI developer guide.