Hello Say,
The reason this happens is because Azure OpenAI does not currently support image output features, such as generating or modifying images (e.g., drawing bounding boxes), even though the underlying o3 (GPT-4o) model can support it on OpenAI’s platform.
This is due to a feature gap in Azure's API offerings, not a limitation of the model itself.
This feature gap exists in Azure Because:
Azure OpenAI is a hosted version of OpenAI models but with governed, secure, and enterprise-ready deployment features. Because of this:
1.Security & Compliance
o Microsoft often restricts advanced outputs (like images/audio) until they can ensure responsible use, security, and regulatory compliance.
2.Platform Maturity
o Azure OpenAI lags slightly behind OpenAI’s platform in releasing cutting-edge capabilities like multimodal image/audio output.
3.Preview Phase
o As of June 2025, GPT-4o on Azure is in Preview, and many features (e.g., image generation or output) are not yet available.
4. Custom Deployment Layers
o Azure wraps OpenAI models with a different API surface (aoai.openai.azure.com) compared to api.openai.com, which limits some direct capabilities.
Please refer to the below documents for more information:
Hope this helps, do let me know for further details.
Thank you!