Question regarding the image generation models Responses API

Hello, I started using the responses API and I noticed that:
It says
Please note that right now, only gpt-image-1 and gpt-image-1-mini are supported in the Responses API. We’re working on support for gpt-image-1.5.

However in another page it says
The model used for the image generation process is always a GPT Image model (gpt-image-1.5, gpt-image-1, or gpt-image-1-mini), but these models can be used as the mainline model in the Responses API as they can reliably call the image generation tool when needed.

but when I try to use any of these 3 (gpt-image-1.5, gpt-image-1, or gpt-image-1-mini ) in the model: field in the request, I get an error saying that the model I used is not available.

Am I missing something? In one page it says that I can use gpt-image-1 and gpt-image-1-mini for the Responses API as mainline and that they’re working on gpt-image-1.5. In another page it says that I can use all 3 as the mainline, but in reality i can only use like gpt-5, gpt-4, gpt-5-mini in the model field of the request. Thanks in advance

Same issue here, attempting to use streaming response method with gpt-image-1.

The responses API doesn’t expose which DALL-E model (3 or 4) generated an image—it’s handled internally based on your subscription and request params. If you’re on ChatGPT Plus/Team, it uses the latest available. No direct way to query it per response right now.

Wait, what, how does the API know you’ve even got a ChatGPT account, the two are distinct?

2 Likes

It looks like the documentation needs an update. Image models cannot currently be used as mainline models in a Responses API call.

If we want to use gpt-image-1.5 for image generation via the Responses API, we need to define a custom tool until the model is integrated as an option for the hosted image generation tool.

Thanks for raising this. I will share this information with the team.

On the Responses API endpoint, you must employ a “chat” AI model, trained on having a conversation and outputting turns as “assistant” and following your instructions.

Then you enable tools you want the AI model to be able to use.

Where the “Image generation tool” documentation page says:

…it is not referring to the names of the image generation models within the tool; the text is referring to the list of recommended AI models that appeared immediately before, such as gpt-4o or gpt-5. They are “chat” models and also models capable of calling internal tools on the Responses API. The text is saying that: you have a more limited selection of AI models available if wanting “chat, with image creation” (behaving like ChatGPT does for free users..). (The interloping model names of image models should be removed from that part of the documentation.)

If you go to the API prompts playground (and ensure you have picked “responses” in the kebab menu drop-down), then you can use internal hosted tools:

Within the “image gen” tool, you then have the tool configuration options also documented when you deeply-expand the API reference, including the image creation models:

If the user than says something where that added image creation tool is useful, such as “create an image of a cute sea monster” (instead of calculate pi to 100 digits..), then the AI may call the tool.

First: be familiar with using Chat Completions, Responses, parsing streams and events as model responses, and delivering responses to users. Then you can move on to enabling tools and receiving their unique events.

Note: highly-undocumented is how this internal tool uses vision and the chat context: many or all images in the chat history are sent to the image tool model, a copy of the whole chat and not just an AI-written prompt. With gpt-image-1.5 seeming permanently billing for “input_fidelity”: “high” at up to 6000+ tokens besides the “vision” price and the “image input” price, a chat with images even unrelated to the latest picture can be an “input” bill of $0.06 for every image recently-discussed before you even pay for a generated image output (which can also be billed when it is refused).

If you want only prompt-based images and controlled costs: make your own function.

Hey everyone, Thank you for flagging. Our engineering team has corrected the documentation. Apologies for the inconvenience.

2 Likes

This topic was automatically closed 2 days after the last reply. New replies are no longer allowed.