GPT-Image-1.5 rolling out in the API and ChatGPT

Announcement from OpenAI Developers

GPT Image 1.5 is now available in the API:

  • More precise image editing and preservation of logos & faces

  • Better instruction following and adherence to prompts

  • Improved text rendering, particularly for denser and smaller text

Learn more in the docs:

This year’s “Little Shipmas” comes with a new image model.
Looking forward to what will be revealed later today.

7 Likes

1 Like

Here’s a first post with images:

I have already updated the first post with these.

Click to show again

https://x.com/fidjissimo/status/2000990080840949955?s=20

Here’s my first image. Considering the generation speed, I would say: it’s live!

1 Like

It handles photorealistic into abstract…
The old model struggled with combining such a thing….

(I bet Sam could get away with that style in Europe somewhere)

2 Likes

Quick clock test: it’s 9:30 not 11:55 :slight_smile:
But that’s also the explanation why the clock is extra-large. I hope.

2 Likes

Image edit:

The man is doing a hand-stand on the couch.

Input image

Comparison to image-gen 1

3 Likes

One-shot the crystals in a thread world I was shooting for back in July :eyes:

I fought for hours then and had to write huge contexts to do it…

Not anymore =D

3 Likes

Is this being rolled out in a staggered manner in the US? Hitting the models endpoint only shows gpt-image-1 available and trying to use gpt-image–1.5 with the v1/images/generation endpoint just results in:

Supported values are: 'gpt-image-1', 'gpt-image-1-mini', 'dall-e-2', and 'dall-e-3'.
2 Likes

I am getting

“error”:
{

“message”:
“The resource you are requesting could not be found”,

“timestamp”:
1765913734939,

“name”:
“NodeApiError”,

“description”:
“The model ‘gpt-image-1.5’ does not exist.”,

2 Likes

Doesn’t seem to be accessible to me yet either, but I assume it’ll be available soon.

My question (I didn’t notice directly addressed in the release) – has mask editing been improved?

Cheers!

1 Like

me the same trying to use gpt-image-1.5 in the API i got a 404 “Not Found” ?

2 Likes

same issue here. just says “model not found”

2 Likes

It appears the model is still rolling out in the API.
Hang in there just a bit longer and sorry for making you wait! It’s definitely worth it.

5 Likes

Yeah, if its any consolation i keep getting:

An error occurred while processing your request. You can retry your request, or contact us through our help center at help.openai.com if the error persists. Please include the request ID req_09zdklasdlfaoipsdf in your message.

you can all test it by using an email and a free account in the chat

prompt:
An infografic explaining newton’s prism experiment in great detail

how many words is that?

niiiice.

3 Likes

I suspect the origin is this couch and a forum thread of putting people on it:

ChatGPT with prior model, months ago

Same input today

Aside from not being party colors - it looks like they broke the couch (which was prompted “sagging”) -and it grew in size!

1 Like

Blog announcement link that first focuses on ChatGPT’s version:

https://openai.com/index/new-chatgpt-images-is-here/

  • “20% cheaper”
  • “Enterprises and startups across industries, including creative tools, e-commerce, marketing software, and more are already using GPT Image 1.5”. (aka Now you as a developer in lower esteem can do so also.)

It’s stronger at image preservation and editing than GPT Image 1.

You’ll see more consistent preservation of branded logos and key visuals across edits, making it well suited for marketing and brand work like graphics and logo creation, and for ecommerce teams generating full product image catalogs (variants, scenes, and angles) from a single-source image.

Image inputs and outputs are now 20% cheaper in GPT Image 1.5 as compared to GPT Image 1, so you can generate and iterate on more images with the same budget.

Let’s test it out as an “edits” model - can it outfill a mask with compatibility with the old image, and not alter unmasked areas (like DALL-E 2 could do)? What has been awaited for a long time?

1. Mask area

By resize and reposition an input in my utility built for DALL-E 2:

2. Send off for edits

The new model name is in effect.

Prompt: “Sliding penguin, with a mountain with a ski lodge in the background on a mountain.”

3. Evaluate the result

  • Did the AI preserve the original image that was without mask?
  • Did the AI recognize and use the mask area exclusively?
  • Did the AI outfill the area designated with the prompt language?
  • Was this pixel-perfect to develop image editing tools upon?

2 second animated GIF of before and after (transparency of original is shown):

peng-trant

Conclusion: your image input is still just part of a prompt to a multimodal model - all is regenerated.

3 Likes

Yeah… logos regenerating completely nullifies the logo…

But the claim was ‘more consistent preservation,’ which maybe we read differently.

My testing vector starts with the question now:

Does it handle logos differently when specified, or are they just referring to more consistent to reference input? I don’t think they actually claim not to regenerate…?

You’re always 5 levels ahead…