I’ve worked very hard on prompt engineering to fix the style of images.
Looking at the system prompts and rules of DALL-E 3, I created a special long prompt that complies with them and generated images based on it. I was thrilled to finally succeed!

However, when I send the same prompt to the DALL-E 3 API, the style gets completely broken for some reason.
I’ve spent about $64 on the DALL-E 3 API alone, and as far as I can tell, the prompt completion for the DALL-E 3 API seems to perform poorly.

I have no idea if this is due to different image generation models being used or different prompt processing systems.
(Is it a difference between GPT-4 and GPT-4-Turbo?)
The differences in the prompt processing behind the scenes are too significant, and I am very dissatisfied.

3 Likes

[Supplement]
I’ve used proper specifications such as “anime style with delicate lines” and “comic style” without using any artist or studio names for both.

I’m encountering a very similar issue. Sending the exact same prompt to DALLE 3 in the ChatGPT app produces a very nice result, but sending the prompt to the DALLE 3 API returns worse images with unwanted artifacts (e.g. text). I compared revised_prompts from both and cannot find any notable differences, so I’m wondering if a different image generation model is used? Or possibly a new worse model was rolled out to the API and ChatGPT app at different times?

3 Likes

@owencmoore
Do you have a plan to resolve this?