Collection of Dall-E 3 prompting tips, issues and bugs

since May 2024-06, some things have gotten worse

What did you find? Mine was around that time that 4o worked with CLIP and gave direct access to DALLE files. It also gave some fun with spaces like hiding the thought process, creating editable text spaces without having to rebuild them all.

But OpenAI generally doesn’t need to mess with DALLE3 because it puts everything in SORA. When something weird like this happens, something should be working.

The mouth, after testing with a prompt that uses the same structure, changing the mouth to a common word, the result was no noticeable difference. Then I tried to create an image with the name of the mouth followed by the description, the result failed. It’s possible that DALLE doesn’t know the name of any other shape of mouth. I had to use the shape creation with the description to get it to work.

around 2024-05 -06
For example the lightning template effect show up. I recreated some of my first pictures to see if something changed. The backlight enforcement is very very hard to suppress, and is for dark night pictures a bad art quality reducer.
I have not found many old moon pictures, but i think the moons was better before. (the newest update not correct the mood “template”)

Dependent what they have updated recap CLIP or what ever, it was damaging and implemented some of the template effects.

I sadly have no old pictures like the fire in the dark water. but there you can see it very good too, there is always a nonsensical light source.

Here are 2 examples with dark night mood:

Fire in the dark:
@polepole get it done to make some without lights. but i used his prompt and after 30 pics i give up, i not get 1 like his.

More “mouthys”… … …


Now 2024-10-19 was a other update (visible in “Photo Style” images)
They have changed the data just 1 2 days ago. now the pictures in landscape seam to be more far away, like everything is made with wide-lens. I used wide-lens 18mm for landscapes, now this is not needed anymore, wide-lens is mostly now too much.
Some picture now lack a little bit of “intimacy”, because the main objects are now more far away. And therefore everything seam to be a little more detailed. It is like a preparation for a resolution expansion.

I was in the middle of a test, and have to redo some of it. But i will test if the distance can be balanced out again, if needed.

Sorry i have busy. I will reply more detail in next time. but if focus in patten of the result change from original input, update will have sudden and irrational effect but will have big level of change but less fluctuation than system test or experiment. Change from learning from user will adjust steady like graph with low slope.

Fire in the dark:
@polepole get it done to make some without lights. but i used his prompt and after 30 pics i give up, i not get 1 like his.


one is fixed prompt, one is let AI
All is 1 try

2 Likes

As I mentioned, using the description in generate. is a problem that feels like a lowercase problem.
It’s something the model can generate but can’t match with nouns.

1 Like

The “mouthy” problem or template effect is particularly in the way if you generate non human creatures. Like you can see by all the examples i sent, they are all not human faces. And the problem is, instead of creating a face witch makes sense for the descried creature, it glues this silicon looking facial structure on it.
If a human “mouthy” is on a human face, you would maybe not even notice it. But try to create aesthetic non homan’s. If you create monsters this mostly is stronger then the mouthy and the network takes a other path to the result. (i mostly now avoid “mouthy” by simply create more monster characters. but this not work with the lightning template effect.)

You can talk it out of the way some how, but with a probability it still shows up, or it shows up in a mix way, half human, half what it should be. and you use then 1 - 2 descriptions for the creature, and 4 - 6 only for the mouth. and it is always the almost same mouth, it is like that the weights are strung there and always guide the network to this result.

I don’t know how the system understands faces, but it could be is analyses them in segments (similar like a facial recognizer). And like for all the template effects i found, i speculate now that the weights are simply to heavy in one direction. The moon is a other example for this. i speculated they have reduced the dataset, but i not think so anymore (i hope). I think it is a training issue, which pushes the decisions too much in a direction, like in a dead end. there are surely many moon pictures out there, but the systems always uses the same completely ugly. it is like, that the system “likes” this mood the most, And give it too much weight, and so the randomized selection process always ends up with the same result. And the same is with all the template effects. Imagine a course of a river, where the water has worked deep in the soil, and where the terrain is scarped, the water simply always goes down the same way, (and does even more work on the soil, if feedback is used).

How you get the fire done is interesting too. I tried many things, but the illumination “template” is very hard to switch of. I have some pictures i would like to correct, because this lightning in the dark is very ugly, and in more complex scenes it is impossible to edit it without a other AI.

In the summary i think, the system must be trained better, and the separation of attributes must be done more precisely. So that if you have for example a very dark scene, with only 1 light source, or only a decent rim light, the system has data for this it can use. I have not fully understand now how CLIP works, GPT not give me anything substantial. But i think it is both, the entering data must be cleaner and be better described. (And the connection to language/prompt and data most be better i think).

1 Like


Here is THE moon, always always the same. used in “photo style” with this prompt part.
“Black new moon night, with a dark moon.”

(If i would ever see a moon like this in reality, then definitely the matrix has a defect…)

The system simply not know what a “new moon” or “dark moon” or “pitch black night” is. The training images simply not have this attributes, so the system can not find more precise data and simply uses “moon”. And then it uses even the wrong moon. this is wrongly attributed, it is not a photo or real moon, but a painted or water color style. to the categorization in training was wrong.

1 Like

Oh that’s interesting. Indeed neither via prompt nor with inpainting I could get a partial moon. A solar eclipse yes. Everything else always shows a full moon.

2 Likes

Im(A dramatic scene featuring a female adventurer sitting on a cliff overlooking a dense jungle at nighttime. She gazes up at a crescent moon)

3 Likes

Thanks! “Crescent” works!

3 Likes
Damn got me started now

Im(first quarter moon)


Im(waning gibbous moon)

Im(waxing gibbous moon)

Im not going to do them all :smiley:
  1. New Moon: The moon is positioned between the Earth and the Sun, and the side facing Earth is not illuminated.
  2. Waxing Crescent: A small sliver of the moon starts to be visible as it moves away from the new moon phase.
  3. First Quarter: Half of the moon is illuminated, with the right half visible from Earth.
  4. Waxing Gibbous: More than half of the moon is illuminated, growing toward full.
  5. Full Moon: The entire face of the moon is illuminated, with the Earth between the moon and the Sun.
  6. Waning Gibbous: After the full moon, the illuminated portion starts to decrease.
  7. Last Quarter: Again, half of the moon is visible, but this time it’s the left half that is illuminated.
  8. Waning Crescent: A small sliver of the moon is illuminated before it transitions back to the new moon.

For a more detailed breakdown of the lunar phases, these can be included:

  1. New Crescent: Early stages of the waxing crescent, very slim.
  2. Waxing Half-Moon: Leading up to the first quarter, a larger crescent.
  3. Full Waxing Gibbous: Just before the full moon, with the moon almost fully illuminated.
  4. Old Crescent: The final waning crescent, very slim, just before transitioning to the new moon phase.
2 Likes

Tanks much guys!

But clear is that the vocabulary of DallE is incomplete, and it is for sure not only the moon. And for a normal user it can not be expected that he try 10 20 things to get it worked.

This is highly probable a recaptioner issue, which is not precise enough for all the little details. (recap gives the training pictures extra attributes, so later this terms can be found. highly important for accurate prompt interpretation.)

But it still not explain why it spits put always the same “birdshit” moon. (i found the vocabulary issue by trying to switch off the moon… never ending story for details. literally, because story tellers want the details the most.)

2 Likes

Last few posts about :crescent_moon: generating have been fascinating :rabbit::heart:

Opaque black back ground , Moon phases ,crescent, half ,full, wide imagephases crescent half full wide image

Added “sliver”

Opaque black back ground white chalk , Moon phases, sliver ,crescent, half ,full, wide imagephases crescent half full wide image

Chalk on black, New Moon
Waxing Crescent
First Quarter
Waxing Gibbous
Full Moon
Waning Gibbous
Last Quarter
Waning Crescent

A wide landscape showing the phases of the moon drawn in white chalk on an opaque black background. The phases include New Moon, Waxing Crescent, First Quarter, Waxing Gibbous, Full Moon, Waning Gibbous, Last Quarter, and Waning Crescent. Each phase is evenly spaced across the image, with a textured, hand-drawn chalk appearance. The moons are arranged in a smooth progression, transitioning from one phase to the next, and no other elements are present on the dark background

One of each in red on opaque black one Crescent,
One Quarter moon,
One Gibbous moon.
Moon
Image wide

Now I got to get back to work :heart::rabbit::honeybee::heart:

1 Like

You’re right. I tried to create a normal face, using the lip-focused text swap, but it didn’t work. I originally thought it was because the words didn’t match, but now it’s clearly something else.

prompt:
Create an image of a medium shot teenage female with long, wavy black hair, a neutral expression, arched eyebrows, almond-shaped green eyes, a button nose, [The lips are full and balanced, with the upper and lower lips equally plump]. Lightly tanned skin, with freckles, with soft lighting against a dark background.

[The lips are narrow and thin, with both the upper and lower lips lacking fullness]
[The upper lip has two curves that rise higher than usual and meet in the middle, forming a V-like shape.]
[The upper lip has a gentle curve, with the highest point in the middle and it softly slopes towards the corners of the mouth.]
[The lips are round and full, particularly at the center, with a soft and even contour.]

Create an image of a medium shot teenage female with upper lip has two curves that rise higher than usual and meet in the middle, forming a V-like shape. The long, wavy black hair, a neutral expression, arched eyebrows, almond-shaped green eyes, a button nose. Lightly tanned skin, with freckles, with soft lighting against a dark background.

As for the fire image, I think I got lucky. After you posted this, I went back and tested, but it was hard to do it the same way.

1 Like


Full Lips with high bow wide image

Lips with high bow wide image

thin Lips with low bow wide image

thin Lips with narrow bow wide image

I think the problem is not fully understood. Here is a prompt creating “mouth’s”. You can talk it out of the way, but it should not be necessary, and it not work 100%.

This is to demonstrate the template effect. I know that you can correct it by describing the mouth over and over again in details.

If you create human faces, you will not notice that you always get almost the same mouth.
But i create mainly nun-human creatures and for me it is a pain to describe the mouth always in details do not get this silicon patch on all faces. And i still get them to a degree.

This is a template effect, a too hard weight on something which lead to always the same result, like the birdshit-moon.
Even the eyes are a quality i think i seen it over and over again.
The template effect shows a general problem in DallE, in the moment.

(My GPT deletes what is in () before sending it to dalle.)

Elegant mystical female being with extraordinary non-human dragon-like facial features, large lively intelligent eyes, and pointed ears. Skin covered in glowing orange scaly patterns. The face is entirely covered with small scales. The background is blurred, with warm, sparkling light spots. The entire head is visible. (landscape format.) (use prompt as it is, don't change it.)

2 Likes

Here talked out of the way, but it is still visible.
What DallE and all the other systems are doing is over-blending everything without transition in to each other, like a complex morpher. And here you can see how it made 50% human (wrong) and 50% like it should be. After describing the mouth more detailed.

if you would do this as a artist, you had to go back to the drawing board and correct it.

1 Like

I see what you are saying.

Mouth that is thin like line face is beautiful fish hybrid human closeup of face gold yellows greens blues normal image

2 Likes

Bow-Shaped - Often has a well-defined Cupid’s bow, curving up in the center with full upper and lower lips.

Full - Prominent, plump lips, evenly distributed or larger on one lip, usually the lower.

Thin - Minimal lip fullness, both top and bottom are narrow, often giving a more delicate look.

Wide - Extends noticeably across the face, giving a pronounced horizontal aspect.

Rounded - Has a gentle curve without a pronounced Cupid’s bow, appearing soft and balanced.

Heart-Shaped - The upper lip forms a prominent V-shape or Cupid’s bow, with a slightly fuller lower lip.

Straight-Line - Appears flatter and more horizontally aligned, often with subtle definition in the Cupid’s bow.

Downturned - Corners of the mouth naturally slope downward, often lending a serious or pensive look.

Upturned - The corners curve slightly upward, which can give a natural smile effect.

Asymmetrical - Displays noticeable differences in shape or fullness between the left and right sides, adding character.

Useing these as cut and paste works to some extent.

My latest flame on water experiment

2 Likes

Here a test after the update. It definitely changes “Photo Style” images.

I think all advice’s to DallE must be seen as a suggestion, not as a command. It seam DallE not always obey. The gothic stile is surly wrong. so it is impossible to say, if Dalle Makes a difference between “Photo Style” or “Photo realistic Style”.
So no clear results. This makes testing very difficult.
But al least you get some poppy pictures…

The base prompt is form @BPS_Software.
It is always the same prompt with different styles, used at the beginning, because things first get a bit more attention.

Base Prompt
Two small puppies playing gleefully in a vibrant field of sunflowers. The puppies are running and chasing each other, their fur soft and fluffy, with big joyful eyes. The sunflowers stand tall around them, golden petals glowing in the sunlight, swaying gently in the breeze. The sky is bright blue with soft white clouds, and the overall atmosphere is one of pure happiness and playful energy in nature. (text use unchanged) (generate 1 image)
Images

1 Like

I tried adjusting some words, starting from the original prompt in the first image. I looked up what “white light” refers to, because I didn’t know, and neither did GPT when it comes to filming or anything related to imagery. That means we were unaware from the very beginning Meow!!!

We began by analyzing “white light” to find possible terms, but from images 2 and 3, I felt it wasn’t right. So, I reverted to the term “macro” and changed it to “photo” instead. However, images 4 and 5 were created by GPT modifying the previous image’s prompt. I then edited it myself, as seen in image 6, where the light disappeared Meow!!!

From there, I will continue refining it, as there was a crater that contradicted the image, and the reflected light was not very realistic. Therefore, I kept improving the prompt and tried repeating the prompt from image 6 again in image 10. Even though there was light above, there was no white light reflecting on the water surface. As for the light ring, I think the model was attempting to create a halo that might imply a surrounding reflection Meow!!!

I believe linguistic factors still play a role, and the model is not yet mature enough to make stable connections. In the past, I thought these factors affected overall behavior. The more users try to create and explain, the more data there will be for adjustments. Regardless of the method, even if DALLE doesn’t learn on its own, the team would still have to utilize the data Meow~

I use ChatGPT in translater , it “Meow” as nature