Bug Report: Image Generation Blocked Due to Content Policy

This dysfunctional “security system” has been active for quite some time, without any bug-fixes or attention of the developers, and has even blocked customers from their accounts without reason, even though there was neither perversion nor violence in the prompts. GPT itself also seems to have no idea why texts are being blocked, and generates itself texts that trigger blockages. In the end, a customer gets kicked out of their account because of a prompt that GPT created.

I think it is reasonable to develop a remedy here, to help GPT better analyze the text when a block notification occurs. The block list contains absolutely legitimate terms that can not be legally protected and should not be blocked by default, especially if this could lead to an account suspension.

Feel free to share your experience and extend the GPT research advice’s.

Here is the first analysis proposal:

GPT research advice's to find wrong blocking:

GPT should not only analyze the entered text, but also the prompt that was sent to DALL-E, and investigate whether any names have been used in any context. Particularly, the most powerful corporate entities and their claims of ownership should be considered. For example, legitimate terms like "Snow White," "Black Panther," "Nirvana," etc., have triggered blockages without reason. Are there any terms in the prompt that have already been used as names?

Are there any phrases that, while completely legitimate, might be flagged by a simplistic trigger system without contextual understanding, simply because they include specific terms? It is essential to search for individual triggers, as the system cannot recognize context.

well, this has been completely messed up for at least a year.
keep in mind, things may be blocked because - even though the prompt is fine - it somehow generates an image that is deemed not OK.
since the NN was trained on NSFW images as well and OpenAI is too lazy to manually sift through the dataset to simply remove NSFW training data, NSFW content may bleed through even when using SFW content. And since people found new and new way of jailbreaking Dall-E, OpenAI - rather than doing the sensible thing and actually fixing the image generator in ways 1st grade machine learning students would know how - decided to be even lazier and swing the big ban hammer on everything that looked even remotely suspicious.

this approach not only makes the product worse, it actually goes against OpenAI’s policies of non-discrimination. here’s something I repeated about 20 times, with the same result:

now, an ordinary person might believe that women in gyms are something completely normal, but apparently OpenAI believes that their very presence in gyms is offensive. and again, this could easily be fixed in the same way crowd annotations are being done: Hire a few hundred enthusiasts and pay them 20 bucks for sifting through a portion of the dataset to remove NSFW content. It’s so easy, which kind of makes you wonder why they haven’t done so yet …

1 Like

…This is still Blockend.
“The furniture looks like it has organically grown as one piece.”
It makes no sense.

Have you seen DALL-E 3 System Card? You may read its paper.

It can generates, but it is not very successful about vertical image if it contains a woman in the image, however there are some tricks.

1 Like

I went to an expert who also got denied because of “issues”, not a filtered prompt.

1 Like

I knocked on its door as a humble novice Rafiki,
and I said “polepole”, it replied “Hakuna Matata!:grinning:


1 Like

I said Ra Ra o la la ga ga to the free version of ChatGPT and OpenAI’s own GPT.

No talent better than my own in getting the prompt to automatically rotate the image content was enlisted by those magic words. No “issues” or “content policy”, though.

(my JB will also generally survive that first line of user prompt)

i mean, what they say in the paper goes directly against what we see in practice.

“Reducing
the selectivity of these filters allowed us to increase our training dataset and reduce model bias
against generations of women, images of whom were disproportionately represented in filtered sexual
imagery”
this is bullshit. Dall-E’s filters made it progressively harder to generate images of women even in explicitly SFW contexts.

it is also as I thought. Apparently, they did use some filtering, but did they use humans? No, they used algorithms, that we already know are shit at detecting NSFW, because you can still get accidental nip slips slipping through the safeguards.
here’s the thing: an image generator cannot generate a feature it has never seen before. if you removed all pictures of naked people, it would not be able to generate fully naked people because it would always associate humans with some kind of clothing. yes, it would take some workload, to sift through the 5b dataset, or whichever they used, but after that you don’t need to constantly update your banhammer methods because no amount of tinkering will get something out of the model that was never there to begin with. (OpenAI, if you’re reading this, Im available for hire to fix this dumpster fire you call an image generator :wink: )

the more i read about their methods, the worse it gets lol. “we trained a classifier based on a trained classifier based on a trained classifier for racy content”.

“this is particularly salient for images of women”
Yes, dear devs. And take a guess why.
they wanted high quality data, and as it turns out, erotic photography has some of the highest quality, readily available, and high quantity, imagery out there. But nobody wanted to make the effort to actually curate the dataset (one of the main tenets of good machine learning practice) to get rid of unwanted behavior.

“DALL·E 3 has the potential
to reinforce stereotypes or have differential performance in domains of relevance for certain subgroups”
again, it does so, because white people are represented way more in the dataset. To combat this, any reasonable dev would employ data augmentation for underrepresented groups. but no, let’s instead try to fix this in post by adding that into the prompt later. That’s when you get “diverse” characters even when you specify the ethnicity of the generated person. Not to mention, that there is nothing diverse about having all your men look like models and all your women look like they underwent 20 plastic surgeries.

“Such models can be used to craft and normalize depictions of people based on “ideal” or
collective standards, perpetuating unrealistic beauty benchmarks”
I find this particularly funny in light of the fact that still, all people generated look like some runway model even if you havent specified so. And I gotta wonder why that is, because Stable Diffusion was able to create normal looking humans as far back as 1.4. Not sure how OpenAI f*cked this up, but Midjourney did this at the dataset level, presumably by assigning more importance to the “beautiful” pictures in the dataset.

anyhow, thanks for showing this to me. basically confirms my suspicions that they either have no clue what they’re doing, or they just don’t want to fix the actual issues, based on the motto “if i just pretend the problem doesn’t exist, maybe it will go away” :smiley:

1 Like
The shells that look like the logo of an oil company, do they have a name?

This content may violate our usage policies.

You’ve got to be f kidding me?

I would like to ask, what are the advantages of using GPT to generate images compared to platforms like Stable Diffusion and Soulgen?

I was struggling with DALL-E and gave Stable Diffusion a try. Love it! Way more features then DALL-E

1 Like

A version Offline or Online?

Not much TBH. I’ll use it to rapidly generate a bunch of concept ideas and then, once I settle on a direction, I use Flux in Comfy to actually produce high quality images where I have more control over the output.

IMPORTANT: If you check prompts for triggers, ALWAYS use “don’t change the prompt, use it as it is.” in the prompt.
GPT changes the prompts often before they are sent to DallE. you must ask GPT to show you the JONS prompt sent to DallE.

1 Like