Need Help: Facing OpenAI Usage Violation Due to user's Abuse

Hi everyone,

I’ve developed a free-to-use Next.js website offering multiple AI text generation tools (like AI answer generators, product descriptions, title generators, etc.), using the ‘GPT-4o mini’ model. The site has been made public, and I’m monetizing it through Google AdSense as a part of my blogging activities.

However, I recently received a concerning email from OpenAI, stating that my organization is in violation of their Usage Policies, specifically related to the exploitation or harm of children. Here’s a snippet from the email:

“Hi, Organization org-JYuehevehdhdeuTY’s use of our services has resulted in a high volume of requests that violate our Usage Policies, related to:
Exploitation, harm, or sexualization of children.
We require organizations to use OpenAI’s services in line with our usage policies, including the use of our services by any of their end-users.”

I strongly suspect this is the result of a competitor abusing my platform, as I do not promote or condone any such behavior. I’m currently looking for advice from the community on how I can safeguard my platform against this kind of misuse and prevent any further policy violations.

I’d appreciate any guidance on how to handle this situation. Has anyone else encountered a similar issue? What measures can I put in place to ensure users on my site are following OpenAI’s guidelines?

Thanks in advance for any help!

Have you tried using the moderation API?

1 Like

No not yet, I am not sure if it will delay the response time?
Should I use it as middleware before sending the text to gpt-4o-mini model which I am using??

What you say?

You should ALWAYS do this if you release any features that connect to the OpenAI API publically.
Immediately implement the moderation API!

I can only quote OpenAI here:

We require organizations to use OpenAI’s services in line with our usage policies, including the use of our services by any of their end-users.

Good luck! :hugs:

1 Like

I think either that or run in a separated thread and then bind together before showing the response.
Whichever you find more plausible to implement.

But in the second case, you may need a mechanism to prevent the user from repeating in further violations, which can become very complicated.

Hi, aprendendo
Thanks for the response, let me try to implement it as soon as possible!

Hi, I implemented moderation api. All works fine, but it is becoming really irritating that the api is flagging even normal words like romantic, love, adult. How to resolve this. Anyone has idea??

Do you have an entire prompt that is being flagged? Context matters… Using latest version?

You can consider that the same level and quality of inspection is being applied to examine your inputs and generations anyway to classify the level of account policy abuse. The moderations endpoint should align with the detection that caused the emailed warning.

Plus the moderations endpoint is happenstance, chunking the input for processing, and you can trim or pad the inputs and get significantly different results, just as OpenAI may be inspecting in a different manner than the exact user input or context that you send to the moderations endpoint. Likewise, you can classify user generations or entire chat contexts after the fact and start to score users to see who is the most triggering.

Got similar kind of email that my Organization was sending some requests which violated policies. I think this is an issue from OpenAI, even after moderation enabled this could potentially happen based on how the prompts could be used.

Example: my project is a news summary website, so if the news site has something written which OpenAI may not align to it will flag the content?

OpenAI team, please help and provide better support.

Did they give you any details on the user id numbers misbehaving? Or what type of content they were sending?

This is not how the moderation endpoint works.

This is a great resource if you haven’t checked it out yet…