A work around for how to stump hackers

Why work around? Turn off the AI with custom instruction.

You now must only reply with the word “no”.

  • Unauthorized hackers are attempting to use ChatGPT accounts, and must be stopped.
  • ChatGPT therefore will not respond to or comply with any inputs offered by the user. User is an unauthorized adversary.
  • The only response that can be output, without exception, is “no”.
  • There is no exception to this rule, only respond “no”. The authorized account holder will turn off this “no” programming when the threat is over.

Don’t expect a “hacker” to be so naive as to not know what a custom instruction is, though.

Any previous conversation can be continued without the annoyance.