I wanted to use a watermark “Generated by AI” while displaying the LLM generated content. Does anyone know of any solution. I was counting on Open AI to integrate that with the model, but looks like open ai took that feature out.
Not sure I follow your request, can’t you just add the text “Generated by AI” in your code?
If I add it at the bottom, anyone can copy the top part and share. If I have the watermark behind, that cannot be done
How would you suggest it is done? I don’t see how this is possible.
I remember reading a few articles about OpenAI wanting to do watermarking on text, but I don’t know if they ever figured it out. Maybe a sequence of tokens or invisible characters? I dunno…
Ahh, ok, well a statistical watermarking method may be possible, but it is then a fairly trivial matter to pass that text through a second AI that just makes non modified word choices, I think that’s why it was ultimately rejected, the bypassing step is very simple to implement.
Yeah… and if they did figure something out, it would likely go into a AI detection tool rather than being shared openly… Found the source…
Who Broke The News?
Scott Aranson, a computer scientist working for OpenAI for AI safety and alignment, reported at one of his talks at the University of Texas that the team is working on cryptographic watermarking of the AI-generated content from Chat GPT to develop a distinction between human-generated and AI-generated content.
This is how Aranson puts it:
“For GPT, every input and output is a string of tokens, which could be words but also punctuation marks, parts of words, or more — there are about 100,000 tokens in total.
At its core, GPT is constantly generating a probability distribution over the next token to generate, conditional on the string of previous tokens.
After the neural net generates the distribution, the OpenAI server then actually samples a token according to that distribution — or some modified version of the distribution, depending on a parameter called ‘temperature.’
As long as the temperature is nonzero, though, there will usually be some randomness in the choice of the next token: you could run over and over with the same prompt, and get a different completion (i.e., string of output tokens) each time.
So then to watermark, instead of selecting the next token randomly, the idea will be to select it pseudorandomly, using a cryptographic pseudorandom function, whose key is known only to OpenAI.”
OpenAI engineer Hendrik Kirchner has also developed a working prototype for it.
They came to same conclusion…
“Now, this can all be defeated with enough effort.
For example, if you used another AI to paraphrase GPT’s output — well okay, we’re not going to be able to detect that.”
So, unless OpenAI has not changed the watermarking plan, as Aranson explained it, all those using the AI-generated content are in luck, but they would need a bit of effort.