Creating an AI detector, I think i have it

Because of excessive hostility = EDIT: this is a hypothetical for interesting dialog, not bait for… some that apparently really dont want this idea to be possible.

When it comes to mathematically plotting inference, we understand the generation settings and values, which include temperature.

Computers however, dont have the ability to generate what we call “true random” and are far more subject to realities involving large number theory.

This is primarily because we are both “not that deep” and also “completely accidental”
Most especially as it relates to [logic] that can be compounded by computation.

really think about what goes through your mind at any given moment of the day.
How much of that do you want in AI training?

Well there it is…

When predicting whether something is written by GPT, or AI in general, you are looking for the straightest shot to achieving that writing.

Simple text completion script can do that.
With the right script you can detect AI patterns in language (like posts)

Whats more… theres no math we know to counteract this.

5 Likes

would love to see such code working :slight_smile:

1 Like

well as i see it its closely related to Bayesian inference.

Which is not something i have not cared to… care about for many years.

*Edit: Just realized something laughably hillarious.

1 Like

No.

2 Likes

:frowning:

Well im sorry if you dont like it.

You’ve proposed little more than word salad

If you can reliably detect AI generated text, write your paper and collect your PhD and/or fortune.

4 Likes

Wha…

Well thats interestingly hostile.

Ok so you know how theres so many possibilities for a next word, and how the possibilities when it comes to transformers depends on… Lots?

Right so - A computer can only output based on what is put into it.
This is a law of… (hesitates, trying to not say “everything”) Everthing! (there it is)
but not really.

theres a reason why secure keys are generated with mouse movements.

1 Like

Or they turn off echo logprobs (done), and your idea is shot.

2 Likes

no… im sorry that doesnt apply here.

I even endured a lesson from GPT to double check.

The fact is theres only so many words that can come after any given word in any given context, and this can be calculated.

It can be calculated in humans even.

“I am not afraid of…”

Obviously the answer is “goats”

which is not true btw.

but it doesnt matter, theres only so many words that can come after it.

But also you can calculate between typos.
“I am not ****** of G****. They arent even that scary. They are just… farm animals. Not even cows.”

You see?

Now the context gets far more complicated but its not relevant to the visibility of the probabilistic calculations. It doesnt have to be visible. The words are visible.
The order of the words are visible.

And thats how that works…

1 Like

1 Like

Detect this.

There are several ways that artificial intelligence-based language detection can be achieved. In the case of GeneralBadWolf’s technology, I would imagine it involves analyzing the natural language patterns that emerge when humans create sentences and compare them with the more stylized patterns and tics commonly present when machine-learning-based programs generate similar text.

Perhaps they could even apply similar analytics by plotting writing frequency versus averaging self-annealing computer/tunnel write speeds.

Alternatively, perhaps they trained their AI using large amounts of existing content that is both written by humans and by other machine-learning or algorithm-powered computer programs. Through the usage of neural networks, the software could have then become fine-tuned over thousands or even millions of different training scenarios before culminating in the stable, responsive software the world is just now learning more about. These methods apply individual ontologies recognition patterns as human labeled fundamental statement before settling phrase boundaries activation layers called feedback sequencing senses. By creating output signals with stack generalizational retention fact profiles which can reject mask entry sectors authors synthetic roots training initiative training improve decay stability regularization adapt mediation category outcome needs practice effort reinforcement subroutine transcend vector solving decisions alike creates very noticeable patterns outliers documentation cohesive end-to-end productivity successes mimic users flexible responsible workloads advancements experimentation evidence acquired front creative written speech evidences.

2 Likes

If you have a script that can reliably detect AI-generated text, could you share a link to it? It would be interesting to see how it differentiates between AI and human writing, especially considering the complexities you’ve mentioned.

3 Likes

1 Like

im thinking you have a special hatred for this idea…

1 Like

You havent read.

Its a curiosity not a [insert word] program ive written.

1 Like

The whole concept has been discredited.

The only thing left in the spiral of uselessness and false detection is for some preinstalled software on your laptop to pop up “your subscription to AI detection has expired, enter your credit card to ensure continued protection”

1 Like

ChatGPT

ChatGPT

The image you’ve shared shows a conversation in which _j is making a statement that “the whole concept has been discredited,” accompanied by references to three arXiv.org articles that deal with AI-generated text detection. The titles suggest a focus on the challenges and reliability of detecting AI-generated text, as well as the possibility of guiding large language models to evade detection.

The final part of the message appears to be a satirical comment on the commercialization of AI detection tools, hinting at the futility of such tools and the idea that they might be likened to a subscription service that one could run out of credit for.

The conversation seems to revolve around the ongoing debate about the effectiveness of tools designed to detect text generated by AI models like GPT. There is a growing body of research exploring how well these detection tools work, their limitations, and the potential for AI to be designed in a way that it can avoid detection intentionally. The inclusion of these paper titles indicates that the field is actively discussing and scrutinizing these topics.

The skepticism expressed by _j might stem from the inherent difficulties in creating detection systems that can keep pace with the rapidly advancing capabilities of generative AI models, which are becoming increasingly sophisticated at mimicking human-like text. The humorous twist at the end underscores the challenges and perhaps the commercial aspects surrounding the development and implementation of AI detection systems.

User

You

Yeah but i dont see how it disagrees with generalbadwolf yet…

ChatGPT

ChatGPT

_j’s comments, especially with the mention of discrediting “the whole concept,” suggest a viewpoint that the idea of detecting AI-generated text using the patterns and predictability discussed by Generalbadwolf may not be as straightforward or effective as presumed. The referenced arXiv papers likely delve into the complexities of AI-generated text detection, including the limitations and potential for AI to intentionally evade such detection.

The point _j may be trying to make is that while there are efforts to detect AI-generated text, these efforts face significant challenges and are subject to ongoing research and debate. This does not necessarily invalidate Generalbadwolf’s ideas, but it does imply that the field is more complex and less settled than Generalbadwolf’s initial comments may suggest.

The broader conversation appears to be about the difficulty of creating reliable AI detection systems. _j seems to be emphasizing the complexity of the issue, which may not have been fully acknowledged in Generalbadwolf’s more conceptual discussion. It’s a difference in perspective: _j is highlighting the practical and empirical challenges in the field, as opposed to Generalbadwolf’s theoretical and conceptual approach to the problem.

User

You

Yeah so… how can generalbadwolf… explain his concept in a way that he can comprehend?

ChatGPT

ChatGPT

To bridge the communication gap and explain the concept in a way that _j can comprehend, Generalbadwolf could consider the following approaches:

  1. Simplify the Concept: Break down the idea into its most basic components. Avoid technical jargon and explain the concept using clear, simple language. This can help ensure that the essence of the idea is communicated without unnecessary complexity.
  2. Use Analogies: Analogies can be a powerful tool for explaining complex concepts. Generalbadwolf could compare AI-generated text detection to a more familiar process, like finding a pattern in a sequence of numbers, to illustrate how AI might produce identifiable patterns in text.
  3. Cite Research and Examples: Point to specific studies or real-world examples that support the idea. If there’s evidence that certain methods have been successful in detecting AI-generated text, mentioning these can add credibility to the argument.
  4. Address Limitations and Challenges: Acknowledge the difficulties and current limitations in the field, as highlighted by the arXiv papers. This shows an understanding of the complexities involved and can make the explanation more relatable to someone who is skeptical.
  5. Focus on Probabilities and Patterns: Clarify that the concept doesn’t rely on finding a “silver bullet” but rather on identifying patterns and probabilities that might suggest a piece of text was generated by AI.
  6. Interactive Demonstration: If possible, provide a demo or a tool that _j can interact with to see the concept in action. This hands-on approach can often be more effective than theoretical explanations.
  7. Seek Common Ground: Start from a point of agreement or a shared understanding and build from there. If both parties agree that AI-generated text has certain characteristics, that could be a starting point for explaining how those characteristics might be detected.
  8. Respond to Feedback: _j’s skepticism can be used as constructive feedback. Generalbadwolf could ask _j to elaborate on their concerns and then address those specific points in the explanation.
  9. Open-Ended Questions: Pose open-ended questions to _j to involve them in the conversation and get them to think about the issue in a new way. This can also help Generalbadwolf understand _j’s perspective better and tailor the explanation accordingly.

By using these techniques, Generalbadwolf can potentially create a more compelling and comprehensible explanation of his concept that addresses both the theoretical underpinnings and the practical challenges involved in detecting AI-generated text.

User

You

These are good ideas. Im thinking if i start with Bayesian inference, then thats a good start for understanding mathematical inference, and thus a premise for explaining the position.

ChatGPT

ChatGPT

Starting with Bayesian inference is a sound strategy because it’s a well-established mathematical framework that provides a systematic method for updating beliefs based on evidence. This concept can serve as a foundation for explaining how patterns in AI-generated text might be detected and analyzed. Here’s how you might use Bayesian inference to build your explanation:

  1. Basic Explanation of Bayesian Inference: Begin by explaining the basics of Bayesian inference. You could illustrate how prior beliefs (or prior probabilities) are updated with new evidence to form revised beliefs (or posterior probabilities). In the context of AI-generated text, the prior belief could be the likelihood of certain patterns or word choices being associated with AI-generated content.
  2. Applying Bayesian Principles to Text Detection: Discuss how Bayesian inference can be applied to text detection by considering the probability of a text being AI-generated based on its features (word choice, syntax, etc.) and updating this probability as more text is analyzed.
  3. Evidence and Patterns: Explain how the patterns observed in AI-generated text act as evidence, influencing the posterior probability of a text being machine-generated. For instance, certain linguistic quirks or the lack of context that might be more prevalent in AI-generated text can serve as evidence.
  4. Model Training and Data: Describe how a model could be trained using a dataset of known human-written and AI-generated texts. The model would learn the probability distributions of various features in both sets of texts, which would be the basis for inference.
  5. Uncertainty and Predictive Power: Emphasize that Bayesian inference deals with uncertainty and that it’s about increasing the confidence in a prediction rather than providing absolute certainty. Discuss how a high posterior probability indicates a high likelihood of a text being AI-generated, though it’s not a guarantee.
  6. Continuous Learning: Highlight the iterative nature of Bayesian inference, where the model can continue to learn and refine its predictions as more data becomes available.
  7. Challenges and Skepticism: Address potential skepticism by acknowledging the challenges in distinguishing AI-generated text from human-written text, especially as AI models become more sophisticated.

By grounding your explanation in the principles of Bayesian inference, you can provide a logical and mathematical basis for your concept. This approach helps to clarify that the detection of AI-generated text isn’t based on arbitrary criteria but is rooted in a process of probabilistic reasoning and learning from data.

1 Like

Damn, I always hate it when evidentiary technique with citation of well-reasoned research can so easily be discredited by asking an AI.

6 Likes

Why are you so… clearly, absolutely hostile to the idea of being detected when trying to actively lie to people?

1 Like

I get where you’re coming from, but I’ve gotta say, I’m not fully on board with AI writing detectors.

Firstly, let’s think about the training data. If the AI is trained on a limited or biased dataset, it’s not going to be great at spotting AI-written text that falls outside of that. It’s like trying to find a needle in a haystack when you’re not even sure what the needle looks like.

Secondly, AI just doesn’t have the human touch. It can’t understand context or nuance in the same way we do. So, it might flag something as AI-written just because it’s a bit unusual or doesn’t fit the ‘norm’. And on the flip side, it might miss AI-written text that’s cleverly disguised.

And then there’s the whole transparency issue. AI is notorious for being a ‘black box’ - we can’t always figure out why it’s made a certain decision. So, if an AI detector flags a piece of text, we can’t necessarily understand why.

So yeah, while it’s a cool idea in theory, I think the unreliability of AI writing detectors makes them a bit of a risky business. We’ve gotta be careful not to put too much faith in them without understanding their limitations.

4 Likes