What did you mean by “random” input?
I had similar idea and experiments, but very recently, different model (GPT 4o) and different memory names and different limites. But same question. There is also the new file upload tool own size mysteries with respect to the model instance processing abilities from the environment back to the conversation memory management (obscure or difficult to find clear statements, specially in the web UI).
So empirical “truth” it came to. For the bots would send me on goose chases, with the upmost confidence I would succeed. But then it would base a bunch of its stuff on vague and wrong assumptions never shared. and then started the long quest to figure out how to stop thinking it understood just because the words, as human words would make it look like it. The reality of its environment, and my independent ability to share some common reality, the text files themselves.
what do chat bot words really mean…
now some context of my understanding:
-
I am talking about ChatGPT web Ui, assuming same memory mechanical model (which is hard to find a clear diagram of anywhere, lots of words in vain, and not sure all makes use of the same terminology, the dates on help pages are not meaning that all the page has been updated, just that something changed).
-
the date now is in October 2024 and the specifications that can be found would be those of GPT4o for my curiosity.
Models - OpenAI API
The language there is “context window” = 128K and “max output tokens” = 16386
I mention that because it seems to have changed, and particularly from the outdated corpus times. Which makes a good discrepancy check experiment about the bot lack of system prompt updating their own self-awareness of their model class and abilities.
They have to be artfully babysat. And the dumber they start the longer the conversation to weed out all the wrong lurking assumptions. And now with huge capacities. So, I did resort to figure out the same way. But I specified to use correct syntax but no semantics. All the words would be tokenizable. Maybe that is obvious.
My version of your idea:
-
I got to generate 3000 tokens outputs, but it might have been around the change of GPT 4o snapshot change. before it was 40xx tokens. I was wondering if one could not do a set of graduated extra chunks to find the upper bound and lower bounds.
-
but the idea to improve on your idea was that one could concatenate the file files upload them, and then ask to stitch them back in context memory, into various combination to “binary” search knowing what to expect. and a single line prompt asking to output various combinations of such set of input sizes (from the context memory then) to the response text chunk.
Parent context of mine (why did I go reality check rabbit hole):
This was proposed to me in one conversation as a mean to deal with long text document substrate for my text transformation projects (translation).
plea:
please fix my words to make more sense. to this community, if my words are not the most resounding shop talk. I am only a month-old at this intensive struggling with finding out what is that semantics they keep talking about. And this very simple reality check, is nagging me. So much obscurity. And not just the bots.
Recap of extended empirical method idea:
So the solution is to work from under and keep going up. but you can accelerate with using the context memory and asking to transfer from the upload file mechanism previous under output limit response back into the conversation. (does not even have to be the same, if you want to not run into aging and declining conversation lengths).
off-shoot question
(given new interface features that did not exist at op time).
Also I wonder about the same question for file uploads, and then transfer into the context memory (context window 128K tokens), a.k.a. conversation history memory or context, a.Not.k.a single turn memory which, I agree with your assessment seems to be sharing the “max output tokens” between user prompt and bot response.