I really need more input length to keep context in novel creation.
The output could even be 2-4k.
I really need more input length to keep context in novel creation.
The output could even be 2-4k.
I made an app called PromptMinutes (Prompt Minutes from meetings etc) available on Apple Store. Cut n paste your API key in the ‘action’ view, and you are good to go. You can record , transcribe, and summarize. My tests show an hour worth of audio, transcription, and summary cost approx. 0.75USD . There is also a field to enter your custom prompt.
This is correct. Check out this working example:
[{“role”:“system”,“content”:“Step 1 - List 10 popular questions about generative AI.\nStep 2 - take the 1st question from the list from Step 1 and write a 1000 word article using markdown formatting, lists and tables where applicable.\nStep 3 - take the 2nd question from the list from Step 1 and write a 1000 word article using markdown formatting, lists and tables where applicable, lists and tables where applicable.\nStep 4 - take the 3rd question from the list from Step 1 and write a 1000 word article using markdown formatting, lists and tables where applicable.\nStep 5 - take the 4th question from the list from Step 1 and write a 1000 word article using markdown formatting.\nStep 6 - take the 5th question from the list from Step 1 and write a 1000 word article using markdown formatting, lists and tables where applicable.\nStep 7 - take the 6th question from the list from Step 1 and write a 1000 word article using markdown formatting, lists and tables where applicable.\nStep 8 - take the 7th question from the list from Step 1 and write a 1000 word article using markdown formatting, lists and tables where applicable.\nStep 9 - take the 8th question from the list from Step 1 and write a 1000 word article using markdown formatting, lists and tables where applicable.\nStep 10 - take the 9th question from the list from Step 1 and write a 1000 word article using markdown formatting, lists and tables where applicable.\nStep 11 - take the 10th question from the list from Step 1 and write a 1000 word article using markdown formatting, lists and tables where applicable.“},{“role”:“user”,“content”:“Execute steps 1-11"}
The GPT-3.5-turbo-16k model has a maximum response length of around 1500 tokens, regardless of the prompt used. Even in the Playground, the Maximum Length slider only goes up to 2048 tokens. The model’s increased token limit primarily benefits the input context rather than the output length. To generate longer responses, you can try sending multiple queries in a chat-like format, providing additional context with each subsequent message. However, the model is not designed to produce excessively long responses, and there is currently no way to force a specific response length, such as a 10k response.
I’m not sure that is completely accurate mosssmo, I can generate a 5000 token response with the 16k model.
Thanks, however increased contextual memory and output are what most would infer from 16K renewed token length.
Would you please share the prompt type used ? I am unable to generate that length even with a direct word count prompt.
Hi, I’m quite new to this world, so don’t mind me too much. I just wanted to comment that it works quite well for me:
The answer I receive usually contains, in a very high percentage, the requested amount of words.
In my case I ask it to write an article and I specify the range of words I want each section of the article to have.
For example, for the title I want it to be between 3 and 12 words.
The lead should be at least 100 words. The body should be x words.
And so on with the rest.
In the end I get an answer with the number of words I am looking for.
Hi Jeffer, how where you able to generate 5K token response from the API ? Because I’m not able to get close…
I’ll locate my prompt when I get back to my laptop but essentially it was asking for 120 paragraphs on a topic where each paragraph is 2 to 4 sentences in length.
Let me know how you go.
For my specific use case this wont work, but thank you for the insights!
The folks over at Future Fiction Academy seem to have this figured out. It is a method and is iterative. Even though the output context seems to be around 1500 words (2048 in tokens?) partial outputs seem to work just fine.
For example the folks over there have this prompt where they take the same chapter written twice but with two different prompt focuses. One focusing more on detail and another that concentrates more on dialog. Which produces a ton of text. With the 16k context window for GPT Turbo they are able to paste both of those into a megaprompt that combines the best aspects of each to very good results.
It rarely finishes in one go usually just abruptly stopping mid sentence. Then you just say continue and it keeps going from where it left off and because the 16K context window allows for all the massive amount of text in the initial prompt and the 1500 words without filling up it is able to keep the narrative coherent.
What’s the length they are able to produce with that? Do you know?