GPT 4 I get this: textPayload: " message: “This model’s maximum context length is 8192 tokens. However, you requested 8808 tokens (617 in the messages, 8191 in the completion). Please reduce the length of the messages or completion.”,"
gpt-3.5-turbo-16k = textPayload: “Error in chatGPTPrompt function: BadRequestError: 400 This model’s maximum context length is 16385 tokens. However, you requested 17001 tokens (617 in the messages, 16384 in the completion). Please reduce the length of the messages or completion.”
Why is this happening? Why is it trying to use the max amount of tokens, rather than the amount it needs? Or is this a bug? Because it was working perfectly a few days ago, with gpt-4 turbo
Hey! Can you give me an example of the prompt that is causing this?
Also, to be clear, every model has a different context window length. GPT-4 right now is defaulted to 8k, GPT 3.5 Turbo 16k has 16k, and GPT-4 Turbo has 128k. So if it worked with 4-turbo, that does not mean it will work with the other models since they have lower context.