Gpt-4-1106-preview in Playground needs some fixes

Same problem. I want to translate books but I had to split it in 80 parts first. The book was around 260.000 tokens. I would love to split into only two parts :frowning: The 128k context window is misleading, because that’s not the in or output limit as I understand it.

Currently I am limited to 4096 tokens max output using the gpt-4-1106-preview API model.

1 Like

Yes, this is lame, but is also exactly what costs actual money that they have curtailed already in AI training: quality attention heads go quadratic with longer lengths

openai.error.InvalidRequestError: max_tokens is too large: 75000. This model supports at most 4096 completion tokens, whereas you provided 75000.

So for input/output tasks like checking spelling, this model gains you nothing except lower quality.


edit, confirmed:

2 Likes

I am in usage tier 4 but am still receiving a 32k limit in the gpt4turbo assistant playground. How do I get to to the 128k limit?

Hi and welcome to the Developer Forum!

This seems to be a technical limit on the code side, not the model, it should get corrected soon.

1 Like