I’ve noticed that since a few days ago the parameter max_tokens does not get respected anymore by the API, neither in GPT-3.5 nor GPT-4. When I set it to an invalid value though, I receive an error… Indicating that how I send the parameter is indeed correct and well… As mentioned it was working until a few days ago.
Now, if I set it to whatever value, the response does not stop with the reason “length” anymore. Anyone else seen this?
Indeed, I checked again and it was the case that in some other part the accounting was changed from response to total tokens so it always returned me the total tokens, my bad for not noticing it sooner…