I’m using gpt-4o-2024-08-06 specifically because it support 16384 completion tokens. However, when I fine tune the model and try to substitute the new one into my code (based on gpt-4o-2024-08-06) I get the error:
BadRequestError: Error code: 400 - {‘error’: {‘message’: ‘max_tokens is too large: 16384. This model supports at most 4096 completion tokens, whereas you provided 16384.’, ‘type’: ‘invalid_request_error’, ‘param’: ‘max_tokens’, ‘code’: None}}
Is this a known issue or am I doing something dumb?
However, when you fine-tune gpt-4o-0806 (which is the only available option for fine-tuning a gpt-4o model), the max output tokens may be shortened back to 4k.