Some time ago, there were limits listed on how many tokens a training example for fine-tuning can have for different models. Theese limits seems to have vanished from the documentation, I cannot find them anymore.
So I thought then the limit must just be the models context length. But no, I just go an “Example too long” error when trying to fine-tune gpt-4.1, which has a context length of 1M tokens.
Where can I find info on the maximum length of a training example for fine-tuning for different models?