Hi,
I would like to fine tune GPT4 and almost all documentation/guides is around GPT 3.5. I have some questions:
-
Where can I find the max_tokens per example for GPT4?
-
Other than the max_tokens_per_example mentioned here Data preparation and analysis for chat model fine-tuning | OpenAI Cookbook, can this be used as is for GPT 4 fine tuning?
-
If I fine tune a model and want to add more examples, from this forum looks like it will start giving more weightage to the newer examples. Is this the case with GPT4 also or does the model give equal weights for examples from both training sessions?
-
If I want structured output- will it suffice to simply have enough examples where the output is always structured, or do I need to do add it to the system message? The thing is that the prompt will be huge and repeated across all examples. The documentation is unclear on this.
It states:
“If you would like to shorten the instructions or prompts that are repeated in every example to save costs, keep in mind that the model will likely behave as if those instructions were included, and it may be hard to get the model to ignore those “baked-in” instructions at inference time.” I don’t fully understand what this means? Are they implying that whatever system prompt we use for fine tuning is exactly what we need to give later?
Thanks!
Anika.