I submitted a fine-tuning job today with davinci as the base model. My token count estimate was 1,431,706. But after finishing the job, in account daily usage breakdown, it is listed as
5,726,824 trained tokens which is 4x my estimate.
The result doc at the end of the job shows 11,806,768 as the
elapsed_tokens (with repeats) in last row. Using the default 4 epochs, that would likely mean 2,951,692 tokens per epoch, which is 2x my estimate so does not match, 0.5x of what is shown in the account so also does not match.
Can anyone offer more clarity into this? It has been 5 hours since I sent the message to their help and no response still so thought I would try out the community forum too. Charge estimate is also showing up as 4x my estimate. Does that mean rate needs to be multiplied by epochs? No information or page I saw suggest that cost needs to be multiplied by epochs.
(Unrelated, someone from OpenAI should really hook up GPT-4 to this forum to make search for answers painless in case it already exists in the forum)