P.S.: I’m using a training file with examples containing functions (function calling). See “Fine-tuning examples > Function calling” in the API Doc: OpenAI Platform
Hi, I have the same problem since an hour and a half, I validated my jsonl file with the script in the documentation and it should be ok. Maybe it could be an openai error?
It could be an issue on the OpenAI side of things, but looking at the example file in this case did not help much as it’s quite large and complex, It would be better to have a very simple file that just contains placeholders so it’s easy to view and test.
Token counts of conversations seem to be nearing the limits allowed, and you also have functions.
Without actually processing the json overhead to chat format overhead, example 2 is 4268 tokens.
I would suggest that this huge converasation only trains the AI how to respond after having received a huge conversation history. You should also have the shorter exchanges that grow to get you there.
Actually my file is quite small, because I am just trying to specialize the model to extract some data from a txt. I have 10 examples with a “system” message, 2 “user” messages that are always the same request, and then the txt from which I need to extract the file and last I have the “assistant” where I specify what data I want.
the structure is the following:
{
"messages": [
{
"role": "system",
"content": "ChatGPT you are a chatbot that extract data"
},
{
"role": "user",
"content": "message where is specified what data i need"
},
{
"role": "user",
"content": "another message where is specified other data i need"
},
{
"role": "user",
"content": "text"
},
{
"role": "assistant",
"content": "data"
}
]
}
after I have compose all the message I validate it in jsol format and send to the api