I have a project where I want to fine-tune an OpenAI GPT-3 model via API. My aim is to create a model that follows the language of my company. To achieve this, I want to fine-tune it with many texts, but I do not have a specific prompt for the texts.
My question is whether I can simply upload the texts to the API for fine-tuning, and whether it’s better to upload them as a txt file or a JSON file. I would appreciate any advice or recommendations from the community on how to proceed with this task.
I’ve worked on this application myself. As far as I understand, you have to create the jsonl file finetuning, but you leave the prompt column blank, and fill the completion column with your test.
Output from this model will be in the style of your training text.
If you want the model to repeat facts and knowledge of you company, you need to use embeddings.
If you want the response to have the same general tone and language of your company documents, you need to create prompt-completion pairs by taking chunks of data, using an AI model to neutralize the text and then set your prompt completion for the fine-tune as:
This is awesome!! Do you know if there is any study or paper published on this technique? (I’m happy if a technique just works, but my professors want a little more background research.)
No paper for tone, other than what was mentioned in this post
But try it and see what you get. I have found that these days GPT-4 has the best performance at neutralizing tone. Turbo get’s apologetic, and adds way too much noise instead of following the directions.