How creat How to format a fine-tuned file and upload it through n8n?

Hey guys!

I’m focused on doing a Fine-Tunning project using n8n to upload assertions in one of the models…

In the end, I will use OpenAI to answer questions asked through Whatsapp.

I imagine the process steps would be:

  1. Create a text file with a summary of the company and the main questions and answers. (done)
  2. Adapt the text to follow fine-tunning best practices (I need help with the pattern I should create to teach AI)
  3. Create json of smaller parts to send to OpenAI. (I need help with the pattern to be created)
  4. The http template to be created in n8n for sending to AI.

Can anyone give me a light on the solution?

It seems that this matter is very simple but it is not clear to me.

I really appreciate the help.

Generally if you are looking to provide additional knowledge to GPT the best way is through embeddings, not fine-tuning (which is better at classification/structure changes). LangChain project is probably the easiest way to get a proof-of-concept running so you can evaluate the results.

Otherwise fine-tuning guide has details on best practices.

Thanks for trying to help me.

I’m actually wanting to teach AI to answer the main questions that are asked in my business.

From the information I read, it said to follow these steps:

  1. Create a basic document with contextualization and as many questions and answers as possible.

  2. Transform it into a language that makes it easier for AI to better understand and absorb the knowledge case.

  3. Raise this base as a fine tuned base.

  4. And continue training.

At this point I need to standardize the file. I’m sure someone must have already developed a way to transform each paragraph into a json compatible with the data model for fine tuning.

And in the sequel to know how this base using an https node from n8n through the api.

And excuse my complete ignorance, but I have no idea what langchain is…

The fine tuning guide I linked has all the information on that format.

Again, the consensus is not to use fine tuning for knowledgebase Q&A chatbot style applications. Here’s a relevant LangChain tutorial. You can search these forums for fine-tuning vs. embeddings for more discussion.

https://python.langchain.com/en/latest/use_cases/question_answering.html