Before I embark on this journey and invest a significant amount of time into this, I’d really appreciate some feedback to ensure what I’m trying to achieve is possible.
Performing a classification problem. Input is a string of text, output is a classification (i.e group 0,1,2,3,4). Both the input and output has real meaning and use of a LLM seems like an excellent solution. I’ve currently performed this using GPT vector embeddings as features for a random forest machine learning model, however I’d like to experiment with GPT finetuning if possible.
Question: Once I’ve finetuned a model once, can I continue to append to that model? I.e: First I finetune on 1000 datapoints, then I finetune the same model again on the next 1000 datapoints, and so on …
Question: Is a finetuned GPT model the right tool for the job here (classification problem).
Thank you! I think this explains why I was unable to get some of the base models like ada, davinci, etc, to work. It would just return model error. Reading this, it looks like I should be using the old API for these models. Cheers