Hi GPT community,
I am a developer from Internet Computer Protocol blockchain and I wish to fine-tune GPT to learn our native language Motoko. We have exponentially growth for new developers, No. 1 GitHub commit as Layer 1 blockchain in last month, so I am very confident our current example code is enough for fine-tune.
My initial approach is to fetch GitHub repo by tag of Motoko language and then use self-instruct technique (like Alpaca from LLaMA) to generate prompt given code. I wish to get some critics before I dive in. How do u think of this approach for dataset generation?
Additionally, I am open for suggestions to create a plug-in rather than fine-tune. If plug-in is more legit, what message my backend sever shall response to users’ inquiries through GPT?
Thank u for suggestions.