I am starting to use the fine-tuning function and so far like the precision I get from it.
In the past, I used Google Natural Language for my classification tasks and really liked the detailed evaluations I received after the training ( Confusion matrix, False positives, False negatives).
Is this information also available somewhere?
I understand there is the following command for getting some training results
openai api fine_tunes.results -i <YOUR_FINE_TUNE_JOB_ID>
However, I do not fully understand how I can leverage this output for improving the training and dataset.
That’s the correct command. Then you can download the results file for a few calculated classification metrics.
If you want something more custom, I recommend you call the fine-tuning endpoint on your validation or test set, to get the predictions, and then apply your custom evaluation function on the predictions.