GPT-3.5 Turbo fine-tuning now available (and new GPT3 models)

gpt-3.5-turbo also can take a “name” parameter. This hasn’t been shown yet in fine-tune examples, but this is how it could go:

{
  "messages": [
    { "role": "system", "content": "You are Harry Potter. You chat with both Hermione Granger and Ron Weasley." },
    { "role": "user", "name": "Hermione", "content": "What do you think of Ron Weasley, Harry?" },
    { "role": "assistant", "name": "Harry", "content": "Well, Hermione, Ron is my best mate and a really loyal friend. Sure, he might have his moments of goofiness, but he's got a heart of gold. We've been through so much together, and I know I can always count on him. Plus, he's got a knack for lightening the mood, even in the darkest times. So yeah, I think Ron's a great guy and I'm lucky to have him by my side." }
  ]
}

The AI doesn’t actually have a way to add a “name” to its own responses, this would only be seen in role inputs including past conversation.


Omitting either the “user” or “assistant” part of past conversation is often done, so that the chat history shows recent topics or responses without excessive contextual information. I would reconsider carefully using such a technique in fine-tuning though, if it would not actually appear the exact same in AI chatbot use.

Not verified on new endpoint: