ETA: Seems there’s more info here on Hacker News… I’ve not personally got the email yet.
So… not sure if it’s Chat endpoint or if they’re bringing back legacy endpoint for it? The lower price is appreciated, though!
From HN…
I haven’t checked an email but using the completions api I do get a response using model “gpt-3.5-turbo-instruct”. Interestingly the completions api is deprecated though? Are they bringing it back?
I see it’s briefly mentioned in their docs under the deprecations section.
Note: The recommended replacement, gpt-3.5-turbo-instruct, has not yet launched. Impacted customers will be notified by email when it becomes available.
I can just copy it here? I do not think there’s anything sensitive in this:
Hello!
We are excited to announce the release of gpt-3.5-turbo-instruct, our latest model that serves as a replacement for several deprecated models, such as text-davinci-003.
Key Features:
Gpt-3.5-turbo-instruct is an InstructGPT 3.5 class model. It’s trained similarly to previous Instruct models such as the text-davinci series while maintaining the same speed as our turbo models.
Pricing:
We are committed to making cutting-edge technology accessible, so we have priced gpt-3.5-turbo-instruct in line with our other turbo GPT-3.5 models with 4K context.
Thank you for being a part of our journey and for building with OpenAI. Your support enables us to continually advance and bring you the best in AI technology.
Best regards,
The OpenAI team
I am simple man - I do not think I understand what “priced in line” exactly means. I already hit the endpoint, hopefully will not go bankrupt.
Not too sure. It works in the playground but also doesn’t send it in the actual payload Maybe they expect us to do it manually? The “inject start text” is added to the payload before actually being sent out. Based on the fact that there was no logprobs added to them makes me believe that yes, we may need to manually do it now.
It could also just have always been the way the playground works
Unfortunately, the completion endpoint doesn’t encode special control tokens such as <|im_sep|> from text, so one can’t completely emulate the chat endpoint (or see if ChatML fine-tune is still there):