I’m starting to do a lot of code gen that asks o1-mini to write code for calling OpenAI models directly. I’ve noticed that it has a tendency to want to use either gpt-4
or even gpt-3.5
when it calls the completion API. That’s presumably because that’s what its seen prior to its knowledge cutoff. The model also doesn’t know about newer features like Structured Outputs or even JSON mode.
Given that OpenAI controls the fine tuning of the models they should fine tune all of their models to know about their latest model names, API endpoints, and features. I mean shouldn’t o1-mini
have a natural preference to want to call itself?