Despite “continued support” stated, this sounds like Chat Completions deprecation to me, when new models can no longer be used…
Is Response API available for all users yet? I can’t access it with my API key, though it has access to gpt-4o and gpt-4o-mini. It returns 403 forbidden “Project … does not have access to model gpt-4o-mini…”, even though with the same API key I can hit chat completion using the same models.
API keys can have allowed endpoints set on them.
You will note that there is no ‘/responses’ showing here in “Restricted” where you must enable individual endpoints:
An “All” API key was immediately applicable to the Responses endpoint for me.
You may need to simply generate a new API key to be placed in your application for the endpoints behind “all” to be properly populated.
Thanks for your reply! Yes I set the permission to “All”, but I can’t seem to hit the endpoint successfully. I’ll try your suggestion, if it’s not succeed I’ll start another thread. Thank you!
Any ETA for code interpreter for Responses API?
Is there a Completions to Responses migration guide? I couldn’t find one … that would be useful.
However, in any case I’m not 100% a fan of switching unless Responses becomes an industry standard, because there are circumstances where you want to be able to support both OpenAI and third parties services, e.g. in generic open source repos.
Today I might want to use GPT 4.1, but tomorrow I might need to switch to Meta’s Llama 4 perhaps, something I can do with far less friction with Completions.
Great find, thanks, very useful!
I understand about not wanting to switch and it’s hard to justify any reason. You get an internal iterator and response id storage for:
- server-side chat state: that can run a new model up to its 1M on chat with no management of the length;
- file search that has injected messages saying “user uploaded files”, destroying most use-case;
- web search that takes control and only produces a dump.
- reasoning “thinking” - then blocked without personal ID verification
…to then take away logprobs, logit_bias, etc.
I think the one benefit you get from the Responses API (if just doing single transaction “completions” and local tools) is reasoning summaries?
This is the discussion I have been looking for, thank you all for your insights!
They added a selector in the right panel for an equivalency of code for both versions in each api docs:
Myself and my team are working on a middleware server.
It lets you use Responses API (+features! We’re working on Web search, file search, etc.) but with any chat-completions compliant AI Provider.
It’s called “openai-responses-server” on GitHub (Can’t include links..)
And we’re adding it as pypi package or docker image.
Any update on the timeline for Code interpreter? This is a key feature keeping us from switching to the Responses API
Hi, I am still waiting for “previous_response_id” (as a way to preserve conversation) to handle truncation of the chain based on message date/time , max number, or max tokens. With the release of models handling 1M tokens, this is even more important to control costs.