I am using OpenAI API and the ChatGPT web GUI for image recognition tasks, and I have noticed a difference in their performance, particularly in the “thinking process.”
When using the web GUI, the model seems to generate code for recognizing specific items within the image and even performs a search for double-checking. On the other hand, the API only provides pure analysis and gives the final conclusion. Specifically, the GUI appears to handle finer details of the image and performs step-by-step reasoning, while the API’s output is more concise and limited to a final analysis.
I would like to make the API perform exactly like the web GUI, especially in terms of the thinking process (e.g., recognizing individual items and performing a double-check). Is there a way to adjust API settings or apply any method to make the performance of both the API and the GUI consistent?
Any advice or suggestions would be greatly appreciated!
That is not possible at this moment because the completions/responses API doesn’t support code interpreter, and the assistants API doesn’t support the newer models used in ChatGPT.
Basically, ChatGPT is a product that although it uses the same models as in the API, internally have a different toolset of extensions that makes it unique. The same behaviour can only be replicated for certain use cases, but not all of them.
Yes, but then you move to inability to verify what is offered is actually what is being run. You are blocked from using any functions (just one type of tool) - and ChatGPT doesn’t have just a functions tool for passing yours, it is loaded with unplaceable tool recipients and their internal instructions with any API model.