Hello. We have created a wrapper chat assistant on top of gpt models. Everything is good, except one thing. We speak to some people and they say that Chatgpt gives better answers with same models.
Now I would like to know how can we also get better answers ? I suspect that we should use temperature, top_p, frequency_penalty parameters. If yes then how and with which values ?
I have gpt-4o, 4o-mini, 4 turbo, o1-mini. Better answers - means like “chatgpt solves their problem”. example: my friend wrote chatgpt to write a blog and in the blog to mention special keywords. it did it.
they same we did in my chat, but it did not mention the keywords.
Just look to these 2 images. One from chatgpt another from my chatbot. both using gpt-4o models. But as you see chatgpt gave much better and detailed answer to this prompt: tell me why a^2+b^2=c^2
Hope I was able to explain the problem.
You’re not using chatgpt-4o-latest in your below example.
In either case, you can probably just ask the model to be more verbose and in-depth with it’s responses. Although this is not something you should be configuring at a system level, but instead letting the user configure.
I tried that model. it gives the same result. chatgpt-4o-latest points to gpt-4o. So it doesn’t matter what model name you write in code. The question - why chatgpt gives better answer. I assume there is something should be done in config or system prompt. The prompt I gave is just one of examples.
chatgpt-4o-latest does not point to gpt-4o. In your own screenshot you will see that it specifically says “GPT-4o used in ChatGPT”
If someone says that ChatGPT is better than your model, then the closest thing you can do is use their model, and then let the user adjust their prompt.
I’m not sure what kind of answer you’re looking for here. Yes, adjust your prompt, yes, adjust your settings. Although if this is a service you’re offering then you should give these controls to the users.
@aidan_mclau@OpenAIDevschatgpt-4o-latest will track our 4o model in chatgpt, and is a chat-optimized model. our model from last week (gpt-4o-2024-08-06) is optimized for api usage (eg. function calling, instruction following) so we generally recommend that!
chatgpt-4o-latest is a chat-optimized model that is used by ChatGPT gpt-4o-[] is optimized for API-usage
Could it be that overtime they have converged? Maybe? But it’s a good first step to matching the quality of ChatGPT.
Based on both your topic title and the example images you provided, it seems like you’re looking for a preference-based response from the model. I say this because terms like “better” and “best” are subjective and depend on individual preferences. What one person finds well-structured and effective may not be as appealing to someone else.
With that in mind, if your preferred structured format aligns with the image you shared, I agree with @anon10827405 regarding the importance of giving clear, specific instructions to the model to ensure a structured output. Personally, I achieve this by explicitly defining how responses should be formatted and by creating a knowledge base that includes a reference template for the model to follow. In my experience, this approach generally yields more consistent and structured responses compared to allowing the model to generate replies without strict formatting guidelines.
That aside, you may want to consider providing your user base with a preferred prompting structure or guide. This can help ensure they receive “better” or more effective and structured responses compared to the outputs they are currently getting.