Not getting best answer from GPT models

Hello. We have created a wrapper chat assistant on top of gpt models. Everything is good, except one thing. We speak to some people and they say that Chatgpt gives better answers with same models.
Now I would like to know how can we also get better answers ? I suspect that we should use temperature, top_p, frequency_penalty parameters. If yes then how and with which values ?

Thanks !

  1. What model are you using
  2. What is meant by “better answers”?

I have gpt-4o, 4o-mini, 4 turbo, o1-mini. Better answers - means like “chatgpt solves their problem”. example: my friend wrote chatgpt to write a blog and in the blog to mention special keywords. it did it.
they same we did in my chat, but it did not mention the keywords.

chatgpt-4o-latest is the model used by ChatGPT. If you have a large system prompt get rid of it

2 Likes

It doesn’t support tool use unfortunately.

I’m not sure what exactly you’re looking for. ChatGPT doesn’t use tools for writing blog posts either unless you’re referring to search or actions.

Based on your link I’m assuming you are trying to build a auto-blog posting AI. You would be looking at agentic systems, not a single model.

You can use chatgpt-4o-latest for conversational abilities and then other powerful models for background tasks like research

Just look to these 2 images. One from chatgpt another from my chatbot. both using gpt-4o models. But as you see chatgpt gave much better and detailed answer to this prompt: tell me why a^2+b^2=c^2
Hope I was able to explain the problem.


You’re not using chatgpt-4o-latest in your below example.

In either case, you can probably just ask the model to be more verbose and in-depth with it’s responses. Although this is not something you should be configuring at a system level, but instead letting the user configure.

I tried that model. it gives the same result. chatgpt-4o-latest points to gpt-4o. So it doesn’t matter what model name you write in code. The question - why chatgpt gives better answer. I assume there is something should be done in config or system prompt. The prompt I gave is just one of examples.

chatgpt-4o-latest does not point to gpt-4o. In your own screenshot you will see that it specifically says “GPT-4o used in ChatGPT

If someone says that ChatGPT is better than your model, then the closest thing you can do is use their model, and then let the user adjust their prompt.

I’m not sure what kind of answer you’re looking for here. Yes, adjust your prompt, yes, adjust your settings. Although if this is a service you’re offering then you should give these controls to the users.

1 Like

chatgpt-4o-latest points to gpt-4o . I am trying to find the answer why the same model in chatgpt gives better answer.

I don’t think you’re fully understanding what the docs are trying to say.

chatgpt-4o-latest points to the same model used by ChatGPT. It’s a derivative of gpt-4o designed to be more communicative.

I don’t understand why you think it would just simply point to gpt4o, why would it exist then?

Your initial post was “My user says ChatGPT is better”. I am saying to you, the first step to be closer to ChatGPT is to use the model.

The first step is to use the same model that ChatGPT does. At least try it out and see the differences.

I can maybe understand because the docs really don’t highlight the differences, but here is a Twitter post about it:


Michelle Pokrass (@michpokrass) on X

@aidan_mclau @OpenAIDevs chatgpt-4o-latest will track our 4o model in chatgpt, and is a chat-optimized model. our model from last week (gpt-4o-2024-08-06) is optimized for api usage (eg. function calling, instruction following) so we generally recommend that!

chatgpt-4o-latest is a chat-optimized model that is used by ChatGPT
gpt-4o-[] is optimized for API-usage


Could it be that overtime they have converged? Maybe? But it’s a good first step to matching the quality of ChatGPT.

3 Likes

Based on both your topic title and the example images you provided, it seems like you’re looking for a preference-based response from the model. I say this because terms like “better” and “best” are subjective and depend on individual preferences. What one person finds well-structured and effective may not be as appealing to someone else.

With that in mind, if your preferred structured format aligns with the image you shared, I agree with @anon10827405 regarding the importance of giving clear, specific instructions to the model to ensure a structured output. Personally, I achieve this by explicitly defining how responses should be formatted and by creating a knowledge base that includes a reference template for the model to follow. In my experience, this approach generally yields more consistent and structured responses compared to allowing the model to generate replies without strict formatting guidelines.

That aside, you may want to consider providing your user base with a preferred prompting structure or guide. This can help ensure they receive “better” or more effective and structured responses compared to the outputs they are currently getting.
:+1:

1 Like