When using the API in python I set the model to GPT-4 as shown below:
response = openai.ChatCompletion.create(
model="gpt-4",
messages=[
{
"role": "system",
"content": "You are a helpful assistant."
},
*[
{"role": "user", "content": msg}
for msg in conversation_history
],
],
max_tokens=2048,
n=1,
temperature=0.8,
)
However when I send a request in Python asking āWhat gpt model is this?ā this is the response:
I am an AI language model based on OpenAIās GPT-3.
When I ask the same question again on the web version with ChatGPT plus, I get this response:
I am based on the GPT-4 architecture, which is an advanced version of OpenAIās language models. The GPT-4 model is designed to generate human-like text and is capable of understanding and responding to a wide range of questions and prompts. However, please note that my knowledge is limited to information available up until September 2021.
I confirmed that I have GPT-4 API access after receiving the invite this morning and under āusageā in my account profile, the requests are shown using āgpt-4-0314ā. But as you can see here the responses are way less detailed compared to the web version. Feels like its not really using GPT-4, anyone else have this problem?
I was having the same challenge and am seeing similar results.
I have tested both gpt-4 and gpt-4-0314 as the model, and only have the gpt-4-0314 requests show up in my usage.
An interesting observation though, was that my tests using āgpt-4ā in my model field, generated a response to my question request / question (although it told me it was gpt-3), yet it did not generate an entry in the Daily usage breakdown, where I can see the gpt-4-0314 appearing.
For me, gpt-4 or gpt-4-0314 doesnāt make a difference, however itās important to know if we are evaluating the right results from the correct model in order to compare them effectively.
Iāve had a moment of clarity on this. I was looking at it the wrong way.
The appropriate test to see if we have the GPT-4 model, should be to see if it can handle greater than 4,096 tokens.
OpenAI already tell us in the documentation that the model is only trained on data Up to Sep 2021. Therefore, from that dataset, it wouldnāt know what GPT-4 is. While there may have been additional fine-tuning to tell it that it is now GPT-4, that would only be a superficial addition to the model.
The advance, comes not from the training corpus, but from the analysis of that same data in an enhanced way.
The available token count, should surely be the measure for us at this point in time?
Itās my best assessment at this time, however Iām open to see if this is a useful & accurate measuring stick!
Iām having the exact same issue. I set the max_tokens to 8096 and it does go through, but I really would like to see if thereās another way to test.
Hey! Iām fairly new to all this, but Iām having the same issue in the Playground. It says itās GPT-3. Also, the replies seem to be way less sophisticated than they are on ChatGPT-4 Plus?
I am not a developer, the only reason I want access to the API is to get the same features as ChatGPT-4 Plus, without the 25 messages per 3 hours limit, since pay as you go seems much more fair to me. But right now, the capabilities of the two models Iām seeing donāt even compareā¦
Were you able to figure out a solution to this issue? I am having a similar issue when using playground. Even though I have gpt-4 as the selected model, the model says its max token limit is 4096.
I have two theories as to the GPTās responses to questions about itself:
It is never trained on the specifics of itself at this level (āYou are GPT-4, now go forth!ā), thus, it has to respond confidently and probably inaccurately from the info it it is trained on which only goes to 2 years ago, or whatever. Thus, it could be GPT-4 and just not aware that it is. The API JSON attribute tends to bear this out. This is a hallucination or two types, not being specifically given a piece of information (which Iām not sure it ever is) so it must guess (not to anthropomorphize it), and lack of recent info generally.
I forget what my second theory was but its probably been spun into theory 1.
Oh, here it is, when GPT-4 gets overloaded, us pleebs get responses from lower versions of the thing, which is kinda explained here there any other places.
But it all changes so fast who can know? BTW - Iām just along for the ride just like all of us. Iām just an old dev keeping the brain cells poppin.
The GPT models are trained with data that finishes in September of 2021, at that time GPT-4 did not exist, so the model is not aware that it exists.
ChatGPT has a system message that tells it which model it is, the API lets you define the system message yourself, so you would need to add a system message saying āYu are GPTR-4 a large language model by OpenAIā¦ā
Please take advantage of the forum search system for other commonly answered question you may have.
hello, im asking the same question to āgpt-4-vision-previewā, and it replies that its gpt-3. your reply about 'not knowing bcz its data upto spt 2021 seems not clear here, right? bcz āgpt-4-vision-previewā data is upto Apr 2023, right? and when i ask it ācan you accept imagesā? it says āno, im only text based modelāā¦ why is so?