Doesn’t matter which model I use, it always answer gpt3.5.
However in my chatgpt plus I do get answer correctly.
is there a setting or parameter i missed?
{
"model": "gpt-4",
"messages": [{"role":"system", "content": "You are ChatGPT, a large language model trained by OpenAI. Follow the user's instructions carefully. Respond using markdown."},{"role": "user", "content": "what is the version of model you running?"}]
}
{
"id": "chatcmpl-7HtDo4GzaQQPnZYubYYP4jtF0v00h",
"object": "chat.completion",
"created": 1684498204,
"model": "gpt-4-0314",
"usage": {
"prompt_tokens": 46,
"completion_tokens": 33,
"total_tokens": 79
},
"choices": [
{
"message": {
"role": "assistant",
"content": "I am running on **OpenAI's ChatGPT**, which is based on the gpt-3.5-turbo version of the model."
},
"finish_reason": "stop",
"index": 0
}
]
}```