When I make an API call from my code using the gpt-4
model like so:
{
"frequency_penalty": 0,
"messages": [{
"content": "...",
"role": "user"
}],
"model": "gpt-4",
"presence_penalty": 0,
"temperature": 0.1
}
I get the following response:
<html>
<head><title>400 Bad Request</title></head>
<body>
<center><h1>400 Bad Request</h1></center>
<hr><center>nginx</center>
</body>
</html>
If I make the same (so far as I can tell) call from the command line, I get an actual response. If I make the call with gpt-3.5-turbo
I get an actual response. I’m nowhere near the token limit on these calls.
The interesting thing is the error is coming from nginx, and I’m getting the error response in HTML (not the JSON error response body I’d get if I had exceeded the token limit, for example.)
Does anyone know what’s going on?