Yes, may be “turbo” it´s a little bit “pretencious” adjective for this model
I´m using curl with PHP on 500 tk max environment and the answers takes arround 30-50 secs to get ready.
Mine, too… I’m also having connection errors like this.
openai.error.APIConnectionError: Error communicating with OpenAI: (‘Connection aborted.’, ConnectionResetError(104, ‘Connection reset by peer’))
Same slowness here, plus occasional 502 Bad Gateway
responses after a long wait.
Sadly, the API is throttled for normal paying users. And at the moment we are getting also a lot of errors. Not very usable in the current state and we hope OpenAI will find a solution soon.
Is there a way to avoid this error?
I got a loop that broke today after 5 minutes and I didn’t even notice when it did.
The best way to adjust, I think is trying to change your solution to avoiding invoke the API or classify your demands to reduce the times calling it to decrease the total amount of time
I came here looking to see if other people were encountering this. I guess it is reassuring that it’s not just me. But also unfortunate because I’m hoping to launch my app in a few weeks and hope this improves.
Was gonna try using an another model but for this feature I need chat API to keep context. Guess I’ll just have to wait it out like everyone else.
I’m using the @backoff.on_exception(backoff.expo, openai.error.RateLimitError)
from backoff
library. Trying
for i in rlist:
try:
#mycode
except TimeoutError:
print("error")
continue
but it still breaks…
How do you get those reports of your queries? Is there an OpenAI webpage for that? I dont see that fine grained results in OpenAI API
mark. Yes,Now he’s incredibly slow, and he’s getting slower and slower.I hope the official can improve it as soon as possible
Performance of the OpenAI API is horrible for the moment. Are there plans to improve this soon because this instability in performance is blocking the roll out of our project.
API responses have been consistently 20-50 seconds for about a week now- unusable when ChatGPT itself seems faster than it has ever been
Is there a way to get someone from OpenAI to comment on this? Why are paying customers being rate limited into unusable latencies? The model is supposed to be “turbo” 30-40 seconds is not very “turbo” for some 100s of tokens. The API is wayyyyyy slower than the free chat? Why? I doubt it’s a technical issue, is that a strategic decision to limit developers? If so, I think OpenAI should be more “open” with the community
I think there are too many people using OpenAI API services. Like, I am a bit shocked that people are now saying ‘gpt-3.5-turbo’ is slow, because I remember ‘gpt-3.5-turbo’ had a good speed, with +1000 tokens. So… I feel like the server is packed now.
But my issue is more serious, because my company is using gpt-4, and gpt-4 is way slower though it is accurate. We are about to launch this internally, and I can imagine that our customer service team might say the chat bot is too slow and complain.
totally suffered from the same problem, awesome response, but awful slow
I just signed up for an OpenAI subscription myself, and I expected a similar response time as what ChatGPT is using, but using the gpt-3.5-turbo
model, the response times are 30-60 seconds, or timeout completely.
I’m only using it for a demo application, but it’s almost unusable due to this performance, and it’s extra disappointing I had to pay for this to experience this.
Actually, now we’re lucky to have the ‘stream’ option, which allows users to see the GPT generating answers. With this way, people can feel like they’re not stuck. Still slow, but now is much better.
How do you achieve the stream option to run?
stream
boolean
Optional
Defaults to false
If set, partial message deltas will be sent, like in ChatGPT. Tokens will be sent as data-only server-sent events as they become available, with the stream terminated by a
data: [DONE]
message. See the OpenAI Cookbook for example code.