ChatGPT API responses are very slow

ChatGPT API responses are very slow, even for short API calls with 200-400 tokens take 20-30 seconds. Is there any way to make the response faster.


Hi @nandha

Yes, things are slow based on the demand. I just checked for you by sending 300 words of lorem epsom text to the chat completion API for you and got these results:


Total Tokens: 826, Completion API Time: 16.17 seconds
Total Tokens: 866, Completion API Time: 14.434 seconds
Total Tokens: 1313,  Completion API Time: 38.629 seconds

I don’t think there is much you can do at the moment as the issue is with the performance of the turbo model(s.) You could switch to another model, which have tested to be faster than turbo these days.



Appendix: Example Completion


hi ruby_coder,
im using api and model gpt 3.5 turbo too. but the response is very slow. im calling api by python.

Yeah, is it slow, for sure now, I tested again for you now, completion time was nearly 22 seconds:

My advice is to relax and do something less frustrating until the issue on the OpenAI infrastructure side improves, if you can.




Yes, may be “turbo” it´s a little bit “pretencious” adjective for this model :slight_smile:
I´m using curl with PHP on 500 tk max environment and the answers takes arround 30-50 secs to get ready.

Mine, too… I’m also having connection errors like this.

openai.error.APIConnectionError: Error communicating with OpenAI: (‘Connection aborted.’, ConnectionResetError(104, ‘Connection reset by peer’))

Same slowness here, plus occasional 502 Bad Gateway responses after a long wait.

Sadly, the API is throttled for normal paying users. And at the moment we are getting also a lot of errors. Not very usable in the current state and we hope OpenAI will find a solution soon.

Is there a way to avoid this error?
I got a loop that broke today after 5 minutes and I didn’t even notice when it did.

The best way to adjust, I think is trying to change your solution to avoiding invoke the API or classify your demands to reduce the times calling it to decrease the total amount of time

I came here looking to see if other people were encountering this. I guess it is reassuring that it’s not just me. But also unfortunate because I’m hoping to launch my app in a few weeks and hope this improves.

Was gonna try using an another model but for this feature I need chat API to keep context. Guess I’ll just have to wait it out like everyone else.

I’m using the @backoff.on_exception(backoff.expo, openai.error.RateLimitError) from backoff library. Trying

for i in rlist: 
    except TimeoutError:

but it still breaks…

1 Like

How do you get those reports of your queries? Is there an OpenAI webpage for that? I dont see that fine grained results in OpenAI API

mark. Yes,Now he’s incredibly slow, and he’s getting slower and slower.I hope the official can improve it as soon as possible


Performance of the OpenAI API is horrible for the moment. Are there plans to improve this soon because this instability in performance is blocking the roll out of our project.


API responses have been consistently 20-50 seconds for about a week now- unusable when ChatGPT itself seems faster than it has ever been

1 Like

Is there a way to get someone from OpenAI to comment on this? Why are paying customers being rate limited into unusable latencies? The model is supposed to be “turbo” 30-40 seconds is not very “turbo” for some 100s of tokens. The API is wayyyyyy slower than the free chat? Why? I doubt it’s a technical issue, is that a strategic decision to limit developers? If so, I think OpenAI should be more “open” with the community


I think there are too many people using OpenAI API services. Like, I am a bit shocked that people are now saying ‘gpt-3.5-turbo’ is slow, because I remember ‘gpt-3.5-turbo’ had a good speed, with +1000 tokens. So… I feel like the server is packed now.

But my issue is more serious, because my company is using gpt-4, and gpt-4 is way slower though it is accurate. We are about to launch this internally, and I can imagine that our customer service team might say the chat bot is too slow and complain. :sob:

1 Like

totally suffered from the same problem, awesome response, but awful slow :smiling_face_with_tear:

I just signed up for an OpenAI subscription myself, and I expected a similar response time as what ChatGPT is using, but using the gpt-3.5-turbo model, the response times are 30-60 seconds, or timeout completely.

I’m only using it for a demo application, but it’s almost unusable due to this performance, and it’s extra disappointing I had to pay for this to experience this.

1 Like