Too long response time on API gpt-3.5-turbo model

Yes I have try exact same input on both. If I have use max_token value as 2000 then Davinci model also take approx 1 Minutes to 1.5 Minutes. Still it’s very high response time.

  • Is there any way to optimize the response time.
  • Can response time will be improve on paid module.

More information: I have tried https://api.openai.com/v1/chat/completions and https://api.openai.com/v1/completions