Hi Community,
I’ve been working on a project that uses the GPT-3.5-Turbo model. I demoed a version of the project mid July and the max response generation timeout set for the project is 8 seconds. For all the use-cases I have, the demo was working perfectly. But running the exact same code today, I see that the use-cases are failing. Specially the chatCompletion requests that result in function call request have a significantly higher latency.
I tried the exact same environment as last month so I’m sure almost nothing has changed on my setup. So I’m wondering if the APIs have gotten slower now. Has anyone else also experienced the same thing?
Thanks!