I need a chatbot that would use streaming response functionality to answer requests on the webapp. There are limitations on concurrent requests for single openai.create generator and I am wondering:
Is it possible to have a parallel streaming request-response webapp with GPT4 without Enterprise solution or Azure?