Hello OpenAI Community,
I’m developing a backend on Kubernetes to communicate with and endpoint connect to OpenAI’s DaVinci engine. Do I need to have my Kubernetes framework be Flask if I’m using Python?
Kind regards,
Julian
Hello OpenAI Community,
I’m developing a backend on Kubernetes to communicate with and endpoint connect to OpenAI’s DaVinci engine. Do I need to have my Kubernetes framework be Flask if I’m using Python?
Kind regards,
Julian
First off, welcome to the OpenAI community @julian2001!
You can use any framework that allows you to call the API endpoints, including Flask or Django. I personally have a project going that uses Django as the webserver and uses Celery to initiate tasks that calls the OpenAI API endpoints under each individual user.
I had plans to make it a full-stack business-grade OpenAI instance for everyone to use for free once I get it working, and maybe I still can here soon, but I’ve unfortunately had to put the project on hold for the time being since I have personal matters that have taken precedence currently.
Feel free to reach out to me if you (or anyone else here) plans to use Django for their project. I can certainly help!
Hi Nicholas,
Thank you that’s kind of you, will do. I hope everything works out for you.
Kind regards,
Julian Reyes
Hi Nicholas,
Does OpenAI support the CoDel ( Controlled Delay Queuing Development) algorithm for networking?
Kind regards,
Julian Reyes
Good question!
If you’re asking about specifics on how OpenAI has designed their infrastructure to handle the huge amount of network traffic, then you’ll receive a more official response by emailing support@openai.com and asking them about that. I unfortunately do not know the answer to that question, and I don’t believe that’s been talked about by OpenAI officials or community members yet.