We’ve been using the Assistants API for a few months now and are generally happy with its ease of use and all the built-in features. However, with the recent latency issues, I’m starting to wonder if switching to Completions might be the way to go for a faster, more responsive chatbot.
I’m looking to connect with anyone who’s built a complete chatbot solution—from setting up a vectorized knowledge base (we’re experimenting with Supabase) to integrating file uploads and managing conversation history. If you’ve tackled these challenges and found a way to streamline your system, I’d really appreciate hearing about your experiences and any tips you have.
This sounds brilliant. BUT, how do we actually do this? And do you have a live example of a chatbot thats built in this way? Would be interesting to try out.
To give you some context, my company provides chatbot solutions for different clients. Some are E-commerce stores, other are real estate companies etc.
With risk of sounding naive, is it possible to build costumer service chatbots for different clients this way? And are these bots faster than just using Assistant API? Because that is my ambition, to build great chatbots faster, and better.
for me the assistants are perenially in Beta. Sometimes they work beautifully, some times they suck (pathetically slow) and some times they don’t work at all.
The chat completion is much more mature and adopted even outside of openai as a standard way of interacting with the LLMs.
If you begin to deep dive into these solutions (esp for clients), you may find it necessary to set session timeouts (with loggedIn user experience) and the ability to control costs. Implementing with chat completion with Threads, Assistants and Messages exclusively as a transactional interface ( as opposed to also inferencing), you get a vastly different control surface, imo