Hello everyone! I am trying to feed gpt-3.5 with my own “custom” knowledge base. But I found out that this is very expensive, because with every message the whole knowledge base has to be given again to the ai which if the knowledge base is really big might be a couple thousand tokens per message. I’ve looked into vdbs but i dont really get how i could use them in a chat completion api call. Could somebody help me out here? Thanks!
For context. This is my current api call data:
{
"model": "gpt-3.5-turbo-0125",
"messages": [
{ "role": "system", "content": knowledge },
{ "role": "user", "content": content }
]
}