I just don’t get it. I am building a ChatGPT for multiple PDFs.
Now I have multiple PDFs which I have created embeddings of and saved in MongoDB database.
Now when a user asks a specific question, I get multiple outputs from GPT-3.5 and a lot of it does not make sense. Even though real answer is still in there.
Is there an API i can use to go through all data of PDFs and figure out which one is relevant and then I run GPT-3.5 only on that specific part of PDFs
The solution works for multiple pdf as well. All the generated embeddings are stored in a single vector db. When you make a query, top K results are fetched as context and sent to ChatGPT which then generates a response