I want to understand the following
-
Does open AI retain the information it gets from these queries and documents? If the documents used to create the vectorized knowledge base if highly sensitive and private, is it advisable to use open AI and its APIs to help process that data?
-
How large can this vectorized database be? Are there any limits? Will the LLM be able to go over a large pool of embedded data and give out correct information?