Documentation says that openai automatically creates the chunks and stores the embeddings.
Retrieval augments the Assistant with knowledge from outside its model, such as proprietary product information or documents provided by your users. Once a file is uploaded and passed to the Assistant, OpenAI will automatically chunk your documents, index and store the embeddings, and implement vector search to retrieve relevant content to answer user queries.
Where does the embeddings stored. Is it stored in some Azure service? Is there a way to access that embedding store?
If you watched the Microsoft keynote today, you may have seen mention of the Azure retrieval service that is used, so yes, it is on Azure services and is part of Microsoft’s vector and embedding service. I think it’s this Vector search - Azure AI Search | Microsoft Learn
No, the GPT Assistant does not operate by extracting text from a PDF, embedding it into vectors, and saving them into a vector database for subsequent searches. Instead, its approach to handling user queries about a PDF is different and does not involve creating or maintaining a persistent database of embeddings. Here’s how it generally works:
this is how GPT response. I dont know what it’s truth now