ccl
1
Retrieval augments the Assistant with knowledge from outside its model, such as proprietary product information or documents provided by your users. Once a file is uploaded and passed to the Assistant, OpenAI will automatically chunk your documents, index and store the embeddings, and implement vector search to retrieve relevant content to answer user queries.
Referencing https://platform.openai.com/docs/assistants/tools/knowledge-retrieval
There are several chunking strategies that are used in the wild (ie. fixed sized, context aware, etc), all with different tradeoffs. What chunking strategy (or strategies) does the assistant knowledge retrieval use?
6 Likes
kesar
2
I would love to know more about the chunking strategy too
Same question here, as well as how pricing works since we have no control over how much of the chunked text is used in model invocations
1 Like