Hi there,
To get a better understanding of how the knowledge retrieval process works behind the scenes, I conducted a series of experiments, each requiring a different approach of the retrieval process. Initially I just aimed to figure out, if the file size and the location of the information within the file has an impact on runtime and accuracy. But the focus shifted towards are more general question on how to best utilize the knowledge retrieval tool. Here are my key findings:
For trivial search tasks, the file size or the location of the information within the file does not matter.
For complex search tasks it does.
Some level of “abstraction” is performed if neccesary. E.g. it searches for synonyms or changes the threshold of the vector search.
Formatting matters!
There was an undocumented limit of 2M tokens for knowledge files (thanks @logankilpatrick for adding it!).
Everything is quite inconsistent and beta and sometimes the assistants just goes rogue and you might end up with $0.50 API calls without a decent answer.
A lot feels like a black box and it would be really helpful to at least get to see what exactly the knowledge retrieval process is adding to the context!! @OpenAI: Please populate the retrievalfield
I wrote a detailed article about the experiments and my findings. You can find it over here
I am curious to hear your thoughts on this and experiences you made with the knowledge retrieval process!