_j
March 3, 2024, 4:37am
11
Yes, it is so. I make note of the extreme expenditures, and others also getting hit with big bills, one run potentially using many dollars of tokens, the week this came out:
Assistants is unshippable.
The chat length (threads) is not under your control, and the assistants backend will fill the context with all you have and all that can fit.
The retrieval function is not under your control, and the assistants backend will fill the context with all you have and all that can fit.
At least that’s what they say about retrieval, when in fact the AI is given a “browse the files” function OpenAI decided to not document where it can autonomously make multiple calls,…
There is no way currently to temper the threshold or amount of data being fed to the AI. Your assistant would permanently know about icons when chatting, because “create assistant” created it with connection to the file IDs.
You would have to disengage files with “modify assistant” - a pointless function to use in runtime when you can’t know when your file will be relevant to a user’s chat query.
If a retrieval file or collection is uploaded and connected to an assistant, and the file is small…