I have noticed that if max_prompt_tokens is set low like ~500 some very strange hallucinations start happening with code interpreter and chat responses. probably due to messages being dropped due to tokens taken up by tool definitons.
If the max_prompt_tokens is set to ~10.000 or less some very strange hallucinations happen with file search.
Hi there. Have the same issue. I set max_prompt_tokens to 5 thousand with file search and get heavy hallucinations! Model started to create random answers with mentioning some strange document references… How to find the right number to strict input tokens? For example I have document with template answers, in file citation it takes only one paragraph from it, but read 10000 tokens and more as it read all file. How to optimise this file search?