max_tokens was certainly handy for keeping prompt+messages to a
rough length. I have had mixed results setting this using just the assistants API. This would mitigate the need for post response checks.
3 Likes