How to get a quick response from chat via API?

Asking a question using the API I usually have to wait a long time for an answer, like 25-30 seconds, which is much longer than through www services. When I turn on streaming responses, it seems very inefficient, because the stream comes very often, every few characters (one token?), but I found that by limiting max_tokens to a small number, e.g. 150, I get a quick answer, but unfortunately the answer is cut off. Setting other values ​​like 1000 or 2000 can also be cut off and you have to wait longer. Would there be something in between streaming and regular: e.g. every 100 tokens I would receive a partial answer, which would be continued?